Alexander Müller. “Optimizing Large Language Model Inference: Strategies for Latency Reduction, Energy Efficiency, and Cybersecurity Applications”. International Journal of Computer Science & Information System 10, no. 11 (November 30, 2025): 93–97. Accessed January 25, 2026. https://scientiamreearch.org/index.php/ijcsis/article/view/214.