Alexander Müller (2025) “Optimizing Large Language Model Inference: Strategies for Latency Reduction, Energy Efficiency, and Cybersecurity Applications”, International Journal of Computer Science & Information System, 10(11), pp. 93–97. Available at: http://scientiamreearch.org/index.php/ijcsis/article/view/214 (Accessed: 12 March 2026).