ALEXANDER MÜLLER. Optimizing Large Language Model Inference: Strategies for Latency Reduction, Energy Efficiency, and Cybersecurity Applications. International Journal of Computer Science & Information System, [S. l.], v. 10, n. 11, p. 93–97, 2025. Disponível em: http://scientiamreearch.org/index.php/ijcsis/article/view/214. Acesso em: 12 mar. 2026.