optimization 2 Efficient Data Encoding for Large Language Models Nov 18, 2025 Optimizing LLM Inference Pipelines with Docker Caching and Model Preloading Oct 8, 2025