Optimizing inference proxy for LLMs
GitHub – algorithmicsuperintelligence/optillm: Optimizing inference proxy for LLMs

“OptiLLM is an OpenAI API-compatible optimizing inference proxy that implements 20+ state-of-the-art techniques to dramatically improve LLM accuracy and performance on reasoning tasks – without requiring any model training or fine-tuning…”
May 14, 2026
Subscribe
Login
Please login to comment
0 Comments