LLM routing aims to reduce costs by directing simple queries to less capable, cheaper models, and complex queries to stronger models. RouteLLM is a framework that uses preference data and data augmentation to train routers that decide which LLM to use, achieving significant cost savings without compromising quality.
https://lmsys.org/blog/2024-07-01-routellm/
@yogthos
Excellent!
Detecting "simple queries" is clearly halting-complete, so someone's solved the halting problem at last!