Reliable model swapping for local LLM servers - seamlessly switch between llama.cpp, vLLM, and compatible backends
2.6K
Stars
191
Forks
Oct 4, 2024
Created
Mar 2, 2026
Last Updated
+6.4% change
stars per day
Normal engagement
stars/day over 519 days
Run LLMs locally in C/C++ with high performance
Run local LLMs on any device with GPT4All
Web UI for local AI with multiple backends and offline capabilities
Run frontier AI models locally across devices using RDMA and tensor parallelism
Sign in to leave a comment or vote
Sign InNo comments yet. Be the first to comment!
Get weekly updates on trending AI coding tools and projects.