Pure C inference engine for Mistral Voxtral 4B speech-to-text model with minimal dependencies
1.5K
Stars
95
Forks
Feb 5, 2026
Created
Feb 15, 2026
Last Updated
+18.6% change
stars per day
Normal engagement
stars/day over 29 days
Run LLMs locally in C/C++ with high performance
Run local LLMs on any device with GPT4All
Web UI for local AI with multiple backends and offline capabilities
Run frontier AI models locally across devices using RDMA and tensor parallelism
Sign in to leave a comment or vote
Sign InNo comments yet. Be the first to comment!
Get weekly updates on trending AI coding tools and projects.