Showing 1-20 of 22 projects
Self-hosted AI platform with Ollama and OpenAI API support
Run LLMs locally in C/C++ with high performance
Llama 2 inference code for running Llama models
PrivateGPT enables private document interaction with GPT without data leaks.
Open-source Grok-1 model for local inference with JAX
Self-hosted, open-source AI alternative to OpenAI with local LLM inference, no GPU required
Extracts structured info from text using LLMs with source grounding
1-bit LLM inference framework for CPU/GPU
Qwen3 is Alibaba Cloud's large language model series with enhanced reasoning and coding capabilities.
Official FLUX.1 inference repo for image generation & editing
On-device multimodal LLM for vision, speech, and live streaming on phones
Faster Whisper transcription with CTranslate2 for efficient speech-to-text
Deploy open-source LLMs as OpenAI-compatible API endpoints using BentoML's model serving framework.
Autonomous AI assistant infrastructure in Zig—fast, minimal, self-contained runtime for building AI agents.
Rust implementation of OpenClaw focusing on privacy-preserving AI model execution and security hardening.
Fast local neural text-to-speech engine for offline voice synthesis
Native multimodal model for high-quality image generation with text-to-image capabilities
Reliable model swapping for local LLM servers - seamlessly switch between llama.cpp, vLLM, and compatible backends
Large language model by Alibaba Cloud Qwen team for advanced NLP and AI applications
Lightweight AI assistant for ESP32 microcontrollers with GPIO, scheduling, custom tools, and memory.
Get weekly updates on trending AI coding tools and projects.