Showing 1-12 of 12 projects
Run LLMs locally in C/C++ with high performance
High-performance C/C++ port of OpenAI's Whisper for speech recognition
A tensor library for machine learning, with a focus on large language models and automatic differentiation.
Unified, production-ready inference API to run open-source, speech, and multimodal models on cloud, on-prem, or your laptop.
A C/C++ implementation of Stable Diffusion and other diffusion models for image generation and processing.
An offline iOS and macOS library for running large language models like LLAMA, GPT-2, and RWKV using the GGML library.
A Vim plugin that provides LLM-assisted code and text completion to boost developer productivity.
Python bindings for Transformer models implemented in C/C++ using the GGML library.
A Go-based command-line tool for managing Ollama LLM models on Linux and macOS.
An efficient C++ implementation of the RWKV language model for fast CPU inference on various bit-width quantizations.
A JavaScript tool for calculating token/s and GPU memory requirements for large language models like LLaMa.
VS Code extension that provides LLM-assisted code and text completion for developers.
Get weekly updates on trending AI coding tools and projects.