Showing 1-20 of 20 projects
DeepSeek-V3 is a large-scale MoE language model with 671B parameters, optimized for efficiency and performance.
High-throughput LLM inference engine for developers
Fine-tuning framework for 100+ LLMs & VLMs
High-performance serving framework for large language and multimodal models
TensorRT LLM provides a Python API and optimizations to efficiently run large language models on NVIDIA GPUs.
A Python library for using and fine-tuning over 900 large language models and multimodal models for various AI tasks.
An unofficial Bangumi (similar to Douban) client app for Android and iOS, built with React Native.
A Python library for serving large language models (LLMs) with high performance, including GPU acceleration and distributed inference.
A foundational AI agent model for building agent-based, reasoning, and coding capabilities.
A simple JavaScript library for adding a customizable visitor counter badge to your website.
DeepSeekMoE is a novel Mixture-of-Experts language model framework for building highly specialized AI coding assistants.
A fast implementation of Mixture of Experts (MoE) for PyTorch, enabling efficient large-scale neural networks.
Custom Home Assistant component for smart thermostat control with AI-powered features
A global, black box optimization engine for real-world metric optimization.
Spiking Brain-inspired Large Models with efficient attention, MoE, and spike encoding for AI and ML developers.
PyTorch implementation of the Sparsely-Gated Mixture-of-Experts layer, a powerful AI model architecture.
Efficient communication library for GPUs, covering collectives, P2P, and EP for AI/ML workloads
Aria is an open-source multimodal AI framework for building vision and language models.
Uni-MoE is a large multimodal model family from Lychee, a Python library for AI model development and deployment.
A library for building Mixture-of-Experts (MoE) models from the LLaMA language model with continual pre-training.
Get weekly updates on trending AI coding tools and projects.