Explore Projects

Discover 2 open source projects

Active filters (1):
Search: distributed-inferenceร—
Clear all

Showing 1-2 of 2 projects

flashinfer-ai/flashinfer

A Python library for serving large language models (LLMs) with high performance, including GPU acceleration and distributed inference.

5.1K
Active
Python
LLM Frameworks
Inference
PyTorch
#llm#inference#cuda

gpustack/gpustack

Optimize AI inference performance on GPUs with this Python library for selecting and tuning inference engines.

4.6K
Active
Python
Inference
CLI Tools
Python
#ai-inference#gpu-acceleration#performance-optimization

Stay in the loop

Get weekly updates on trending AI coding tools and projects.