Explore Projects

Discover 5 open source projects

Active filters (1):
Search: kv-cacheร—
Clear all

Showing 1-5 of 5 projects

BlinkDL/RWKV-LM

RWKV is an RNN-based language model with high performance, fast training, and flexible transformer-like architecture.

14.4K
Active
Python
LLM Frameworks
PyTorch
#language-model#rnn#transformer

LMCache/LMCache

Supercharge your large language models (LLMs) with the fastest key-value cache layer for lightning-fast inference.

7.5K
Active
Python
LLM Wrappers & SDKs
Caching
PyTorch
#llm#inference#cache

microsoft/LLMLingua

This project aims to speed up large language model (LLM) inference and enhance their understanding of key information through prompt and KV-Cache compression.

5.9K
Stable
Python
LLM Frameworks
Inference
Python
#llm#language-model#inference-optimization

HDT3213/godis

A high-performance, production-ready Redis server and cluster implementation in Go.

3.8K
Stable
Go
Caching
Databases
Go
#redis#cluster#go

Zefan-Cai/KVCache-Factory

Unified compression methods for KV caching in autoregressive language models like GPT-3.

1.3K
Archived
Python
LLM Frameworks
Caching
Python
#kv-cache#kv-cache-compression#llm

Stay in the loop

Get weekly updates on trending AI coding tools and projects.