turboderp/exllama

A memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.

Python
AI & Machine Learning
LLM Wrappers & SDKs
MIT

2.9K

Stars

221

Forks

May 4, 2023

Created

Sep 30, 2023

Last Updated

Project Analytics

Stars Growth (1 Month)

+6

+0.2% change

Avg Daily Growth (1 Month)

+0.2

stars per day

Fork/Star Ratio (All Time)

7.6%

Normal engagement

Lifetime Growth

2.8

stars/day over 1.0K days

Stars Over Time

Forks Over Time

Open Issues Over Time

Pull Requests Over Time

Commits Over Time

AI-Generated Tags

llama
transformers
quantized-weights
memory-efficient
language-model

Comments (0)

Sign in to leave a comment or vote

Sign In

No comments yet. Be the first to comment!

Stay in the loop

Get weekly updates on trending AI coding tools and projects.