Explore Projects

Discover 3 open source projects

Active filters (1):
Search: awqร—
Clear all

Showing 1-3 of 3 projects

mit-han-lab/llm-awq

A library for efficient weight quantization of large language models to accelerate inference on edge devices.

3.5K
Experimental
Python
LLM Frameworks
LLM Wrappers & SDKs
Python
#llm#compression#acceleration

intel/neural-compressor

Optimizes large language models for low-bit precision and sparsity, improving model compression techniques.

2.6K
Active
Python
LLM Frameworks
PyTorch
#quantization#post-training-quantization#sparsity

casper-hansen/AutoAWQ

Implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference, for AI coding tools.

2.3K
Experimental
Python
Inference
AI Code Generation
Python
#quantization#speedup#inference

Stay in the loop

Get weekly updates on trending AI coding tools and projects.