openai/evals

A framework for evaluating large language models (LLMs) and an open-source registry of benchmarks.

Python
AI & Machine Learning
LLM Frameworks
NOASSERTION

17.9K

Stars

2.9K

Forks

Jan 23, 2023

Created

Nov 3, 2025

Last Updated

Project Analytics

Stars Growth (1 Month)

+314

+1.8% change

Avg Daily Growth (1 Month)

+11.2

stars per day

Fork/Star Ratio (All Time)

16.2%

Good engagement

Lifetime Growth

15.8

stars/day over 1.1K days

Stars Over Time

Forks Over Time

Open Issues Over Time

Pull Requests Over Time

Commits Over Time

AI-Generated Tags

llm
evaluation
benchmarking
testing
api-clients
open-source

Comments (0)

Sign in to leave a comment or vote

Sign In

No comments yet. Be the first to comment!

Stay in the loop

Get weekly updates on trending AI coding tools and projects.