Explore Projects

Discover 4 open source projects

Active filters (1):
Search: evaluation-metricsร—
Clear all

Showing 1-4 of 4 projects

confident-ai/deepeval

A Python framework for evaluating and benchmarking large language models (LLMs) and their capabilities.

13.9K
Active
Python
LLM Frameworks
Python
#llm-evaluation#benchmarking#python-framework

AgentOps-AI/agentops

Python SDK for AI agent monitoring, LLM cost tracking, benchmarking, and more for vibe coders.

5.3K
Stable
Python
LLM Frameworks
Agents & Orchestration
Python
#ai#agents#cost-tracking

datawhalechina/tiny-universe

An open-source guide for building a 'Tiny-Universe' of large language models and AI tools.

4.6K
Stable
Jupyter Notebook
LLM Frameworks
Agents & Orchestration
Jupyter Notebook
#large-language-models#diffusion#evaluation-metrics

xinshuoweng/AB3DMOT

An open-source Python implementation for 3D multi-object tracking, with KITTI benchmarking and new evaluation metrics.

1.8K
Archived
Python
Computer Vision
API Frameworks
#3d-object-tracking#computer-vision#robotics

Stay in the loop

Get weekly updates on trending AI coding tools and projects.