Explore Projects

Discover 6 open source projects

Active filters (1):
Search: llm-evalร—
Clear all

Showing 1-6 of 6 projects

promptfoo/promptfoo

A framework for testing and evaluating large language models, prompts, and AI agents for security and performance.

10.8K
Active
TypeScript
LLM Frameworks
TypeScript
#llm-evaluation#prompt-engineering#red-teaming

Arize-ai/phoenix

AI observability and evaluation tooling for developers building with large language models and AI agents.

8.8K
Active
Jupyter Notebook
LLM Frameworks
Agents & Orchestration
Jupyter Notebook
#ai-monitoring#ai-observability#llm-evaluation

Giskard-AI/giskard-oss

Open-source evaluation and testing library for LLM Agents

5.1K
Active
Python
LLM Frameworks
React
#evaluation#testing#LLM

datachain-ai/datachain

Comprehensive analytics, versioning, and ETL toolkit for multimodal data (video, audio, PDFs, images)

2.7K
Active
Python
Computer Vision
ETL & Pipelines
Python
#data-analytics#data-wrangling#embeddings

uptrain-ai/uptrain

An open-source platform for evaluating and improving Generative AI applications with 20+ preconfigured checks and root cause analysis.

2.3K
Archived
Python
LLM Frameworks
Testing
Python
#llm-eval#prompt-engineering#root-cause-analysis

AI-QL/tuui

A desktop tool for orchestrating AI models across vendors using the Model Context Protocol (MCP)

1.1K
Active
TypeScript
MCP Frameworks
Agents & Orchestration
React
#ai-integration#mcp#llm-orchestration

Stay in the loop

Get weekly updates on trending AI coding tools and projects.