Showing 1-6 of 6 projects
A framework for testing and evaluating large language models, prompts, and AI agents for security and performance.
AI observability and evaluation tooling for developers building with large language models and AI agents.
Open-source evaluation and testing library for LLM Agents
Comprehensive analytics, versioning, and ETL toolkit for multimodal data (video, audio, PDFs, images)
An open-source platform for evaluating and improving Generative AI applications with 20+ preconfigured checks and root cause analysis.
A desktop tool for orchestrating AI models across vendors using the Model Context Protocol (MCP)
Get weekly updates on trending AI coding tools and projects.