Showing 161-180 of 222 projects
A JavaScript library for parsing and evaluating mathematical expressions.
An open-source Java toolkit for building, evaluating, and deploying sophisticated AI agents.
A dataset of mathematical reasoning problems for evaluating AI systems.
A Lisp editing library for Emacs focused on providing a streamlined and efficient coding experience.
An open-source toolkit for general OCR research and applications, with integrated training, evaluation, and production-ready OCR systems.
This is a set of model analysis tools for TensorFlow, enabling developers to evaluate and optimize their machine learning models.
Guidance for training an AWS DeepRacer model using Amazon SageMaker, providing developers full control over the process.
An autonomous web application evaluation agent powered by MCP and Playwright for vibe coders.
A powerful tool for automated LLM fuzzing to help developers and security researchers identify and mitigate potential jailbreaks.
A collection of novel jailbreak methods for large language models (LLMs) focused on privacy and safety.
An explainability toolbox for developers building machine learning models with interpretability and fairness in mind.
Aesara is a Python library for defining, optimizing, and efficiently evaluating mathematical expressions involving multi-dimensional arrays.
Train, evaluate, optimize, and deploy computer vision models with OpenVINO, a toolkit for accelerating deep learning on edge devices.
A collection of resources for quantitative finance, including factor-based stock quantitative framework and AI-finance related materials.
A Python library for fine-tuning and evaluating large language models with visual prompts.
bloom is a Python library for evaluating any behavior immediately, focused on AI safety research.
This is an MCU framework for the STM32F4 series, providing a complete set of drivers, middleware, and sample projects.
A Python library for evaluating multi-object tracking algorithms using HOTA and other metrics.
A toolbox for quantitative trajectory evaluation of visual odometry and visual-inertial odometry algorithms.
FuzzBench is a benchmarking framework for evaluating fuzzer performance and security.
Get weekly updates on trending AI coding tools and projects.