Showing 101-120 of 222 projects
A curated collection of must-read papers and blogs on large language model-based long-context modeling.
A benchmarking tool for evaluating the performance of popular machine learning algorithms and libraries.
Petastorm enables training and evaluation of deep learning models from Apache Parquet datasets.
A Python library for computing BERT-based text generation evaluation metrics.
A platform for designing and evaluating Graph Neural Networks (GNNs) using Python.
A lightweight C library for parsing, compiling, and evaluating mathematical expressions.
Comprehensive resources for developers working with Generative AI, including projects, use cases, and interview prep.
A Python library that helps prevent CUDA out of memory errors in PyTorch models with just 1 line of code.
An open-source Python implementation for 3D multi-object tracking, with KITTI benchmarking and new evaluation metrics.
Official repository for C-Eval, a Chinese evaluation suite for foundation models.
CodeXGLUE is a comprehensive benchmark for evaluating the performance of large language models on a variety of coding-related tasks.
A framework for evaluating large language models, focused on AI and machine learning tools.
An AI-powered security toolkit for LLM vulnerability scanning and red teaming.
A feedback management tool for education, built with Java and Angular, used by universities and teachers.
A benchmark for evaluating language understanding models and datasets for the Chinese language.
A security audit tool to assess and improve cybersecurity posture.
A plugin that tracks occurrences and numbers in your Obsidian notes, providing charts and quantified self-tracking.
Reference implementations of MLPerfยฎ training benchmarks for evaluating machine learning performance.
VPN/proxy node recommendation and evaluation guide for airport services
A graphical Python debugger that lets you easily view the values of all evaluated expressions.
Get weekly updates on trending AI coding tools and projects.