Showing 1-4 of 4 projects
A curated list of responsible machine learning resources for interpretable AI development.
A safe reinforcement learning from human feedback (RLHF) system for aligning large language models with human values.
A Python library for exploring secrets of RLHF (Reward-Weighted Maximum Likelihood Estimation) in large language models
A Python package for uncertainty quantification and hallucination detection in large language models (LLMs)
Get weekly updates on trending AI coding tools and projects.