Showing 1-2 of 2 projects
A safe reinforcement learning from human feedback (RLHF) system for aligning large language models with human values.
OmniSafe is an infrastructural framework for accelerating safe reinforcement learning research.
Get weekly updates on trending AI coding tools and projects.