Explore Projects

Discover 1 open source projects

Active filters (1):
Search: safe-rlhfร—
Clear all

Showing 1-1 of 1 projects

PKU-Alignment/safe-rlhf

A safe reinforcement learning from human feedback (RLHF) system for aligning large language models with human values.

1.6K
Stable
Python
LLM Frameworks
Reinforcement Learning
#ai-safety#large-language-models#reinforcement-learning

Stay in the loop

Get weekly updates on trending AI coding tools and projects.