PKU-Alignment/safe-rlhf

A safe reinforcement learning from human feedback (RLHF) system for aligning large language models with human values.

Python
AI & Machine Learning
LLM Frameworks
Apache-2.0

1.6K

Stars

130

Forks

May 15, 2023

Created

Nov 24, 2025

Last Updated

Project Analytics

Stars Growth (1 Month)

+7

+0.4% change

Avg Daily Growth (1 Month)

+0.3

stars per day

Fork/Star Ratio (All Time)

8.2%

Normal engagement

Lifetime Growth

1.5

stars/day over 1.0K days

Stars Over Time

Forks Over Time

Open Issues Over Time

Pull Requests Over Time

Commits Over Time

AI-Generated Tags

ai-safety
large-language-models
reinforcement-learning
reinforcement-learning-from-human-feedback
safe-reinforcement-learning

Comments (0)

Sign in to leave a comment or vote

Sign In

No comments yet. Be the first to comment!

Stay in the loop

Get weekly updates on trending AI coding tools and projects.