anthropics/hh-rlhf

Repository containing human preference data for training a helpful and harmless AI assistant.

AI & Machine Learning
LLM Frameworks
MIT

1.8K

Stars

152

Forks

Apr 10, 2022

Created

Jun 17, 2025

Last Updated

Project Analytics

Stars Growth (1 Month)

+8

+0.4% change

Avg Daily Growth (1 Month)

+0.3

stars per day

Fork/Star Ratio (All Time)

8.3%

Normal engagement

Lifetime Growth

1.3

stars/day over 1.4K days

Stars Over Time

Forks Over Time

Open Issues Over Time

Pull Requests Over Time

Commits Over Time

AI-Generated Tags

language-model
reinforcement-learning
human-feedback
safety
assistant

Comments (0)

Sign in to leave a comment or vote

Sign In

No comments yet. Be the first to comment!

Stay in the loop

Get weekly updates on trending AI coding tools and projects.