Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
PythonPKU-Alignment/safe-rlhf

safe-rlhf

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback

53.2/100
1.6KForks: 130
View on GitHubHomepage →
Loading report...

Similar Projects

unsloth

93

Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek, Qwen, Llama, Gemma, TTS 2x faster with 70% less VRAM.

Python53.5K

litgpt

89

20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.

Python13.2K

OpenRLHF

89

An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)

Python9.1K

lmdeploy

85

LMDeploy is a toolkit for compressing, deploying, and serving LLMs.

Python7.7K
Back to List