Welcome to OpenRLHF’s documentation!

OpenRLHF is the first easy-to-use, high-performance RLHF framework built on Ray, vLLM, ZeRO-3 and HuggingFace Transformers, designed to make RLHF training simple and accessible:

  • Distributed Architecture with Ray OpenRLHF leverages Ray for efficient distributed scheduling. It separates the Actor, Reward, Reference, and Critic models across different GPUs, enabling scalable training for models up to 70B parameters. It also supports Hybrid Engine scheduling, allowing all models and vLLM engines to share GPU resources—minimizing idle time and maximizing GPU utilization.

  • vLLM Inference Acceleration + AutoTP RLHF training spends 80% of the time on the sample generation stage. Powered by vLLM and Auto Tensor Parallelism (AutoTP), OpenRLHF delivers high-throughput, memory-efficient samples generation. Native integration with HuggingFace Transformers ensures seamless and fast generation, making it the fastest RLHF framework available.

  • Memory-Efficient Training with ZeRO-3 Built on DeepSpeed’s ZeRO-3 and deepcompile, OpenRLHF enables large model training without heavyweight frameworks. It works directly with HuggingFace for easy loading and fine-tuning of pretrained models.

  • Optimized PPO Implementation Incorporates advanced PPO tricks inspired by practical guides and community best practices, enhancing training stability and reward quality in RLHF workflows. Referencing Zhihu and Advanced Tricks for Training Large Language Models with Proximal Policy Optimization.

For more technical details, see our technical report and slides.

Check out the Quick Start section for further information, including how to Installation the project.

Note

This project is under active development.

Contents