About OpenRLHF
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & VLM & TIS & vLLM & Ray & Async RL)
Topics
large-language-modelsproximal-policy-optimizationraylibreinforcement-learningreinforcement-learning-from-human-feedbacktransformersvisual-language-modelsvllm
No rules target OpenRLHF yet
No published rules, MCP servers, or skills target OpenRLHF yet. If you maintain a tool that works well with this project, you can publish for free during beta.
Why this page exists
RuleSell tracks the AI-coding ecosystem so you don't have to. When a repo like OpenRLHF picks up momentum, we surface the Claude Code skills, Cursor rules, MCP servers, and agent configs that target it — with real author attribution, SPDX license badges, and quality scores. Every listing ships with copy-paste install for each environment.