rLLM v0.2 release #251
jeffreysijuntan
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
rLLM v0.2 Release (Blog Post)
We are excited to release rLLM v0.2, a major upgrade of our RL training framework. In v0.1, rLLM provided agent and OpenAI Gym-like environment abstractions to support training ReACT-style agents. In v0.2, we additionally introduce
AgentWorkflowEngineandAgentWorkflowTrainer—more general abstractions that enable arbitrary agentic programs to be trained. Agent builders and researchers can now define multi-agent systems, complex workflows (e.g., solver-judge, planner executor, MCTS), and agentic programs with custom reward functions, and train them with reinforcement learning without rewriting their production code.Key Features in v0.2
verl==0.5.0as training backend, no custom verl fork anymore!verl==0.5.0comes with support of the following features which are now supported in rLLM (@kylemontgomery1):AgentWorkflowEngine, which enables passing in arbitrary agentic programs for training. (@kylemontgomery1)What's Changed
New Contributors
Full Changelog: https://github.com/rllm-org/rllm/commits/v0.2.0
This discussion was created from the release rLLM v0.2 release.
Beta Was this translation helpful? Give feedback.
All reactions