Skip to content

Releases: rllm-org/rllm

rLLM: v0.2.1.post1

18 Dec 23:51
618fa7d

Choose a tag to compare

What's Changed

  • update docs & add curves by @thwu1 in #343
  • Fix import for colorful_print in agent_sdk_engine.py and agent_sdk_trainer.py by @wht0703 in #345
  • Unblock sdk installation by overriding dependices by @wht0703 in #348
  • [Doc] Update README and fix a few installation related issues by @listar2000 in #347
  • fix: keyerror completion_ids by @kxfan2002 in #353
  • Fix: Enable GPU acceleration for dense retrieval in search agent by @Gitsamshi in #349

New Contributors

Full Changelog: v0.2.1...v0.2.1.post1

rLLM: v0.2.1

11 Dec 22:58
960d573

Choose a tag to compare

rLLM v0.2.1: Tinker backend, VLM training, Eval Protocol, and SDK (preview)

We are excited to release rLLM v0.2.1. This new version comes with the following exciting features:

  • rLLM SDK (preview): The rLLM SDK enables you to transform agents written in frameworks such as LangGraph, SmolAgent, or Strands into trainable workflows. Check out this LangGraph RAG example, which builds a RAG agent and trains it with the rLLM SDK.

  • Tinker training backend: In addition to verl, rLLM now supports Tinker as a training backend. You can use the same abstractions for building agents and easily switch between different backends for training.

  • VLM training: rLLM supports Vision-Language Model training with the verl backend. See the Geo3K training example for reference.

  • LoRA fine-tuning: rLLM supports LoRA training in both the verl and Tinker backends. See the GSM8K LoRA example for how to enable LoRA training with a single config change.

  • Eval Protocol Integration We integrate with the Eval Protocol from Fireworks AI. Users can now train on any environments supported by the Eval Protocol. See this example that uses Eval Protocol in rLLM to train a Frozenlake agent.

A big shoutout to @thwu1 @kylemontgomery1 @listar2000 @xzrderek for their outstanding work on these features.

What's Changed

New Contributors

Full Changelog: v0.2.0...v0.2.1

rLLM: v0.2.0

16 Oct 21:24
52efedc

Choose a tag to compare

rLLM v0.2: RL Training over General Agentic Programs (Blog Post)

We are excited to release rLLM v0.2, a major upgrade of our RL training framework. In v0.1, rLLM provided agent and OpenAI Gym-like environment abstractions to support training ReACT-style agents. In v0.2, we additionally introduce AgentWorkflowEngine and AgentWorkflowTrainer—more general abstractions that enable arbitrary agentic programs to be trained. Agent builders and researchers can now define multi-agent systems, complex workflows (e.g., solver-judge, planner executor, MCTS), and agentic programs with custom reward functions, and train them with reinforcement learning without rewriting their production code.

Key Features in v0.2

  1. Support the official verl==0.5.0 as training backend, no custom verl fork anymore! verl==0.5.0 comes with support of the following features which are now supported in rLLM (@kylemontgomery1):
    • Megatron training support (@jeewoo-lee)
    • SGLang as the rollout engine, in addition to vLLM.
  2. Introduce AgentWorkflowEngine, which enables passing in arbitrary agentic programs for training. (@kylemontgomery1)
  3. Support more agents and environments
  4. Integration with other agentic framework/SDK
    • Strands SDK from AWS
    • SmolAgents

What's Changed

New Contributors

Full Changelog: https://github.com/rllm-org/rllm/commits/v0.2.0