Thanks to visit codestin.com
Credit goes to github.com

Skip to content

[NeurIPS 2025] Flow x RL. Official Implementation of "ReinFlow: Fine-tuning Flow Policy with Online Reinforcement Learning". Fully open-sourced. Support VLAs e.g., pi0

License

ReinFlow/ReinFlow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

ReinFlow: Fine-tuning Flow Matching Policy with Online Reinforcement Learning

πŸ’ Paper accepted at NeurIPS 2025

Tonghe Zhang$^1$, Chao Yu$^{2,3}$, Sichang Su$^4$, Yu Wang$^2$

$^1$ Carnegie Mellon University $^2$ Tsinghua University $^3$ Beijing Zhongguancun Academy $^4$ University of Texas at Austin


Architecture Diagram

Shortcut Flow Can Shortcut Transport


Installation | Quick Start | Implementation Details | Add Dataset/Environment
Debug & Known Issues | License | Acknowledgement | Citation

This is the official implementation of "ReinFlow: Fine-tuning Flow Matching Policy with Online Reinforcement Learning".

If you like our work, we'll be happy if you give us a star ⭐!

πŸ”₯ ReinFlow can now scale to fine-tune 3B VLA models like $\pi_0$ via massively parallel RL.

πŸš€ About ReinFlow

ReinFlow is a flexible policy gradient framework for fine-tuning flow matching policies at any denoising step.

How does it work?
πŸ‘‰ First, train flow policies using imitation learning (behavior cloning).
πŸ‘‰ Then, fine-tune them with online reinforcement learning using ReinFlow!

🧩 Supports:

  • βœ… 1-Rectified Flow
  • βœ… Shortcut Models
  • βœ… Any other policy defined by ODEs (in principle)

πŸ“ˆ Empirical Results: ReinFlow achieves strong performance across a variety of robotic tasks:

  • 🦡 Legged Locomotion (OpenAI Gym)
  • βœ‹ State-based manipulation (Franka Kitchen)
  • πŸ‘€ Visual manipulation (Robomimic)

🧠 Key Innovation: ReinFlow trains a noise injection network end-to-end:

  • βœ… Makes policy probabilities tractable, even with very few denoising steps (e.g., 4, 2, or 1)
  • βœ… Robust to discretization and Monte Carlo approximation errors

Learn more on our πŸ”— project website or check out the arXiv paper.

πŸ“’ News

  • [2025/10/16] We scaled up ReinFlow to fine-tune large VLA like $\pi_0$. Code, hyperparamters in LIBERO environment released at RLinf.
  • [2025/09/18] Paper accepted at NeurIPS 2025.
  • [2025/08/18] All training metrics (losses, reward, etc) released in WandB to help you reproduce our results.
  • [2025/07/30] Fixed the rendering bug in Robomimic. Now supports rendering at 1080p resolution.
  • [2025/07/29] Add tutorial on how to record videos during evaluation in the docs
  • [2025/06/14] Updated webpage for a detailed explanation to the algorithm design.
  • [2025/05/28] Paper is posted on arXiv!

πŸš€ Installation

Please follow the steps in installation/reinflow-setup.md.

πŸš€ Quick Start: Reproduce Our Results

To fully reproduce our experiments, please refer to ReproduceExps.md.

To download our training data and reproduce the plots in the paper, please refer to ReproduceFigs.md.

πŸš€ Implementation Details

Please refer to Implement.md for descriptions of key hyperparameters of FQL, DPPO, and ReinFlow.

πŸš€ Adding Your Own Dataset or Environment

Please refer to Custom.md.

πŸš€ Debug Aid and Known Issues

Please refer to KnownIssues.md to see how to resolve errors you encounter.

⭐ Todo

  • Release Pi0 fine-tuning results.
  • Support fine-tuning Mean Flow with online RL
  • Release videos
  • Release WandB metrics
  • Release docs
  • Release checkpoints
  • Release codebase

License

This repository is released under the MIT license. See LICENSE. If you use our code, we appreciate it if you paste the license at the beginning of the script.

Acknowledgement

This repository was developed from multiple open-source projects. Major references include:

For more references, please refer to Acknowledgement.md.

Cite our work

@misc{zhang2025reinflowfinetuningflowmatching,
    title={ReinFlow: Fine-tuning Flow Matching Policy with Online Reinforcement Learning},
    author={Tonghe Zhang and Chao Yu and Sichang Su and Yu Wang},
    year={2025},
    eprint={2505.22094},
    archivePrefix={arXiv},
    primaryClass={cs.RO},
    url={https://arxiv.org/abs/2505.22094},
}

Star History

Star History Chart

About

[NeurIPS 2025] Flow x RL. Official Implementation of "ReinFlow: Fine-tuning Flow Policy with Online Reinforcement Learning". Fully open-sourced. Support VLAs e.g., pi0

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published