There is enormous potential for reinforcement learning and other data-driven control paradigms for controlling large scale fluid flows. But RL research on such systems is often hindered by a complex and brittle software pipeline consisting of external solvers and multiple code bases, making this exciting field inaccessible for many RL researchers.
To tackle this challenge, we (primarily Jannis from TU Dortmund) have developed a standalone, fully differentiable, plug-and-play benchmark for RL in active flow control, implemented in a single PyTorch codebase via PICT, without external solver dependencies.
Our FluidGym comes with a collection of standardized environment configurations spanning diverse 3D and multi-agent control tasks. We perform an extensive experimental study with multiple seeds, randomized initial conditions and separate train/validate/test sets. We compare the default implementations of the two most popular algorithms PPO and SAC in the single and multi-agent settings, and also investigate the potential for transfer learning.
We hope that this may be of interest to a large number of reinforcement learning researchers who are keen on assessing the most recent trends in basic RL research on a new set of challenging tasks, but otherwise find it difficult to enter the field of fluid mechanics.
Repository: https://github.com/safe-autonomous-systems/fluidgym
Paper: https://arxiv.org/abs/2601.15015

