TY - JOUR
T1 - Physics-guided actor-critic reinforcement learning for swimming in turbulence
AU - Koh, Christopher
AU - Pagnier, Laurent
AU - Chertkov, Michael
N1 - Publisher Copyright:
© 2025 authors.
PY - 2025/1
Y1 - 2025/1
N2 - Turbulent diffusion causes particles placed in proximity to separate. We investigate the required swimming efforts to maintain an active particle close to its passively advected counterpart. We explore optimally balancing these efforts by developing a novel physics-informed reinforcement learning strategy and comparing it with prescribed control and physics-agnostic reinforcement learning strategies. Our scheme, coined the actor-physicist, is an adaptation of the actor-critic algorithm in which the neural network parameterized critic is replaced with an analytically derived physical heuristic function, the physicist. We validate the proposed physics-informed reinforcement learning approach through extensive numerical experiments in both synthetic Batchelor-Kraichnan and more realistic Arnold-Beltrami-Childress flow environments, demonstrating its superiority in controlling particle dynamics when compared to standard reinforcement learning methods.
AB - Turbulent diffusion causes particles placed in proximity to separate. We investigate the required swimming efforts to maintain an active particle close to its passively advected counterpart. We explore optimally balancing these efforts by developing a novel physics-informed reinforcement learning strategy and comparing it with prescribed control and physics-agnostic reinforcement learning strategies. Our scheme, coined the actor-physicist, is an adaptation of the actor-critic algorithm in which the neural network parameterized critic is replaced with an analytically derived physical heuristic function, the physicist. We validate the proposed physics-informed reinforcement learning approach through extensive numerical experiments in both synthetic Batchelor-Kraichnan and more realistic Arnold-Beltrami-Childress flow environments, demonstrating its superiority in controlling particle dynamics when compared to standard reinforcement learning methods.
UR - http://www.scopus.com/inward/record.url?scp=85217082004&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85217082004&partnerID=8YFLogxK
U2 - 10.1103/PhysRevResearch.7.013121
DO - 10.1103/PhysRevResearch.7.013121
M3 - Article
AN - SCOPUS:85217082004
SN - 2643-1564
VL - 7
JO - Physical Review Research
JF - Physical Review Research
IS - 1
M1 - 013121
ER -