TY - GEN
T1 - Comparative Evaluation of Reinforcement Learning and Model Predictive Control for 6DoF Position Control of an Autonomous Underwater Vehicle
AU - Tunçay, Sümer
AU - Andres, Alain
AU - Carlucho, Ignacio
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
PY - 2026
Y1 - 2026
N2 - Autonomous Underwater Vehicles (AUVs) require precise and robust control strategies for 3D pose regulation in dynamic underwater environments. In this study, we present a comparative evaluation of model-free and model-based control methods for AUV position control. Specifically, we analyze the performance of neural network controllers trained by three Reinforcement Learning (RL) algorithms—Proximal Policy Optimization (PPO), Twin Delayed Deep Deterministic Policy Gradient (TD3), and Soft Actor-Critic (SAC)—alongside a Model Predictive Control (MPC) baseline. We train our RL methods in a simplified AUV simulator implemented in PyTorch, while our evaluation is done in a realistic marine robotics simulator called Stonefish. Controllers are evaluated on the basis of tracking accuracy, robustness to disturbances, and generalization capabilities. Our results show that, MPC suffers from unmodeled dynamics such as disturbances, whereas RL demonstrates adaptation capabilities to disturbances. Also, although MPC demonstrates strong control performance, it requires an accurate model, high compute power and a careful implementation to run in real-time whereas the control frequency of RL policies is only bound by the inference time of the policy network. Among RL-based controllers, PPO achieves the best overall performance, both in terms of training stability and control accuracy. This study provides insight into the feasibility of RL-based controllers for AUV position control, offering guidance for selecting suitable control strategies in real-world marine robotics applications.
AB - Autonomous Underwater Vehicles (AUVs) require precise and robust control strategies for 3D pose regulation in dynamic underwater environments. In this study, we present a comparative evaluation of model-free and model-based control methods for AUV position control. Specifically, we analyze the performance of neural network controllers trained by three Reinforcement Learning (RL) algorithms—Proximal Policy Optimization (PPO), Twin Delayed Deep Deterministic Policy Gradient (TD3), and Soft Actor-Critic (SAC)—alongside a Model Predictive Control (MPC) baseline. We train our RL methods in a simplified AUV simulator implemented in PyTorch, while our evaluation is done in a realistic marine robotics simulator called Stonefish. Controllers are evaluated on the basis of tracking accuracy, robustness to disturbances, and generalization capabilities. Our results show that, MPC suffers from unmodeled dynamics such as disturbances, whereas RL demonstrates adaptation capabilities to disturbances. Also, although MPC demonstrates strong control performance, it requires an accurate model, high compute power and a careful implementation to run in real-time whereas the control frequency of RL policies is only bound by the inference time of the policy network. Among RL-based controllers, PPO achieves the best overall performance, both in terms of training stability and control accuracy. This study provides insight into the feasibility of RL-based controllers for AUV position control, offering guidance for selecting suitable control strategies in real-world marine robotics applications.
KW - Autonomous Underwater Vehicles
KW - Optimal Control
KW - Reinforcement Learning
UR - https://www.scopus.com/pages/publications/105017374263
U2 - 10.1007/978-3-032-01486-3_29
DO - 10.1007/978-3-032-01486-3_29
M3 - Conference contribution
AN - SCOPUS:105017374263
SN - 9783032014856
T3 - Lecture Notes in Computer Science
SP - 381
EP - 394
BT - Towards Autonomous Robotic Systems - 26th Annual Conference, TAROS 2025, Proceedings
A2 - Cavalcanti, Ana
A2 - Foster, Simon
A2 - Richardson, Robert
PB - Springer Science and Business Media Deutschland GmbH
T2 - 26th Annual Conference on Towards Autonomous Robotic Systems, TAROS 2025
Y2 - 20 August 2025 through 22 August 2025
ER -