TY - JOUR
T1 - Improving reinforcement learning performance in spacecraft guidance and control through meta-learning
T2 - a comparison on planetary landing
AU - Federici, Lorenzo
AU - Furfaro, Roberto
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
PY - 2024
Y1 - 2024
N2 - This paper investigates the performance and computational complexity of recurrent neural networks (RNNs) trained via meta-reinforcement learning (meta-RL) as onboard spacecraft guidance and control systems. The paper first presents the theoretical background behind meta-RL with RNNs, highlighting the features that make it suitable for real-world spacecraft guidance and control applications. A thorough comparison of meta-RL with a standard RL approach that uses fully connected neural networks is carried out on a benchmark problem related to spacecraft guidance and control, namely a pin-point planetary landing. The focus is on evaluating the optimality of the control policy, the ability to handle constraints, and the robustness of the approach to different kinds and levels of uncertainties, such as unmodeled dynamics, navigation uncertainties, control errors, and engine failures, to highlight the superiority of meta-RL in both nominal and off-nominal operating conditions.
AB - This paper investigates the performance and computational complexity of recurrent neural networks (RNNs) trained via meta-reinforcement learning (meta-RL) as onboard spacecraft guidance and control systems. The paper first presents the theoretical background behind meta-RL with RNNs, highlighting the features that make it suitable for real-world spacecraft guidance and control applications. A thorough comparison of meta-RL with a standard RL approach that uses fully connected neural networks is carried out on a benchmark problem related to spacecraft guidance and control, namely a pin-point planetary landing. The focus is on evaluating the optimality of the control policy, the ability to handle constraints, and the robustness of the approach to different kinds and levels of uncertainties, such as unmodeled dynamics, navigation uncertainties, control errors, and engine failures, to highlight the superiority of meta-RL in both nominal and off-nominal operating conditions.
KW - Meta-reinforcement learning
KW - Planetary landing
KW - Recurrent neural network
KW - Spacecraft guidance and control
UR - http://www.scopus.com/inward/record.url?scp=85212521055&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85212521055&partnerID=8YFLogxK
U2 - 10.1007/s00521-024-10520-8
DO - 10.1007/s00521-024-10520-8
M3 - Article
AN - SCOPUS:85212521055
SN - 0941-0643
JO - Neural Computing and Applications
JF - Neural Computing and Applications
M1 - 105746
ER -