Abstract
Closed-loop feedback-driven control laws can be used to solve low-thrust many-revolution trajectory design and guidance problems with minimal computational cost. Lyapunov-based control laws offer the benefits of increased stability whilst their optimality can be increased by tuning their parameters. In this paper, a reinforcement learning framework is used to make the parameters of the Lyapunov-based Q-law state-dependent, increasing its optimality. The Jacobian of these state-dependent parameters is available analytically and, unlike in other optimisation approaches, can be used to enforce stability throughout the transfer. The results focus on GTO–GEO and LEO–GEO transfers in Keplerian dynamics, including the effects of eclipses. The impact of the network architecture on the behaviour is investigated for both time- and mass-optimal transfers. Robustness to navigation errors and thruster misalignment is demonstrated using Monte Carlo analyses. The resulting approach offers potential for on-board autonomous transfers and orbit reconfiguration.
Original language | English (US) |
---|---|
Pages (from-to) | 511-528 |
Number of pages | 18 |
Journal | Acta Astronautica |
Volume | 187 |
DOIs | |
State | Published - Oct 2021 |
Keywords
- Low-thrust
- Lyapunov control
- Reinforcement learning
- Stability
- State-dependent
ASJC Scopus subject areas
- Aerospace Engineering