LOW-THRUST TRAJECTORY DESIGN USING STATE-DEPENDENT CLOSED-LOOP CONTROL LAWS AND REINFORCEMENT LEARNING

Harry Holt, Roberto Armellin, Nicola Baresi, Andrea Scorsoglio, Roberto Furfaro

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Closed-loop feedback-driven control laws can be used to solve low-thrust many-revolution trajectory design and guidance problems with minimal computational cost. They treat the problem from a targeting perspective and hence value stability over optimality. The optimality can be increased by making the parameters state-dependent at the cost of reduced stability. In this paper, an actor-critic reinforcement learning framework is used to make the parameters of the Lyapunov-based Q-law state-dependent. A single-layer neural network ensures the Jacobian of these state-dependent parameters can be calculated and used to enforce stability throughout the transfer. The current results focus on GTO-GEO and LEO-GEO transfers in Keplerian dynamics. A trade-off between optimality and stability is observed for the first, but the added stability increases optimality for the later. Robustness to uncertainties in position and velocity are also investigated, along with the effects of eclipses and dynamical perturbations such as J2, Sun and Moon third body attractions.

Original languageEnglish (US)
Title of host publicationASTRODYNAMICS 2020
EditorsRoby S. Wilson, Jinjun Shan, Kathleen C. Howell, Felix R. Hoots
PublisherUnivelt Inc.
Pages131-149
Number of pages19
ISBN (Print)9780877036753
StatePublished - 2021
Externally publishedYes
EventAAS/AIAA Astrodynamics Specialist Conference, 2020 - Virtual, Online
Duration: Aug 9 2020Aug 12 2020

Publication series

NameAdvances in the Astronautical Sciences
Volume175
ISSN (Print)0065-3438

Conference

ConferenceAAS/AIAA Astrodynamics Specialist Conference, 2020
CityVirtual, Online
Period8/9/208/12/20

ASJC Scopus subject areas

  • Aerospace Engineering
  • Space and Planetary Science

Fingerprint

Dive into the research topics of 'LOW-THRUST TRAJECTORY DESIGN USING STATE-DEPENDENT CLOSED-LOOP CONTROL LAWS AND REINFORCEMENT LEARNING'. Together they form a unique fingerprint.

Cite this