Low-Trust Trajectory Design from LEO to GEO Using Reinforcement Learning
In this paper, In the preliminary phase of space mission design, the selection of the spacecraft's trajectory is critical. This study simulates the dynamics of low-thrust orbital transfers within a two-dimensional orbital plane, employing a set of ordinary differential equations to represent the continuum of the spacecraft's orbital elements. These elements are encapsulated by six orbital parameters, manipulated under a defined thrust vector strategy within the action space, adhering to a specified policy framework. An agent, trained via a reinforcement learning algorithm within an actor-critic network architecture, is tasked with executing a low-thrust transfer between Low Earth Orbit (LEO) and Geostationary Orbit (GEO). The algorithm dynamically adjusts the spacecraft's trajectory, informed by initial orbital conditions and mission-specific constraints, to derive an optimal thrust angle trajectory and corresponding adjustments in orbital elements for the maneuver. To validate the algorithm's efficacy and robustness, a comparative analysis is conducted by implementing an alternative transfer mode at a varied orbital altitude. Additionally, the study explores the impact of adjusting the algorithm's degradation coefficient hyperparameter on the learning efficacy. Conclusively, the findings suggest that the agent, once adequately trained within the specified dynamical model, is capable of autonomously executing analogous orbital transfers. This is achieved without necessitating reiteration of the dynamical simulations, contingent solely upon the stipulation of initial and terminal orbital parameters.