Abstract
This paper describes a Markov decision process approach to a robust spacecraft mission control policy that maximizes the expected value of science reward assuming a circular orbit. The control policy that governs mission steps can be computed off-board or onboard depending upon the availability of communication bandwidth and on-board computational resources. This paper considers a sample science mission, where the spacecraft collects data from celestial objects viewable only within a certain orbit true anomaly window. Science data collection requires the spacecraft to slew its instrument(s) toward each target, and continue pointing in the direction of the target while the spacecraft traverses its orbit. Robustness and stochastic optimization of scientific reward, is achieved at the cost of computational complexity. Approximate dynamic programming (ADP) is exploited to reduce the computational time and effort to manageable levels and to treat larger problem sizes. The proposed ADP algorithm partitions the state-space based on true anomaly regions, enabling grouping of adjacent science targets. Results of a simulation case study demonstrate that our proposed ADP approach performs quite well for reasonable ranges of key problem parameters.
| Original language | English |
|---|---|
| Article number | 8101308 |
| Pages (from-to) | 923-934 |
| Number of pages | 12 |
| Journal | IEEE Transactions on Systems, Man, and Cybernetics: Systems |
| Volume | 50 |
| Issue number | 3 |
| DOIs | |
| State | Published - Mar 2020 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2017 IEEE.
Keywords
- Contingency response
- Markov decision process (MDP)
- science-optimal control
- spacecraft mission control
ASJC Scopus subject areas
- Software
- Control and Systems Engineering
- Human-Computer Interaction
- Computer Science Applications
- Electrical and Electronic Engineering