Technical Papers
Dec 1, 2021

Attitude Control of a Moving Mass–Actuated UAV Based on Deep Reinforcement Learning

Publication: Journal of Aerospace Engineering
Volume 35, Issue 2

Abstract

A moving mass–actuated unmanned aerial vehicle (MAUAV) is controlled by mass sliders installed inside the airframe and has the advantages of high aerodynamic efficiency and good stealth performance. However, designing a controller for it faces severe challenges due to the strong nonlinearity and coupling of its dynamics. To this end, we proposed an attitude controller based on deep reinforcement learning for the MAUAV. It directly maps the states to the needed deflection of the actuators and is an end-to-end controller. For the sparse reward problem, the reward function required for training is reasonably designed through reward shaping to hasten the algorithm’s training speed. In training, random initialization and parameter perturbation are used to strengthen the final policy’s robustness further. The simulation results tentatively demonstrate that the proposed controller is not only robust but suboptimal. Compared with an active disturbance rejection controller (ADRC) optimized by the particle swarm algorithm, our controller still guarantees a 100% success rate in multiple unlearned scenarios, meaning it has good generalization ability.

Get full access to this article

View all available purchase options and get full access to this article.

Data Availability Statement

Some or all data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

This study was supported by the National Natural Science Foundation of China (No. 11572097).

References

Abbeel, P., A. Coates, M. Quigley, and A. Y. Ng. 2006. “An application of reinforcement learning to aerobatic helicopter flight.” In Proc., 19th Int. Conf. on Neural Information Processing Systems, 1–8. Cambridge, MA: MIT Press.
Bou-Ammar, H., H. Voos, and W. Ertel. 2010. “Controller design for quadrotor UAVs using reinforcement learning.” In Proc., 2010 IEEE Int. Conf. on Control Applications, 2130–2135. New York: IEEE.
Calhoun, P., and E. Queen. 2002. “Entry vehicle control system design for the Mars smart lander.” In Proc., AIAA Atmospheric Flight Mechanics Conf. and Exhibit. San Diego: American Institute of Aeronautics and Astronautics.
Chen, L., G. Zhou, X. J. Yan, and D. P. Duan. 2012. “Composite control strategy of stratospheric airships with moving masses.” J. Aircr. 49 (3): 794–801. https://doi.org/10.2514/1.C031364.
Darvishpoor, S., J. Roshanian, and M. Tayefi. 2020. “A novel concept of VTOL bi-rotor UAV based on moving mass control.” Aerosp. Sci. Technol. 107 (Dec): 106238. https://doi.org/10.1016/j.ast.2020.106238.
Dogan, A., and S. A. Erturk. 2014. “Trimming mass-actuated airplane in turns with zero side slip angle.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S., O. Daskiran, and A. Dogan. 2012. “Trim analysis of a moving-mass actuated airplane.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S. A., and A. Dogan. 2013. “Trim analysis of a moving-mass actuated airplane in steady turn.” In Proc., 51st AIAA Aerospace Sciences Meeting Including the New Horizons Forum and Aerospace Exposition. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S. A., and A. Dogan. 2015. “Propeller torque effect on cruise trim of standard and mass-actuated airplane.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S. A., and A. Dogan. 2016. “Propeller torque effect on steady-state turn trim of standard and mass-actuated airplane.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S. A., and A. Dogan. 2017a. “Trim analyses of mass-actuated airplane in cruise and steady-state turn.” J. Aircr. 54 (4): 1587–1594. https://doi.org/10.2514/1.C034200.
Erturk, S. A., and A. Dogan. 2017b. “Trim analyses of mass-actuated airplane in steady-state climb and descent.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Erturk, S. A., and A. Dogan. 2017c. “Dynamic simulation and control of mass-actuated airplane.” J. Guidance Control Dyn. 40 (8): 1939–1953. https://doi.org/10.2514/1.G002658.
Furfaro, R., A. Scorsoglio, R. Linares, and M. Massari. 2020. “Adaptive generalized ZEM-ZEV feedback guidance for planetary landing via a deep reinforcement learning approach.” Acta Astronaut. 171 (Jun): 156–171. https://doi.org/10.1016/j.actaastro.2020.02.051.
Gaudet, B., R. Linares, and R. Furfaro. 2020a. “Adaptive guidance and integrated navigation with reinforcement meta-learning.” Acta Astronaut. 169 (Apr): 180–190. https://doi.org/10.1016/j.actaastro.2020.01.007.
Gaudet, B., R. Linares, and R. Furfaro. 2020b. “Deep reinforcement learning for six degree-of-freedom planetary landing.” Adv. Space Res. 65 (7): 1723–1741. https://doi.org/10.1016/j.asr.2019.12.030.
Gaudet, B., R. Linares, and R. Furfaro. 2020c. “Terminal adaptive guidance via reinforcement meta-learning: Applications to autonomous asteroid close-proximity operations.” Acta Astronaut. 171 (Jun): 1–13. https://doi.org/10.1016/j.actaastro.2020.02.036.
Gong, L., Q. Wang, C. Hu, and C. Liu. 2020. “Switching control of morphing aircraft based on Q-learning.” Chin. J. Aeronaut. 33 (2): 672–687. https://doi.org/10.1016/j.cja.2019.10.005.
Guo, T., N. Jiang, B. Li, X. Zhu, Y. Wang, and W. Du. 2021. “UAV navigation in high dynamic environments: A deep reinforcement learning approach.” Chin. J. Aeronaut. 34 (2): 479–489. https://doi.org/10.1016/j.cja.2020.05.011.
Haus, T., M. Orsag, and S. Bogdan. 2016a. “Design considerations for a large quadrotor with moving mass control.” In Proc., 2016 Int. Conf. on Unmanned Aircraft Systems, 1327–1334. New York: IEEE.
Haus, T., N. Prkut, K. Borovina, B. Maric, M. Orsag, and S. Bogdan. 2016b. “A novel concept of attitude control for large multirotor-UAVs based on moving mass control.” In Proc., 2016 24th Mediterranean Conf. on Control and Automation, 832–839. New York: IEEE.
Huang, X., J. Liu, C. Jia, Z. Wang, and J. Zhang. 2020. “Deep deterministic policy gradient for UAV control.” [In Chinese.] Acta Aeronaut. Astronaut. Sin. 42 (10): 524688. https://doi.org/10.7527/S1000-6893.2020.24688.
Li, B., and T.-C. Su. 2017. “Heading autopilot of autonomous underwater vehicles with internal moving mass.” J. Comput. Nonlinear Dyn. 12 (2): 021003. https://doi.org/10.1115/1.4034727.
Li, J., C. Gao, T. Feng, and W. Jing. 2018. “Novel moving mass flight vehicle and its equivalent experiment.” J. Dyn. Syst. Meas. Control 140 (11): 111010. https://doi.org/10.1115/1.4040326.
Li, J., C. Gao, W. Jing, and P. Wei. 2017. “Dynamic analysis and control of novel moving mass flight vehicle.” Acta Astronaut. 131 (Feb): 36–44. https://doi.org/10.1016/j.actaastro.2016.11.023.
Li, Q., S. Xie, J. Luo, P. Wu, and J. Chen. 2014. “Pitch reduction system design and control for an underwater vehicle.” In Proc., 2014 IEEE Int. Conf. on Mechatronics and Automation. New York: IEEE.
Lillicrap, T. P., J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra. 2015. “Continuous control with deep reinforcement learning.” Preprint, submitted September 09, 2015. http://arxiv.org/abs/1509.02971.
Menon, P. K., G. D. Sweriduk, E. J. Ohlmeyer, and D. S. Malyevac. 2004. “Integrated guidance and control of moving-mass actuated kinetic warheads.” J. Guidance Control Dyn. 27 (1): 118–126. https://doi.org/10.2514/1.9336.
Mnih, V., K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller. 2013. “Playing atari with deep reinforcement learning.” Preprint, submitted December 19, 2013. http://arxiv.org/abs/1312.5602.
Petsopoulos, T., F. J. Regan, and J. Barlow. 1996. “Moving-mass roll control system for fixed-trim re-entry vehicle.” J. Spacecraft Rockets 33 (1): 54–60. https://doi.org/10.2514/3.55707.
Qiu, X., M. Zhang, W. Jing, and C. Gao. 2021. “Dynamics and adaptive sliding mode control of a mass-actuated fixed-wing UAV.” Int. J. Aeronaut. Space Sci. 22 (Mar): 886–897. https://doi.org/10.1007/s42405-020-00344-w.
Robinett, R. D., B. R. Sturgis, and S. A. Kerr. 1996. “Moving mass trim control for aerospace vehicles.” J. Guidance Control Dyn. 19 (5): 1064–1070. https://doi.org/10.2514/3.21746.
Rogers, J., and M. Costello. 2008. “Control authority of a projectile equipped with a controllable internal translating mass.” J. Guidance Control Dyn. 31 (5): 1323–1333. https://doi.org/10.2514/1.33961.
Schulman, J., F. Wolski, P. Dhariwal, A. Radford, and O. Klimov. 2017. “Proximal policy optimization algorithms.” Preprint, submitted July 20, 2017. http://arxiv.org/abs/1707.06347.
Sutton, R. S., and A. G. Barto. 2018. Reinforcement learning: An introduction. Cambridge, MA: MIT Press.
Sutton, R. S., A. G. Barto, and R. J. Williams. 1992. “Reinforcement learning is direct adaptive optimal control.” IEEE Control Syst. Mag. 12 (2): 19–22. https://doi.org/10.1109/37.126844.
Tan, D., P. Yavarow, and A. Erturk. 2017. “Nonlinear structural dynamics of macro-fiber composite cantilevers for resonant actuation.” In Proc., ASME 2017 Conf. on Smart Materials, Adaptive Structures and Intelligent Systems. New York: ASME.
Vengate, S. R., S. A. Erturk, and A. Dogan. 2016. “Development and flight test of moving-mass actuated unmanned aerial vehicle.” In Proc., AIAA Atmospheric Flight Mechanics Conf. San Diego: American Institute of Aeronautics and Astronautics.
Wang, C., L. Wu, C. Yan, Z. Wang, H. Long, and C. Yu. 2020. “Coactive design of explainable agent-based task planning and deep reinforcement learning for human-UAVs teamwork.” Chin. J. Aeronaut. 33 (11): 2930–2945. https://doi.org/10.1016/j.cja.2020.05.001.
Wang, X., W. Zhou, R. Mu, and Z. Wu. 2020a. “Modeling and simulation of mass-actuated flexible aircraft for roll control.” Aerosp. Sci. Technol. 107 (Dec): 106254. https://doi.org/10.1016/j.ast.2020.106254.
Wang, X., W. Zhou, R. Mu, and Z. Wu. 2020b. “A new deformation control approach for flexible wings using moving masses.” Aerosp. Sci. Technol. 106 (Nov): 106118. https://doi.org/10.1016/j.ast.2020.106118.
Xiang, X., C. Yan, C. Wang, and D. Yi. 2021. “Towards coordination control for fixed-wing UAV formation through deep reinforcement learning.” [In Chinese.] Acta Aeronaut. Astronaut. Sin. 42 (4): 524009. https://doi.org/10.7527/S1000-6893.2020.24009.

Information & Authors

Information

Published In

Go to Journal of Aerospace Engineering
Journal of Aerospace Engineering
Volume 35Issue 2March 2022

History

Received: Jan 19, 2021
Accepted: Sep 27, 2021
Published online: Dec 1, 2021
Published in print: Mar 1, 2022
Discussion open until: May 1, 2022

Permissions

Request permissions for this article.

Authors

Affiliations

Ph.D. Candidate, School of Astronautics, Harbin Institute of Technology, Harbin, Heilongjiang 150001, China. ORCID: https://orcid.org/0000-0002-6912-316X. Email: [email protected]
Changsheng Gao [email protected]
Professor, School of Astronautics, Harbin Institute of Technology, Harbin, Heilongjiang 150001, China (corresponding author). Email: [email protected]
Ph.D. Candidate, School of Astronautics, Harbin Institute of Technology, Harbin, Heilongjiang 150001, China. Email: [email protected]
Wuxing Jing [email protected]
Professor, School of Astronautics, Harbin Institute of Technology, Harbin, Heilongjiang 150001, China. Email: [email protected]

Metrics & Citations

Metrics

Citations

Download citation

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited by

  • Trajectory Tracking Control for Fixed-Wing UAV Based on DDPG, Journal of Aerospace Engineering, 10.1061/JAEEEZ.ASENG-5286, 37, 3, (2024).
  • Pressure control of Once-through steam generator using Proximal policy optimization algorithm, Annals of Nuclear Energy, 10.1016/j.anucene.2022.109232, 175, (109232), (2022).

View Options

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Media

Figures

Other

Tables

Share

Share

Copy the content Link

Share with email

Email a colleague

Share