Technical Papers
Jun 2, 2022

An Integrated Tracking Control Approach Based on Reinforcement Learning for a Continuum Robot in Space Capture Missions

Publication: Journal of Aerospace Engineering
Volume 35, Issue 5

Abstract

In this paper, an integrated tracking control approach was developed for a continuum robot in space capture missions. For the configuration of a three-module cable-driven continuum robot, the nonlinear dynamics equations were derived. The uncertain movement of noncooperative debris requires a real-time trajectory planning solution. Therefore, an adaptive controller based on deep reinforcement learning (DRL) is proposed to generate a dynamic controller in continuous action space, where the trajectory planning function is simultaneously integrated into the dynamic solution. To obtain an efficient policy network for the highly nonlinear dynamics model, the rolling optimization method was combined in the DRL method of the deep deterministic policy gradient (DDPG). The DRL controller generated an appropriate control sequence according to the long-term control performance of the robot system and then executed optimal control input according to the rolling optimization. The simulation result shows that the proposed policy network of the improved DDPG controller can reasonably provide the tracking control solution in the noncooperative debris capture mission.

Get full access to this article

View all available purchase options and get full access to this article.

Data Availability Statement

All data, models, and code generated or used during the study appear in the published article.

Acknowledgments

This work was supported by the National Natural Science Foundation Key Foundation No. 91748203 and National Outstanding Youth Science Fund Project of the National Natural Science Foundation of China No. 11922203.

References

Böttcher, G., S. Lilge, and J. Burgner-Kahrs. 2021. “Design of a reconfigurable parallel continuum robot with tendon-actuated kinematic chains.” IEEE Rob. Autom. Lett. 6 (2): 1272–1279. https://doi.org/10.1109/LRA.2021.3057557.
Burgner-Kahrs, J., D. C. Rucker, and H. Choset. 2015. “Continuum robots for medical applications: A survey.” IEEE Trans. Rob. 31 (6): 1261–1280. https://doi.org/10.1109/TRO.2015.2489500.
Chu, M., and X. Wu. 2018. “Modeling and self-learning soft-grasp control for free-floating space manipulator during target capturing using variable stiffness method.” IEEE Access 6: 7044–7054. https://doi.org/10.1109/ACCESS.2017.2788400.
Eren, U., A. Prach, B. B. Koçer, S. V. Raković, E. Kayacan, and B. Açıkmeşe. 2017. “Model predictive control in aerospace systems: Current state and opportunities.” J. Guid. Control Dyn. 40 (7): 1541–1566. https://doi.org/10.2514/1.G002507.
Fabio, M., T. Felix, G. Michael, and P. Jan. 2018. “Domain randomization for simulation-based policy optimization with transferability assessment.” In Proc., Conf. on Robot Learning, 700–713. Seattle: Amazon.
Fazel, M., R. Ge, S. M. Kakade, and M. Mesbahi. 2018. “Global convergence of policy gradient methods for the linear quadratic regulator.” Preprint, submitted January 15, 2018. http://arxiv.org/abs/1801.05039.
Flores-Abad, A., M. A. García Terán, I. U. Ponce, and M. Nandayapa. 2021. “Compliant force sensor-less capture of an object in orbit.” IEEE Trans. Aerosp. Electron. Syst. 57 (1): 497–505. https://doi.org/10.1109/TAES.2020.3027108.
Greer, J., T. K. Morimoto, A. M. Okamura, and E. W. Hawkes. 2017. “Series pneumatic artificial muscles (sPAMs) and application to a soft continuum robot.” In Proc., 2017 IEEE Int. Conf. on Robotics and Automation (ICRA), 5503–5510. New York: IEEE. https://doi.org/10.1109/ICRA.2017.7989648.
Gu, S., E. Holly, T. Lillicrap, and S. Levine. 2017. “Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates.” In Proc., 2017 IEEE Int. Conf. on Robotics and Automation (ICRA), 3389–3396. New York: IEEE. https://doi.org/10.1109/ICRA.2017.798938.
Hartmann, G., Z. Shiller, and A. Azaria. 2019. “Deep reinforcement learning for time optimal velocity control using prior knowledge.” In Proc., 2019 IEEE 31st Int. Conf. on Tools with Artificial Intelligence (ICTAI), 186–193. New York: IEEE. https://doi.org/10.1109/ICTAI.2019.00034.
Heidari, H., M. Pouria, S. Sharifi, and M. Karami. 2018. “Design and fabrication of robotic gripper for grasping in minimizing contact force.” Adv. Space Res. 61 (5): 1359–1370. https://doi.org/10.1016/j.asr.2017.12.024.
Huang, P., D. Wang, Z. Meng, F. Zhang, and Z. Liu. 2016. “Impact dynamic modeling and adaptive target capturing control for tethered space robots with uncertainties.” IEEE/ASME Trans. Mechatron. 21 (5): 2260–2271. https://doi.org/10.1109/TMECH.2016.2569466.
Jiang, D., Z. Cai, H. Peng, and Z. Wu. 2021. “Coordinated control based on reinforcement learning for dual-arm continuum manipulators in space capture missions.” J. Aerosp. Eng. 34 (6): 04021087. https://doi.org/10.1061/(ASCE)AS.1943-5525.0001335.
Kahn, G., A. Villaflor, B. Ding, P. Abbeel, and S. Levine. 2018. “Self-supervised deep reinforcement learning with generalized computation graphs for robot navigation.” In Proc., 2018 IEEE Int. Conf. on Robotics and Automation (ICRA), 5129–5136. New York: IEEE. https://doi.org/10.1109/ICRA.2018.8460655.
Kupcsik, A., M. P. Deosemroth, J. Peter, A. P. Loh, and G. Neumann. 2017. “Model-based contextual policy search for data-efficient generalization of robot skills.” Artif. Intell. 247 (Jun): 415–439. https://doi.org/10.1016/j.artint.2014.11.005.
Li, H., W. Yan, and Y. Shi. 2017. “Continuous-time model predictive control of under-actuated spacecraft with bounded control torques.” Automatica 75 (Jan): 144–153. https://doi.org/10.1016/j.automatica.2016.09.024.
Lillicrap, T. P., J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra. 2016. “Continuous control with deep reinforcement learning.” In Proc., Int. Conf. on Learning Representations. Menlo Park, CA: Facebook.
Mu, Z., T. Liu, W. Xu, Y. Lou, and B. Liang. 2019. “Dynamic feedforward control of spatial cable-driven hyper-redundant manipulators for on-orbit servicing.” Robotica 37 (1): 18–38. https://doi.org/10.1017/S026357471800084X.
Mu, Z., H. Yuan, W. Xu, T. Liu, and B. Liang. 2020. “A segmented geometry method for kinematics and configuration planning of spatial hyper-redundant manipulators.” IEEE Trans. Syst. Man Cybern. Syst. 50 (5): 1746–1756. https://doi.org/10.1109/TSMC.2017.2784828.
Nagabandi, A., G. Kahn, R. S. Fearing, and S. Levine. 2018. “Neural network dynamics for model-based deep reinforcement learning with model-free fine-tuning.” In Proc., 2018 IEEE International Conference on Robotics and Automation (ICRA), 7559–7566. New York: IEEE. https://doi.org/10.1109/ICRA.2018.8463189.
Nishida, S., and K. Satomi. 2011. “Strategy for capturing of a tumbling space debris.” Acta Astronaut. 68 (1–2): 113–120. https://doi.org/10.1016/j.actaastro.2010.06.045.
Peng, H., F. Li, J. Li, and Z. Ju. 2020a. “A symplectic instantaneous optimal control for robot trajectory tracking with differential-algebraic equation models.” IEEE Trans. Ind. Electron. 67 (5): 3819–3829. https://doi.org/10.1109/TIE.2019.2916390.
Peng, J., W. Xu, T. Liu, H. Yuan, and B. Liang. 2021. “End-effector pose and arm-shape synchronous planning methods of a hyper-redundant manipulator for spacecraft repairing.” Mech. Mach. Theory 155 (Jan): 104062. https://doi.org/10.1016/j.mechmachtheory.2020.104062.
Peng, J., W. Xu, T. Yang, Z. Hu, and B. Liang. 2020b. “Dynamic modeling and trajectory tracking control method of segmented linkage cable-driven hyper-redundant robot.” Nonlinear Dyn. 101 (1): 233–253. https://doi.org/10.1007/s11071-020-05764-7.
Robinson, G., and J. Davies. 1999. “Continuum robots: A state of the art.” In Vol. 4 of Proc., 1999 IEEE Int. Conf. on Robotics and Automation, 2849–2854. New York: IEEE. https://doi.org/10.1109/ROBOT.1999.774029.
Rone, W., and P. Ben-Tzvi. 2014. “Continuum robot dynamics utilizing the principle of virtual power.” IEEE Trans. Rob. 30 (1): 275–287. https://doi.org/10.1109/TRO.2013.2281564.
Sangeetha, G., K. Nishank, P. R. Hari, and S. Sasikumar. 2018. “Implementation of a stereo vision based system for visual feedback control of robotic arm for space manipulations.” Procedia Comput. Sci. 133: 1066–1073. https://doi.org/10.1016/j.procs.2018.07.031.
Seleem, I., H. El-Hussieny, S. F. M. Assal, and H. Ishii. 2020. “Development and stability analysis of an imitation learning-based pose planning approach for multi-section continuum robot.” IEEE Access 8: 99366–99379. https://doi.org/10.1109/ACCESS.2020.2997636.
Shah, P., M. Fiser, A. Faust, J. Kew, and D. Hakkani-Tur. 2018. “FollowNet: Robot navigation by following natural language directions with deep reinforcement learning.” Preprint, submitted May 16, 2018. http://arxiv.org/abs/1805.06150.
Shahid, A., L. Roveda, D. Piga, and F. Braghin. 2020. “Learning continuous control actions for robotic grasping with reinforcement learning.” In Proc., 2020 IEEE Int. Conf. on Systems, Man, and Cybernetics (SMC), 4066–4072. New York: IEEE. https://doi.org/10.1109/SMC42975.2020.9282951.
Thuruthel, T., E. Falotico, M. Cianchetti, F. Renda, and C. Laschi. 2016. “Learning global inverse statics solution for a redundant soft robot.” In Vol. 2 of Proc., 13th Int. Conf. on Informatics in Control, Automation and Robotics, 303–310. Berlin: Springer.
Williams, G., A. Aldrich, and E. A. Theodorou. 2017. “Model predictive path integral control: From theory to parallel computation.” J. Guid. Control Dyn. 40 (2): 344–357. https://doi.org/10.2514/1.G001921.
Yip, M., J. Sganga, and D. Camarillo. 2017. “Autonomous control of continuum robot manipulators for complex cardiac ablation tasks.” J. Med. Rob. Res. 2 (1): 1750002. https://doi.org/10.1142/S2424905X17500027.
Zhang, T., G. Kahn, S. Levine, and P. Abbeel. 2016 “Learning deep control policies for autonomous aerial vehicles with MPC-guided policy search.” In Proc., 2016 IEEE Int. Conf. on Robotics and Automation (ICRA), 528–535. New York: IEEE. https://doi.org/10.1109/ICRA.2016.7487175.
Zhou, M., X. Qu, and S. Jin. 2017. “On the impact of cooperative autonomous vehicles in improving freeway merging: A modified intelligent driver model-based approach.” IEEE Trans. Intell. Transp. Syst. 18 (6): 1422–1428. https://doi.org/10.1109/TITS.2016.2606492.
Zhu, Y., D. Zhao, H. He, and J. Ji. 2017. “Event-triggered optimal control for partially unknown constrained-input systems via adaptive dynamic programming.” IEEE Trans. Ind. Electron. 64 (5): 4101–4109. https://doi.org/10.1109/TIE.2016.2597763.
Zhu, Y., D. Zhao, X. Yang, and Q. Zhang. 2018. “Policy iteration for H optimal control of polynomial nonlinear systems via sum of squares programming.” IEEE Trans. Cybern. 48 (2): 500–509. https://doi.org/10.1109/TCYB.2016.2643687.

Information & Authors

Information

Published In

Go to Journal of Aerospace Engineering
Journal of Aerospace Engineering
Volume 35Issue 5September 2022

History

Received: Sep 25, 2021
Accepted: Jan 18, 2022
Published online: Jun 2, 2022
Published in print: Sep 1, 2022
Discussion open until: Nov 2, 2022

Permissions

Request permissions for this article.

Authors

Affiliations

Ph.D. Student, Dept. of Engineering Mechanics, Dalian Univ. of Technology, Dalian 116023, PR China. Email: [email protected]
Professor, Dept. of Engineering Mechanics, Dalian Univ. of Technology, Dalian 116023, PR China (corresponding author). Email: [email protected]
Zhongzhen Liu [email protected]
Ph.D. Student, Dept. of Engineering Mechanics, Dalian Univ. of Technology, Dalian 116023, PR China. Email: [email protected]
Haijun Peng [email protected]
Professor, Dept. of Engineering Mechanics, Dalian Univ. of Technology, Dalian 116023, PR China. Email: [email protected]
Professor, State Key Laboratory of Structural Analysis for Industrial Equipment, Dalian Univ. of Technology, Dalian 116023, PR China. Email: [email protected]

Metrics & Citations

Metrics

Citations

Download citation

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited by

  • Stability Analysis for Incremental Adaptive Dynamic Programming with Approximation Errors, Journal of Aerospace Engineering, 10.1061/JAEEEZ.ASENG-5097, 37, 1, (2024).

View Options

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Media

Figures

Other

Tables

Share

Share

Copy the content Link

Share with email

Email a colleague

Share