TY - JOUR
T1 - Space manipulator optimal impedance control using integral reinforcement learning
AU - Wu, Han
AU - Hu, Qinglei
AU - Shi, Yongxia
AU - Zheng, Jianying
AU - Sun, Kaipeng
AU - Wang, Jiawen
PY - 2023/8
Y1 - 2023/8
N2 - This paper examines the optimal impedance control problem for large-scale space manipulator operational tasks with unknown contact dynamics and partial measurements. More specifically, by quantifying the interaction performance using a discounted value function, the optimal impedance control problem is tactfully transformed into a linear quadratic tracking problem. By resorting to the historical inputs and outputs, an improved state reconstruction method is presented, which obviates the velocity measurement. Unlike the existing state reconstruction method for continuous-time systems, the estimation bias caused by probing noise is completely eliminated under the improved state reconstruction method. Based on this, a novel model-free value iteration integral reinforcement learning algorithm is developed to approximate optimal impedance parameters. Compared with the earlier integral reinforcement learning algorithms, the proposed algorithm not only averts any prior contact dynamics knowledge and full-state measurements, but also eliminates the heavy dependence on the specific initial stabilization control. In addition, the implementation and convergence of the proposed algorithm are discussed successively. Finally, numerical simulations verify the effectiveness of the theoretical results. © 2023 Elsevier Masson SAS.
AB - This paper examines the optimal impedance control problem for large-scale space manipulator operational tasks with unknown contact dynamics and partial measurements. More specifically, by quantifying the interaction performance using a discounted value function, the optimal impedance control problem is tactfully transformed into a linear quadratic tracking problem. By resorting to the historical inputs and outputs, an improved state reconstruction method is presented, which obviates the velocity measurement. Unlike the existing state reconstruction method for continuous-time systems, the estimation bias caused by probing noise is completely eliminated under the improved state reconstruction method. Based on this, a novel model-free value iteration integral reinforcement learning algorithm is developed to approximate optimal impedance parameters. Compared with the earlier integral reinforcement learning algorithms, the proposed algorithm not only averts any prior contact dynamics knowledge and full-state measurements, but also eliminates the heavy dependence on the specific initial stabilization control. In addition, the implementation and convergence of the proposed algorithm are discussed successively. Finally, numerical simulations verify the effectiveness of the theoretical results. © 2023 Elsevier Masson SAS.
KW - Integral reinforcement learning (IRL)
KW - Optimal impedance control
KW - Space manipulator
KW - State reconstruction
KW - Unknown contact dynamics
UR - http://www.scopus.com/inward/record.url?scp=85159854704&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-85159854704&origin=recordpage
U2 - 10.1016/j.ast.2023.108388
DO - 10.1016/j.ast.2023.108388
M3 - RGC 21 - Publication in refereed journal
SN - 1270-9638
VL - 139
JO - Aerospace Science and Technology
JF - Aerospace Science and Technology
M1 - 108388
ER -