Data-Driven Human-Robot Interaction Without Velocity Measurement Using Off-Policy Reinforcement Learning | IEEE Journals & Magazine | IEEE Xplore

Data-Driven Human-Robot Interaction Without Velocity Measurement Using Off-Policy Reinforcement Learning


Abstract:

In this paper, we present a novel data-driven design method for the human-robot interaction (HRI) system, where a given task is achieved by cooperation between the human ...Show More

Abstract:

In this paper, we present a novel data-driven design method for the human-robot interaction (HRI) system, where a given task is achieved by cooperation between the human and the robot. The presented HRI controller design is a two-level control design approach consisting of a task-oriented performance optimization design and a plant-oriented impedance controller design. The task-oriented design minimizes the human effort and guarantees the perfect task tracking in the outer-loop, while the plant-oriented achieves the desired impedance from the human to the robot manipulator end-effector in the inner-loop. Data-driven reinforcement learning techniques are used for performance optimization in the outer-loop to assign the optimal impedance parameters. In the inner-loop, a velocity-free filter is designed to avoid the requirement of end-effector velocity measurement. On this basis, an adaptive controller is designed to achieve the desired impedance of the robot manipulator in the task space. The simulation and experiment of a robot manipulator are conducted to verify the efficacy of the presented HRI design framework.
Published in: IEEE/CAA Journal of Automatica Sinica ( Volume: 9, Issue: 1, January 2022)
Page(s): 47 - 63
Date of Publication: 13 September 2021

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Robotic manipulators have played an important role in engineering applications, such as master-slave teleoperation systems [1], construction automation [2], space engineering [3], to name a few. When the human is restricted by the physical limitation or when operating in a harsh environment is required, the robot manipulators can handle strenuous and dangerous work that is difficult or impossible for human operators. However, robot manipulators cannot completely replace the human because of robots' lack of high-level of learning and inference ability. Therefore, the interaction between the human and the robot could take the best of human and robot capabilities. Successful applications of human-robot interaction (HRI) system designs include physiotherapy [4], human amplifier [5], assistive haptic device [6]. In this paper, we employ the two-level design framework [7] for the HRI systems with novel outer-loop and inner-loop designs to obviate the requirement of velocity measurement and the complete knowledge of system dynamics.

Select All
1.
E. Nuño, R. Ortega and L. Basañez, "An adaptive controller for nonlinear teleoperators", Automatica, vol. 46, no. 1, pp. 155-159, 2010.
2.
S. Cai, Z. Ma, M. J. Skibniewski and S. Bao, "Construction automation and robotics for high-rise buildings over the past decades: A comprehensive review", Advanced Engineering Informatics, vol. 42, no. 100989, 2019.
3.
D. Han, P. Huang, X. Liu and Y. Yang, "Combined spacecraft stabilization control after multiple impacts during the capture of a tumbling target by a space robot", Acta Astronautica, vol. 176, pp. 24-32, 2020.
4.
S. E. Fasoli, H. I. Krebs, J. Stein, W. R. Frontera and N. Hogan, "Effects of robotic therapy on motor impairment and recovery in chronic stroke", Archives of Physical Medicine and Rehabilitation, vol. 84, no. 4, pp. 477-482, 2003.
5.
J. C. Perry, J. Rosen and S. Burns, "Upper-limb powered exoskeleton design", IEEE/ASME Transactions on Mechatronics, vol. 12, no. 4, pp. 408-417, 2007.
6.
M. Bergamasco, B. Allotta, L. Bosio, L. Ferretti, G. Parrini, G. Prisco, et al., "An arm exoskeleton system for teleoperation and virtual environments applications", Proc. IEEE Int. Conf. Robotics and Automation, pp. 1449-1454, 1994.
7.
H. Modares, I. Ranatunga, F. L. Lewis and D. O. Popa, "Optimized assistive human-robot interaction using reinforcement learning", IEEE Trans. Cybernetics, vol. 46, no. 3, pp. 655-667, 2015.
8.
K. Guo, Y. Pan, D. Zheng and H. Yu, "Composite learning control of robotic systems: A least squares modulated approach", Automatica, vol. 111, no. 108612, 2020.
9.
T. Sun and Y. Pan, "Robust adaptive control for prescribed performance tracking of constrained uncertain nonlinear systems", J. Franklin Institute, vol. 356, no. 1, pp. 18-30, 2019.
10.
K. Dupree, P. M. Patre, Z. D. Wilcox and W. E. Dixon, "Asymptotic optimal control of uncertain nonlinear euler-lagrange systems", Automatica, vol. 47, no. 1, pp. 99-107, 2011.
11.
Z. Li, J. Liu, Z. Huang, Y. Peng, H. Pu and L. Ding, "Adaptive impedance control of human-robot cooperation using reinforcement learning", IEEE Trans. Industrial Electronics, vol. 64, no. 10, pp. 8013-8022, 2017.
12.
T. Sun, L. Peng, L. Cheng, Z. Hou and Y. Pan, "Stability-guaranteed variable impedance control of robots based on approximate dynamic inversion", IEEE Trans. Systems Man and Cybernetics: Systems, vol. 51, no. 7, pp. 4193-4200, 2019.
13.
T. Sun, L. Cheng, L. Peng, Z. Hou and Y. Pan, "Learning impedance control of robots with enhanced transient and steady-state control performances", Science China Information Sciences, vol. 63, no. 9, pp. 1-13, 2020.
14.
T. Sun, L. Peng, L. Cheng, Z. Hou and Y. Pan, "Composite learning enhanced robot impedance control", IEEE Trans. Neural Networks and Learning Systems, vol. 31, no. 3, pp. 1052-1059, 2020.
15.
R. Colbaugh, H. Seraji and K. Glass, "Direct adaptive impedance control of robot manipulators", J. Robotic Systems, vol. 10, no. 2, pp. 217-248, 1993.
16.
S. Ge, C. Hang, L. Woon and X. Chen, "Impedance control of robot manipulators using adaptive neural networks", Int. J. Intelligent Control and Systems, vol. 2, no. 3, pp. 433-452, 1998.
17.
C. Wang, Y. Li, S. S. Ge and T. H. Lee, "Reference adaptation for robots in physical interactions with unknown environments", IEEE Transactions on Cybernetics, vol. 47, no. 11, pp. 3504-3515, 2016.
18.
W.-S. Lu and Q.-H. Meng, "Impedance control with adaptation for robotic manipulations", IEEE Trans. Robotics and Automation, vol. 7, no. 3, pp. 408-415, 1991.
19.
H. N. Rahimi, I. Howard and L. Cui, "Neural impedance adaption for assistive human-robot interaction", Neurocomputing, vol. 290, pp. 50-59, 2018.
20.
Y. Wang, W. Sun, Y. Xiang and S. Miao, "Neural network-based robust tracking control for robots", Intelligent Automation & Soft Computing, vol. 15, no. 2, pp. 211-222, 2009.
21.
R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction, USA:A Bradford Book, 2018.
22.
Y. Yang, D. Wunsch and Y. Yin, "Hamiltonian-driven adaptive dynamic programming for continuous nonlinear dynamical systems", IEEE Trans. Neural Networks and Learning Systems, vol. 28, no. 8, pp. 1929-1940, 2017.
23.
Y. Yang, K. G. Vamvoudakis, H. Modares, Y. Yin and D. C. Wunsch, "Hamiltonian-driven hybrid adaptive dynamic programming", IEEE Trans. Systems Man and Cybernetics: Systems, 2020.
24.
Y. Yang, K. G. Vamvoudakis, H. Modares, Y. Yin and D. C. Wunsch, "Safe intermittent reinforcement learning with static and dynamic event generators", IEEE Trans. Neural Networks and Learning Systems, vol. 31, no. 12, pp. 5441-5455, 2020.
25.
D. Wang and X. Zhong, "Advanced policy learning near-optimal regulation", IEEE/CAA J Automa Sinica, vol. 6, no. 3, pp. 743-749, 2019.
26.
Y. Yang, Z. Guo, H. Xiong, D. Ding, Y. Yin and D. C. Wunsch, "Datadriven robust control of discrete-time uncertain linear systems via offpolicy reinforcement learning", IEEE Trans. Neural Networks and Learning Systems, vol. 30, no. 12, pp. 3735-3747, 2019.
27.
D. Wang, D. Liu, C. Mu and Y. Zhang, "Neural network learning and robust stabilization of nonlinear systems with dynamic uncertainties", IEEE Trans. Neural Networks and Learning Systems, vol. 29, no. 4, pp. 1342-1351, 2018.
28.
Q. Zhang and D. Zhao, "Data-based reinforcement learning for nonzerosum games with unknown drift dynamics", IEEE Trans. Cybernetics, vol. 49, no. 8, pp. 2874-2885, 2019.
29.
H. Modares, F. L. Lewis and Z.-P. Jiang, " H ∞ tracking control of completely unknown continuous-time systems via off-policy reinforcement learning ", IEEE Trans. Neural Networks and Learning Systems, vol. 26, no. 10, pp. 2550-2562, 2015.
30.
B. Luo, H.-N. Wu and T. Huang, "Off-policy reinforcement learning for H8 control design", IEEE trans. Cybernetics, vol. 45, no. 1, pp. 65-76, 2014.

Contact IEEE to Subscribe

References

References is not available for this document.