摘要
为了提高力控制的性能,使机器人高效自主地学习执行力控制任务,本文提出一种学习变阻抗控制方法。该方法采用基于模型的强化学习算法学习最优阻抗调节策略,使用高斯过程模型作为系统的变换动力学模型,允许概率化的推理与规划,并在成本函数中加入能量损失项,实现误差和能量的权衡。仿真实验结果表明:该学习变阻抗控制方法具有高效性,仅需数次交互即可成功学习完成力控制任务,大大减少了所需的交互次数与交互时间,且学习得到的阻抗控制策略具有仿生特性,可用于学习执行力敏感型任务。
To improve the performance of force control and enable robots to learn how to execute force control tasks autonomously and efficiently,this paper presents an efficient variable impedance control method.The proposed method learns the optimal impedance regulation strategy using a model-based reinforcement learning algorithm.The Gaussian process model is used as a transformation dynamics model in the system.This model permits probabilistic inference and planning.Also,an energy consumption item is added to cost function to achieve a trade-off between error and energy.The simulation results show the efficiency of the proposed method,requiring only a few interactions to successfully learn how to complete force control tasks.Furthermore,the required number of interactions and interaction time are significantly reduced.The learned impedance control strategy features bionic characteristics,which are applicable in learning how to perform force-sensitive tasks.
作者
李超
张智
夏桂华
谢心如
朱齐丹
刘琦
LI Chao;ZHANG Zhi;XIA Guihua;XIE Xinru;ZHU Qidan;LIU Qi(College of Automation,Harbin Engineering University,Harbin 150001,China;Institute of Chemical Materials,China Academy of Engineering Physics,Mianyang 621000,China)
出处
《哈尔滨工程大学学报》
EI
CAS
CSCD
北大核心
2019年第2期304-311,共8页
Journal of Harbin Engineering University
基金
国家自然科学基金项目(U1530119).
关键词
机器人
阻抗控制
力控制
控制策略
强化学习
高效
高斯过程
成本函数
robot
impedance control
force control
control strategy
reinforcement learning
efficient
Gaussian process
cost function
作者简介
李超,男,博士研究生;通信作者:张智,男,副教授,硕士生导师,E-mail:zhangzhi1981@hrbeu.edu.cn.