publication . Article . Other literature type . 2015

Reinforcement learning control for coordinated manipulation of multi-robots

Li, Yanan; Chen, Long; Tee, Keng Peng; Li, Qingquan;
Open Access
  • Published: 10 Jul 2015 Journal: Neurocomputing, volume 170, pages 168-175 (issn: 0925-2312, Copyright policy)
  • Publisher: Elsevier BV
  • Country: United Kingdom
Abstract
In this paper, coordination control is investigated for multi-robots to manipulate an object with a common desired trajectory. Both trajectory tracking and control input minimization are considered for each individual robot manipulator, such that possible disagreement between different manipulators can be handled. Reinforcement learning is employed to cope with the problem of unknown dynamics of both robots and the manipulated object. It is rigorously proven that the proposed method guarantees the coordination control of the multi-robots system under study. The validity of the proposed method is verified through simulation studies.
Subjects
arXiv: Computer Science::Robotics
free text keywords: Reinforcement learning control, Robot control, Machine learning, computer.software_genre, computer, Control theory, Reinforcement learning, Robot manipulator, Robot, Trajectory, Minification, Artificial intelligence, business.industry, business, Robot learning, Mathematics
Related Organizations
28 references, page 1 of 2

2.2. Preliminary For the approximation of a continuous function h(Z) : Rr → R, the following Radial Function Basis (RBF) Neural Networks (NN) are used [26]:

[1] J. P. Merlet (Ed.), Parallel robots, Vol. 74, Springer, 2001.

[2] R. Bonitz, T. Hsia, Robust internal-force based impedance control for coordinating manipulators-theory and experiments, in: Proceedings of IEEE International Conference on Robotics and Automation, Vol. 1, 1996, pp. 622-628.

[3] Z. Li, P. Y. Tao, S. S. Ge, M. Adams, W. Wijesoma, Robust adaptive control of cooperating mobile manipulators with relative motion, IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics 39 (1) (2009) 103-116.

[4] Y. Li, S. S. Ge, Impedance learning for robots interacting with unknown environments, IEEE Transactions on Control Systems Technology 22 (4) (2014) 1422-1432.

[5] H. Kwakernaak, R. Sivan, Linear Optimal Control Systems, John Wiley & Sons, Inc. New York, NY, USA, 1972.

[6] R. S. Sutton, A. G. Barto, R. J. Williams, Reinforcement learning is direct adaptive optimal control, IEEE Control Systems 12 (2) (1992) 19-22.

[7] X. Liu, S. Balakrishnan, Convergence analysis of adaptive critic based optimal control, in: Proceedings of American Control Conference, Vol. 3, 2000, pp. 1929-1933.

[8] J. J. Murray, C. J. Cox, G. G. Lendaris, R. Saeks, Adaptive dynamic programming, IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews 32 (2) (2002) 140-153.

[9] M. T. Rosenstein, A. G. Barto, Supervised actor-critic reinforcement learning, Learning and Approximate Dynamic Programming: Scaling Up to the Real World (2004) 359-380.

[10] P. J. Werbos, Using ADP to understand and replicate brain intelligence: The next level design, Proceedings of the 2007 IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning (2007) 209-216. [OpenAIRE]

[11] P. J. Werbos, Intelligence in the brain: A theory of how it works and how to build it, Neural Networks 22 (3) (2009) 200-212. [OpenAIRE]

[12] D. Vrabie, F. Lewis, Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems, Neural Networks 22 (3) (2009) 237-246.

[13] D. P. Bertsekas, D. P. Bertsekas, D. P. Bertsekas, D. P. Bertsekas, Dynamic programming and optimal control, Vol. 1, Athena Scientific Belmont, MA, 1995.

[14] K. Doya, Reinforcement learning in continuous time and space, Neural computation 12 (1) (2000) 219-245. [OpenAIRE]

28 references, page 1 of 2
Powered by OpenAIRE Research Graph
Any information missing or wrong?Report an Issue