Grondman I, Vaandrager M, Buşoniu L, Babûska R, Schuitema E (2011) Actor-critic control with reference model learning. In: 51st IEEE conference on decision and control (CDC), pp 5272–5277. Grondman I, Buşoniu L, Babûska R (2012) Model learning actor-critic algorithms: performance evaluation in a motion control task. Galicki M (2016) Robust task space trajectory tracking control of robotic manipulators. Galicki M (2016) Finite-time trajectory tracking control in task space of robotic manipulators. Int J Comput Commun Control 7(3):459–472. įeng Y, Yao-nan W, Yi-min Y (2012) Inverse kinematics solution for robot manipulator based on neural network under joint subspace. Petru Maior University of Tirgu Mures, Romania. In: Procedia technology, the 7th international conference interdisciplinarity in engineering, INTER-ENG 2013. ĭuka A (2014) Neural network based inverse kinematics solution for trajectory tracking of a robotic arm. In: Proceedings of the 28th international conference on machine learning, Bellevue, WAA, USAĭeisenroth MP, Neumann G, Peters J (2011) A survey on policy search for robotics. ĭeisenroth M, Rasmussen C (2011) PILCO: A model-based and data-efficient approach to policy search. In: 24th international conference on mechatronics and machine vision in practice. Ĭsistzar A, Eilers J, Verl A (2017) On solving the inverse kinematics problem using neural networks. CRC Press, Boca RatonĬheah C, Li X (2011) Singularity-robust task-space tracking control of robot. Automation and Control Engineering Series. īuşoniu L, Babûska R, De Schutter B, Ernst D (2010) Reinforcement learning and dynamic programming using function approximators. Lecture notes in computer science, vol 310. In: Srinivasan D, Jain L (eds) Innovations in multi-agent systems and applications-1. īuşoniu L, Babûska R, De Schutter B (2010) Multi-agent reinforcement learning: an overview. IEEE/RSJ Int Conf Intell Robots Syst (IROS). īitzer S, Howard M, Vijayakumar S (2010) Using dimensionality reduction to exploit constraints in reinforcement learning. īcsi B, Nguyen-Tuong D, Csat L, Schlkopf B, Peters J (2011) Learning inverse kinematics with structured prediction. Īxinte D, Dong X, Palmer D, Rushworth A, Guzman S, Olarra A (2018) Miror-miniaturized robotic systems for holisticin-siturepair and maintenance works in restrained and hazardous environments. Ītashzar S, Tavakoli M, Patel R (2018) A computational-model-based study of supervised haptics-enabled therapist-in-the-loop training for upper-limb poststroke robotic rehabilitation. In: 6th IEEE RAS/EMBS international conference on biomedical robotics and biomechatronics (BioRob). Īnsari Y, Falotico E (2016) A multiagent reinforcement learning approach for inverse kinematics oh high dimensional manipulators with precision positioning. The experimental results show that our MARL is much more better compared with the classic methods such as Jacobian-based methods and neural networks.Īhmadi S, Fateh M (2018) Task-space asymptotic tracking control of robots using a direct adaptive Taylor series controller. The convergence property of the proposed MARL is analyzed. The fully cooperative MARL uses a kinematic learning to avoid function approximators and large learning space. Each joint of the robot is regarded as one agent. In this paper, we propose a fully cooperative multi-agent reinforcement learning (MARL) to solve the kinematic problem of redundant robots. However, NN needs big data and classical RL is not suitable for multi-link robots controlled in task space. Intelligent learning methods, such as neural networks (NN) and reinforcement learning (RL) can learn the inverse kinematics solution. However, they are not always available for redundant robots because there are more joint degrees-of-freedom than Cartesian degrees-of-freedom. Task-space control needs the inverse kinematics solution or Jacobian matrix for the transformation from task space to joint space.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |