Learning reaching strategies through reinforcement for a sensor-based manipulator
A neural controller is presented that learns goal-oriented obstacle-avoiding reaction strategies for a multilink robot arm. It acquires these strategies on-line from local sensory data. The controller consists of two neural network modules: an actor-critic module and a module for differential inverse kinematics (DIV). The input codification for the controller exploits the inherent symmetry of the robot arm kinematics. The actor-critic module generates actions with regard to the shortest path vector (SPV) to the closest goal in the configuration space. However, the computation of the SPV is cumbersome for manipulators with more than two links. The DIV module aims to overcome the SPV calculation. This module provides a goal vector by means of the inversion of a neural network that has been previously trained to approximate the manipulator forward kinematics. Results for a two-link robot arm show that the combination of both modules speeds up the learning process.
Bibliographic Reference: Article: Neural Networks
Record Number: 199711363 / Last updated on: 1997-10-23
Original language: en
Available languages: en