Apprenticeship learning via soft local homomorphisms

Abd Boularias, Brahim Chaib-draa
2010 2010 IEEE International Conference on Robotics and Automation  
We consider the problem of apprenticeship learning when the expert's demonstration covers only a small part of a large state space. Inverse Reinforcement Learning (IRL) provides an efficient solution to this problem based on the assumption that the expert is optimally acting in a Markov Decision Process (MDP). However, past work on IRL requires an accurate estimate of the frequency of encountering each feature of the states when the robot follows the expert's policy. Given that the complete
more » ... cy of the expert is unknown, the features frequencies can only be empirically estimated from the demonstrated trajectories. In this paper, we propose to use a transfer method, known as soft homomorphism, in order to generalize the expert's policy to unvisited regions of the state space. The generalized policy can be used either as the robot's final policy, or to calculate the features frequencies within an IRL algorithm. Empirical results show that our approach is able to learn good policies from a small number of demonstrations.
doi:10.1109/robot.2010.5509717 dblp:conf/icra/BoulariasC10 fatcat:43onzr4yivgb7baxtyy6mamb5e