Learning a Transferable World Model by Reinforcement Agent in Deterministic Observable Grid-World Environments

Jurgita Kapočiūtė-Dzikienė, Gailius Raškinis
2012 Information Technology and Control  
Reinforcement-based agents have difficulties in transferring their acquired knowledge into new different environments due to the common identities-based percept representation and the lack of appropriate generalization capabilities. In this paper, the problem of knowledge transferability is addressed by proposing an agent dotted with decision tree induction and constructive induction capabilities and relying on decomposable properties-based percept representation. The agent starts without any
more » ... ior knowledge of its environment and of the effects of its actions. It learns a world model (the set of decision trees) that corresponds to the set of explicit action definitions predicting action effects in terms of agent's percepts. Agent's planning component uses predictions of the world model to chain actions via a breadth-first search. The proposed agent was compared to the Q-learning and Adaptive Dynamic Programming based agents and demonstrated better ability to achieve goals in static observable deterministic gridworld environments different from those in which it has learnt its world model.
doi:10.5755/j01.itc.41.4.915 fatcat:5pgeedmyjbhdzgmidzvkyneh4i