A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2017; you can also visit the original URL.
The file type is application/pdf
.
Improving reinforcement learning function approximators via neuroevolution
2005
Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems - AAMAS '05
Reinforcement learning problems are commonly tackled with temporal difference methods, which use dynamic programming and statistical sampling to estimate the long-term value of taking each action in each state. In most problems of realworld interest, learning this value function requires a function approximator, which represents the mapping from stateaction pairs to values via a concise, parameterized function and uses supervised learning methods to set its parameters. Function approximators
doi:10.1145/1082473.1082794
dblp:conf/atal/Whiteson05
fatcat:d3ahainaybfc3ftrpfgortlxu4