Font Size: a A A

Automatic basis function construction for reinforcement learning and approximate dynamic programming

Posted on:2009-02-16Degree:M.ScType:Thesis
University:McGill University (Canada)Candidate:Keller, Philipp WFull Text:PDF
GTID:2448390002997953Subject:Operations Research
Abstract/Summary:
We address the problem of automatically constructing basis functions for linear approximation of the value function of a Markov decision process (MDP). Our work builds on results by Bertsekas and Castanon (1989) who proposed a method for automatically aggregating states to speed up value iteration. We propose to use neighbourhood component analysis, a dimensionality reduction technique created for supervised learning, in order to map a high-dimensional state space to a low-dimensional space, based on the Bellman error, or on the temporal difference (TD) error. We then place basis functions in the lower-dimensional space. These are added as new features for the linear function approximator. This approach is applied to a high-dimensional inventory control problem, and to a number of benchmark reinforcement learning problems.
Keywords/Search Tags:Basis, Function
Related items