Feature Selection for Neuro-Dynamic Programming

Feature Selection for Neuro-Dynamic Programming

Neuro-Dynamic Programming encompasses techniques from both reinforcement learning and approximate dynamic programming. Feature selection refers to the choice of basis that defines the function class that is required in the application of these techniques. This chapter reviews two popular approaches to neuro-dynamic programming, TD-learning and Q-learning. The main goal of the chapter is to demonstrate how insight from idealized models can be used as a guide for feature selection for these algorithms. Several approaches are surveyed, including fluid and diffusion models, and the application of idealized models arising from mean-field game approximations. The theory is illustrated with several examples. Book chapter,  D. Huang, W. Chen, P. Mehta, S. Meyn, and A. Surana. Feature selection for neuro-dynamic programming. In F. Lewis, editor, Reinforcement Learning and Approximate Dynamic Programming for Feedback Control. Wiley, 2011

 

2013 journal submission

Related references from the Illinois Archive:

@incollection{huachemehmeysur11,
Author = {Huang, D. and Chen, W. and Mehta, P. and Meyn, S. and Surana, A.},
Booktitle = {Reinforcement Learning and Approximate Dynamic Programming for Feedback Control},
Editor = {Lewis, F.},
Publisher = {Wiley},
Title = {Feature Selection for Neuro-Dynamic Programming},
Year = {2011}}