Institutional Repository
Technical University of Crete
EN  |  EL

Search

Browse

My Space

Least-squares methods in reinforcement learning for control

Lagoudakis Michael, Parr Ronald, Littman Michael L.

Full record


URI: http://purl.tuc.gr/dl/dias/E2365192-8D33-48EE-8069-2097118F293B
Year 2002
Type of Item Conference Full Paper
License
Details
Bibliographic Citation M. Lagoudakis , R. Parr , M. Littman ," Least-Squares Methods in Reinforcement Learning for Control," in 2002 Proceedings of the 2nd Hellenic Conference on Artificial Intelligence (SETN), pp. 249-260, doi: 10.1007/3-540-46014-4_23 https://doi.org/10.1007/3-540-46014-4_23
Appears in Collections

Summary

Least-squares methods have been successfully used for prediction problems in the context of reinforcement learning, but little has been done in extending these methods to control problems. This paper presents an overview of our research efforts in using least-squares techniques for control. In our early attempts, we considered a direct extension of the Least-Squares Temporal Difference (LSTD) algorithm in the spirit of Q-learning. Later, an effort to remedy some limitations of this algorithm (approximation bias, poor sample utilization) led to the Least- Squares Policy Iteration (LSPI) algorithm, which is a form of model-free approximate policy iteration and makes efficient use of training samples collected in any arbitrary manner. The algorithms are demonstrated on a variety of learning domains, including algorithm selection, inverted pendulum balancing, bicycle balancing and riding, multiagent learning in factored domains, and, recently, on two-player zero-sum Markov games and the game of Tetris.

Services

Statistics