Institutional Repository
Technical University of Crete
EN  |  EL

Search

Browse

My Space

Learning in zero–sum team Markov games using factored value functions

Lagoudakis Michael, Parr, R.

Full record


URI: http://purl.tuc.gr/dl/dias/FBB6EA9E-B181-4D39-8F6C-4DDB3B0278DA
Year 2002
Type of Item Conference Full Paper
License
Details
Bibliographic Citation M.G. Lagoudakis and R.Parr. (2002, Dec.).Learning in zero–sum team Markov games using factored value functions. [Online]. Available: http://machinelearning.wustl.edu/mlpapers/paper_files/CN15.pdf
Appears in Collections

Summary

We present a new method for learning good strategies in zero-sumMarkov games in which each side is composed of multiple agents collaboratingagainst an opposing team of agents. Our method requires fullobservability and communication during learning, but the learned policiescan be executed in a distributed manner. The value function is representedas a factored linear architecture and its structure determines thenecessary computational resources and communication bandwidth. Thisapproach permits a tradeoff between simple representations with little orno communication between agents and complex, computationally intensiverepresentations with extensive coordination between agents. Thus,we provide a principled means of using approximation to combat theexponential blowup in the joint action space of the participants. The approachis demonstrated with an example that shows the efficiency gainsover naive enumeration.

Services

Statistics