Multiagent reinforcement learning using function approximation

Citation
O. Abul et al., Multiagent reinforcement learning using function approximation, IEEE SYST C, 30(4), 2000, pp. 485-497
Citations number
38
Categorie Soggetti
AI Robotics and Automatic Control
Journal title
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS
ISSN journal
10946977 → ACNP
Volume
30
Issue
4
Year of publication
2000
Pages
485 - 497
Database
ISI
SICI code
1094-6977(200011)30:4<485:MRLUFA>2.0.ZU;2-A
Abstract
Learning in a partially observable and nonstationary environment is still o ne of the challenging problems In the area of multiagent (MA) learning. Rei nforcement learning is a generic method that suits the needs of MA learning in many aspects. This paper presents two new multiagent based domain indep endent coordination mechanisms for reinforcement learning; multiple agents do not require explicit communication among themselves to learn coordinated behavior. The first coordination mechanism Is perceptual coordination mech anism, where other agents are included in state descriptions and coordinati on information is Learned from state transitions. The second is observing c oordination mechanism, which also includes other agents in state descriptio ns and additionally the rewards of nearby agents are observed from the envi ronment. The observed rewards and agent's own reward are used to construct an optimal policy. This way, the latter mechanism tends to increase region- wide joint rewards. The selected experimented domain is adversarial food-co llecting world (AFCW), which can be configured both as single and multiagen t environments, Function approximation and generalization techniques are us ed because of the huge state space. Experimental results show the effective ness of these mechanisms.