Please use this identifier to cite or link to this item: http://hdl.handle.net/1942/729
Full metadata record
DC FieldValueLanguage
dc.contributor.authorTUYLS, Karl-
dc.contributor.authorVerbeeck, Katja-
dc.contributor.authorLenaerts, Tom-
dc.date.accessioned2005-04-20T06:47:26Z-
dc.date.available2005-04-20T06:47:26Z-
dc.date.issued2003-
dc.identifier.citationInternational Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems. p. 693-700.-
dc.identifier.isbn1-58113-683-8-
dc.identifier.urihttp://hdl.handle.net/1942/729-
dc.description.abstractAlthough well understood in the single-agent framework, the use of traditional reinforcement learning (RL) algorithms in multi-agent systems (MAS) is not always justified. The feedback an agent experiences in a MAS, is usually influenced by the other agents present in the system. Multi agent environments are therefore non-stationary and convergence and optimality guarantees of RL algorithms are lost. To better understand the dynamics of traditional RL algorithms we analyze the learning process in terms of evolutionary dynamics. More specifically we show how the Replicator Dynamics (RD) can be used as a model for Q-learning in games. The dynamical equations of Q-learning are derived and illustrated by some well chosen experiments. Both reveal an interesting connection between the exploitation-exploration scheme from RL and the selection-mutation mechanisms from evolutionary game theory.-
dc.description.sponsorshipSIGART: ACM Special Interest Group on Artificial Intelligence ACM: Association for Computing Machinery-
dc.language.isoen-
dc.publisherAssociation for Computing Machinery-
dc.titleA selection-mutation model for q-learning in multi-agent systems-
dc.typeProceedings Paper-
local.bibliographicCitation.conferencenameInternational Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems-
local.bibliographicCitation.conferenceplaceMelbourne, Australia-
dc.identifier.epage700-
dc.identifier.spage693-
local.bibliographicCitation.jcatC1-
local.type.specifiedProceedings Paper-
dc.bibliographicCitation.oldjcatC2-
dc.identifier.urlhttp://doi.acm.org/10.1145/860575.860687-
local.bibliographicCitation.btitleInternational Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems-
item.contributorTUYLS, Karl-
item.contributorVerbeeck, Katja-
item.contributorLenaerts, Tom-
item.accessRightsClosed Access-
item.fullcitationTUYLS, Karl; Verbeeck, Katja & Lenaerts, Tom (2003) A selection-mutation model for q-learning in multi-agent systems. In: International Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems. p. 693-700..-
item.fulltextNo Fulltext-
Appears in Collections:Research publications
Show simple item record

Page view(s)

96
checked on Nov 7, 2023

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.