Please use this identifier to cite or link to this item:
http://hdl.handle.net/1942/729
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | TUYLS, Karl | - |
dc.contributor.author | Verbeeck, Katja | - |
dc.contributor.author | Lenaerts, Tom | - |
dc.date.accessioned | 2005-04-20T06:47:26Z | - |
dc.date.available | 2005-04-20T06:47:26Z | - |
dc.date.issued | 2003 | - |
dc.identifier.citation | International Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems. p. 693-700. | - |
dc.identifier.isbn | 1-58113-683-8 | - |
dc.identifier.uri | http://hdl.handle.net/1942/729 | - |
dc.description.abstract | Although well understood in the single-agent framework, the use of traditional reinforcement learning (RL) algorithms in multi-agent systems (MAS) is not always justified. The feedback an agent experiences in a MAS, is usually influenced by the other agents present in the system. Multi agent environments are therefore non-stationary and convergence and optimality guarantees of RL algorithms are lost. To better understand the dynamics of traditional RL algorithms we analyze the learning process in terms of evolutionary dynamics. More specifically we show how the Replicator Dynamics (RD) can be used as a model for Q-learning in games. The dynamical equations of Q-learning are derived and illustrated by some well chosen experiments. Both reveal an interesting connection between the exploitation-exploration scheme from RL and the selection-mutation mechanisms from evolutionary game theory. | - |
dc.description.sponsorship | SIGART: ACM Special Interest Group on Artificial Intelligence ACM: Association for Computing Machinery | - |
dc.language.iso | en | - |
dc.publisher | Association for Computing Machinery | - |
dc.title | A selection-mutation model for q-learning in multi-agent systems | - |
dc.type | Proceedings Paper | - |
local.bibliographicCitation.conferencename | International Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems | - |
local.bibliographicCitation.conferenceplace | Melbourne, Australia | - |
dc.identifier.epage | 700 | - |
dc.identifier.spage | 693 | - |
local.bibliographicCitation.jcat | C1 | - |
local.type.specified | Proceedings Paper | - |
dc.bibliographicCitation.oldjcat | C2 | - |
dc.identifier.url | http://doi.acm.org/10.1145/860575.860687 | - |
local.bibliographicCitation.btitle | International Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems | - |
item.accessRights | Closed Access | - |
item.contributor | TUYLS, Karl | - |
item.contributor | Verbeeck, Katja | - |
item.contributor | Lenaerts, Tom | - |
item.fullcitation | TUYLS, Karl; Verbeeck, Katja & Lenaerts, Tom (2003) A selection-mutation model for q-learning in multi-agent systems. In: International Conference on Autonomous Agents . Proceedings of the second international joint conference on Autonomous agents and multiagent systems. p. 693-700.. | - |
item.fulltext | No Fulltext | - |
Appears in Collections: | Research publications |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.