Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/109431
Citations
Scopus Web of ScienceĀ® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorNguyen, D.-
dc.contributor.authorWhite, L.-
dc.contributor.authorNguyen, H.-
dc.contributor.editorKang, B.-
dc.contributor.editorBai, Q.-
dc.date.issued2016-
dc.identifier.citationLecture Notes in Artificial Intelligence, 2016 / Kang, B., Bai, Q. (ed./s), vol.9992 LNAI, pp.29-41-
dc.identifier.isbn9783319501260-
dc.identifier.issn0302-9743-
dc.identifier.issn1611-3349-
dc.identifier.urihttp://hdl.handle.net/2440/109431-
dc.descriptionLNAI 9992-
dc.description.abstractWe propose a novel adaptive reinforcement learning (RL) procedure for multi-agent non-cooperative repeated games. Most existing regret-based algorithms only use positive regrets in updating their learning rules. In this paper, we adopt both positive and negative regrets in reinforcement learning to improve its convergence behaviour. We prove theoretically that the empirical distribution of the joint play converges to the set of correlated equilibrium. Simulation results demonstrate that our proposed procedure outperforms the standard regret-based RL approach and a well-known state-of-the-art RL scheme in the literature in terms of both computational requirements and system fairness. Further experiments demonstrate that the performance of our solution is robust to variations in the total number of agents in the system; and that it can achieve markedly better fairness performance when compared to other relevant methods, especially in a large-scale multiagent system.-
dc.description.statementofresponsibilityDuong D. Nguyen, B, Langford B. White, and Hung X. Nguyen-
dc.language.isoen-
dc.publisherSpringer-
dc.relation.ispartofseriesLecture notes in computer science-
dc.rightsSpringer International Publishing AG 2016-
dc.source.urihttp://dx.doi.org/10.1007/978-3-319-50127-7_3-
dc.subjectMultiagent systems; Reinforcement Learning; Game theory; Correlated equilibrium; No regret-
dc.titleAdaptive multiagent reinforcement learning with non-positive regret-
dc.typeConference paper-
dc.contributor.conference29th Australasian Joint Conference on Artificial Intelligence (AI) (5 Dec 2016 - 8 Dec 2016 : Hobart, Tas)-
dc.identifier.doi10.1007/978-3-319-50127-7_3-
dc.relation.granthttp://purl.org/au-research/grants/arc/LP100200493-
pubs.publication-statusPublished-
dc.identifier.orcidNguyen, D. [0000-0003-1048-5825]-
dc.identifier.orcidWhite, L. [0000-0001-6660-0517]-
dc.identifier.orcidNguyen, H. [0000-0003-1028-920X]-
Appears in Collections:Aurora harvest 3
Electrical and Electronic Engineering publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.