A fully-decentralised general-sum approach for multi-agent reinforcement learning using minimal modelling

dc.contributor.authorKruger, Marcel Matthew Anthony
dc.contributor.supervisorRosman, Benjamin
dc.contributor.supervisorJames, Steven
dc.contributor.supervisorShipton, Jarrod
dc.date.accessioned2024-10-26T19:31:22Z
dc.date.available2024-10-26T19:31:22Z
dc.date.issued2023-08
dc.descriptionA dissertation submitted in fulfilment of the requirements for the degree of Master of Science, to the Faculty of Science, School of Computer Science & Applied Mathematics, University of the Witwatersrand, Johannesburg, 2023.
dc.description.abstractMulti-agent reinforcement learning is a prominent area of research in machine learning, extending reinforcement learning to scenarios where multiple agents concurrently learn and interact within the same environment. Most existing methods rely on centralisation during training, while others employ agent modelling. In contrast, we propose a novel method that adapts the role of entropy to assist in fully-decentralised training without explicitly modelling other agents using additional information to which most centralised methods assume access. We augment entropy to encourage more deterministic agents, and instead, we let the non-stationarity inherent in MARL serve as a mode for exploration. We empirically evaluate the performance of our method across five distinct environments, each representing unique challenges. Our assessment encompasses both cooperative and competitive cases. Our findings indicate that the approach of penalising entropy, rather than rewarding it, enables agents to perform at least as well as the prevailing standard of entropy maximisation. Moreover, our alternative approach achieves several of the original objectives of entropy regularisation in reinforcement learning, such as increased sample efficiency and potentially better final rewards. Whilst entropy has a significant role, our results in the competitive case indicate that position bias is still a considerable challenge.
dc.description.submitterMM2024
dc.facultyFaculty of Science
dc.identifier0000-0003-4803-5037
dc.identifier.citationKruger, Marcel Matthew Anthony. (2023). A fully-decentralised general-sum approach for multi-agent reinforcement learning using minimal modelling. [Master's dissertation, University of the Witwatersrand, Johannesburg]. https://hdl.handle.net/10539/41978
dc.identifier.urihttps://hdl.handle.net/10539/41978
dc.language.isoen
dc.publisherUniversity of the Witwatersrand, Johannesburg
dc.rights©2023 University of the Witwatersrand, Johannesburg. All rights reserved. The copyright in this work vests in the University of the Witwatersrand, Johannesburg. No part of this work may be reproduced or transmitted in any form or by any means, without the prior written permission of University of the Witwatersrand, Johannesburg.
dc.rights.holderUniversity of the Witwatersrand, Johannesburg
dc.schoolSchool of Computer Science and Applied Mathematics
dc.subjectReinforcement Learning
dc.subjectMulti-Agent Reinforcement Learning
dc.subjectDeep Learning
dc.subjectMulti-Agent Systems
dc.subjectEntropy
dc.subjectUCTD
dc.subject.otherSDG-9: Industry, innovation and infrastructure
dc.titleA fully-decentralised general-sum approach for multi-agent reinforcement learning using minimal modelling
dc.typeDissertation
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Kruger_Fully-decentralised_2023.pdf
Size:
16.88 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.43 KB
Format:
Item-specific license agreed upon to submission
Description: