Show simple item record

dc.contributor.authorAloor, Jasmine
dc.contributor.authorNayak, Siddharth Nagar
dc.contributor.authorDolan, Sydney
dc.contributor.authorBalakrishnan, Hamsa
dc.date.accessioned2024-11-14T21:09:42Z
dc.date.available2024-11-14T21:09:42Z
dc.date.issued2024-10-29
dc.identifier.issn2833-0528
dc.identifier.urihttps://hdl.handle.net/1721.1/157544
dc.description.abstractMulti-agent systems are trained to maximize shared cost objectives, which typically reflect system-level efficiency. However, in the resource-constrained environments of mobility and transportation systems, efficiency may be achieved at the expense of fairness --- certain agents may incur significantly greater costs or lower rewards compared to others. Tasks could be distributed inequitably, leading to some agents receiving an unfair advantage while others incur disproportionately high costs. It is, therefore, important to consider the tradeoffs between efficiency and fairness in such settings. We consider the problem of fair multi-agent navigation for a group of decentralized agents using multi-agent reinforcement learning (MARL). We consider the reciprocal of the coefficient of variation of the distances traveled by different agents as a measure of fairness and investigate whether agents can learn to be fair without significantly sacrificing efficiency (i.e., increasing the total distance traveled). We find that by training agents using min-max fair distance goal assignments along with a reward term that incentivizes fairness as they move towards their goals, the agents (1) learn a fair assignment of goals and (2) achieve almost perfect goal coverage in navigation scenarios using only local observations. For goal coverage scenarios, we find that, on average, the proposed model yields a 14% improvement in efficiency and a 5% improvement in fairness over a baseline model that is trained using random assignments. Furthermore, an average of 21% improvement in fairness can be achieved by the proposed model as compared to a model trained on optimally efficient assignments; this increase in fairness comes at the expense of only a 7% decrease in efficiency. Finally, we extend our method to environments in which agents must complete coverage tasks in prescribed formations and show that it is possible to do so without tailoring the models to specific formation shapes.en_US
dc.publisherACMen_US
dc.relation.isversionofhttp://dx.doi.org/10.1145/3702012en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleCooperation and Fairness in Multi-Agent Reinforcement Learningen_US
dc.typeArticleen_US
dc.identifier.citationAloor, Jasmine, Nayak, Siddharth Nagar, Dolan, Sydney and Balakrishnan, Hamsa. 2024. "Cooperation and Fairness in Multi-Agent Reinforcement Learning." ACM Journal on Autonomous Transportation Systems.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Aeronautics and Astronauticsen_US
dc.relation.journalACM Journal on Autonomous Transportation Systemsen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2024-11-01T07:45:59Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-11-01T07:45:59Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record