Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp018p58pg99w
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorRamadge, Peter
dc.contributor.authorMatthews, Oliver
dc.date.accessioned2020-10-02T21:30:21Z-
dc.date.available2020-10-02T21:30:21Z-
dc.date.created2020-05-02
dc.date.issued2020-10-02-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp018p58pg99w-
dc.description.abstractMulti-agent reinforcement learning is an attractive and challenging prospect. Multi-agent settings are non-stationary by nature, and multi-agent contributions to reward introduce ambiguity not present in single-agent learning. In this paper, we start by looking at Policy Gradient based reinforcement learning techniques in a single-agent setting. We then show why these techniques fail in a multi-agent setting and suggest improvements to the algorithms to address these problems. Experiments are done on simulations of cyclists, with the aim of agents learning to work together to cycle efficiently as a pack. The improvements suggested include: adding a varying normaliser to deal with state distribution drift; using a value function which takes as input the actions of other agents to deal with non-stationarity; and an automatic tuning algorithm for learning hyperparameters. We find that our techniques show improvements on the traditional algorithms, and suggest a more complex environment and improvements on our algorithms to further demonstrate the advantages of the approach.
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.titleExtending Classical Deep Reinforcement Learning Techniques for use in Multi-Agent Systems
dc.typePrinceton University Senior Theses
pu.date.classyear2020
pu.departmentElectrical Engineering
pu.pdf.coverpageSeniorThesisCoverPage
pu.contributor.authorid920227248
Appears in Collections:Electrical Engineering, 1932-2020

Files in This Item:
File Description SizeFormat 
MATTHEWS-OLIVER-THESIS.pdf9.17 MBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.