Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01w3763695z
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorBubeck, Sebastien-
dc.contributor.authorFong, Christian-
dc.date.accessioned2014-07-16T19:34:40Z-
dc.date.available2014-07-16T19:34:40Z-
dc.date.created2014-06-
dc.date.issued2014-07-16-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01w3763695z-
dc.description.abstractSo far, mutli-armed bandit researchers have restricted their attention to games with a single player. I consider a multiplayer multi-armed bandit problem with infinitely many Bernoulli arms. The mean rewards of these arms follow the uniform distribution. However, players do not have perfect freedom in communicating their actions and payoffs. Instead, players are arranged in a star such the root player at the center of the star can observe all other players, but the peripheral players can only observe the root player. This thesis adapts the algorithm provided by Bonald and Proutière for this multiplayer setting. The root player creates a pool of arms whose posterior distribution on mean rewards is beta, allowing the peripheral players achieve better regret (by a constant multiplicative factor) by pulling from this pool of arms. Through coordination, the players are on average able to beat the lower bound from the single player setting.en_US
dc.format.extent39en_US
dc.language.isoen_USen_US
dc.titleInfinite-Armed Bandits with Multiple Playersen_US
dc.typePrinceton University Senior Theses-
pu.date.classyear2014en_US
pu.departmentOperations Research and Financial Engineeringen_US
Appears in Collections:Operations Research and Financial Engineering, 2000-2020

Files in This Item:
File SizeFormat 
Fong, Christian Final Thesis.pdf469.45 kBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.