dc.contributor.author | Hansen, Viktor | |
dc.date.accessioned | 2018-10-22T16:30:39Z | |
dc.date.available | 2018-10-22T16:30:39Z | |
dc.date.issued | 2018-06-26 | |
dc.date.submitted | 2018-06-25T22:00:11Z | |
dc.identifier.uri | http://hdl.handle.net/1956/18665 | |
dc.description.abstract | Classical Multi-Armed Bandit solutions often assumes independent arms as a simplification of the problem. This has shown great results in many different fields of practice, but could in some cases, presumably leave untapped potential. In this paper I explore network based MAB solutions using explore-exploit algorithms as nodes to further minimize regret, and take advantage of inter-Bandit dependencies. I explore two network approaches; Hierarchical and Flat network. As well as a special cases of the Bernoulli Bandit with dependent arms, referred to as Symbiotic Bandit. The results show that some networked solutions prevail the single node versions in both the Bernoulli Bandit and the Symbiotic Bandit regret wise. | en_US |
dc.language.iso | eng | eng |
dc.publisher | The University of Bergen | eng |
dc.subject | Networks | eng |
dc.subject | Online Learning | eng |
dc.subject | MAB | eng |
dc.title | Multi-Armed Bandit Networks: Exploring Online Learning with Networks | eng |
dc.type | Master thesis | en_US |
dc.date.updated | 2018-06-25T22:00:11Z | |
dc.rights.holder | Copyright the author. All rights reserved. | en_US |
dc.description.degree | Masteroppgave i informasjonsvitenskap | |
dc.description.localcode | INFO390 | |
dc.subject.nus | 735115 | eng |
fs.subjectcode | INFO390 | |
fs.unitcode | 15-17-0 | |