Multi-Armed Bandit Networks: Exploring Online Learning with Networks
Master thesis
Permanent lenke
http://hdl.handle.net/1956/18665Utgivelsesdato
2018-06-26Metadata
Vis full innførselSamlinger
Sammendrag
Classical Multi-Armed Bandit solutions often assumes independent arms as a simplification of the problem. This has shown great results in many different fields of practice, but could in some cases, presumably leave untapped potential. In this paper I explore network based MAB solutions using explore-exploit algorithms as nodes to further minimize regret, and take advantage of inter-Bandit dependencies. I explore two network approaches; Hierarchical and Flat network. As well as a special cases of the Bernoulli Bandit with dependent arms, referred to as Symbiotic Bandit. The results show that some networked solutions prevail the single node versions in both the Bernoulli Bandit and the Symbiotic Bandit regret wise.