Skip to content

meezys/Bernoulli-Bandits

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Simple implementations of Bandit Algorithms, specifically for Bernoulli Bandits.
The implemenentation includes methods such as Explore-Then-Commit, Greedy, Thompson Sampling, UCB, MOSS, AdaUCB, KL-UCB.
Uncomment and alter the line at the bottom to specify the experiment(s) desired
There's also a link to my Google Colab: https://colab.research.google.com/drive/1yz5aTEi3j84E9Z0X4QN8CeAjNdvzgcNa?usp=sharing
Implementations are based on pseudocode given from the textbooks 'A Tutorial on Thompson Sampling'(https://arxiv.org/abs/1707.02038) and 'Bandit Algorithms' (1.Lattimore T, Szepesvári C. Stochastic Bandits with Finitely Many Arms. In: Bandit Algorithms. Cambridge University Press; 2020:73-74.)

About

An implementation of Bandit Algorithms, focusing on the case of Bernoulli Rewards.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages