Projects that are tagged with bandits.


Logo contextual 0.9.8.4

by robinvanemden - July 27, 2020, 16:05:32 CET [ Project Homepage BibTeX Download ] 18088 views, 4252 downloads, 2 subscriptions

About: R package facilitating the simulation and evaluation of context-free and contextual Multi-Armed Bandit policies.

Changes:

Minor update.


Logo SMPyBandits 0.9.2

by Naereen - March 20, 2018, 20:12:13 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 11564 views, 2146 downloads, 0 subscriptions

About: SMPyBandits: an Open-Source Research Framework for Single and Multi-Players Multi-Arms Bandits (MAB) Algorithms in Python

Changes:

Initial Announcement on mloss.org.


Logo Obandit 0.2

by fre - November 6, 2017, 14:33:02 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 5080 views, 1085 downloads, 0 subscriptions

About: Obandit is an Ocaml module for multi-armed bandits. It supports the EXP, UCB and Epsilon-greedy family of algorithms.

Changes:

Initial Announcement on mloss.org.


Logo BayesOpt, a Bayesian Optimization toolbox 0.8.2

by rmcantin - December 9, 2015, 04:53:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 61721 views, 11477 downloads, 0 subscriptions

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python.

Changes:

-Fixed bug in save/restore. -Fixed bug in initial design.