Projects that are tagged with bandits.


Logo contextual 0.9.8.2

by robinvanemden - July 9, 2019, 14:22:28 CET [ Project Homepage BibTeX Download ] 3004 views, 957 downloads, 2 subscriptions

About: R package facilitating the simulation and evaluation of context-free and contextual Multi-Armed Bandit policies.

Changes:

Minor update


Logo SMPyBandits 0.9.2

by Naereen - March 20, 2018, 20:12:13 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3804 views, 716 downloads, 0 subscriptions

About: SMPyBandits: an Open-Source Research Framework for Single and Multi-Players Multi-Arms Bandits (MAB) Algorithms in Python

Changes:

Initial Announcement on mloss.org.


Logo Obandit 0.2

by fre - November 6, 2017, 14:33:02 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2382 views, 544 downloads, 0 subscriptions

About: Obandit is an Ocaml module for multi-armed bandits. It supports the EXP, UCB and Epsilon-greedy family of algorithms.

Changes:

Initial Announcement on mloss.org.


Logo BayesOpt, a Bayesian Optimization toolbox 0.8.2

by rmcantin - December 9, 2015, 04:53:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 45532 views, 8170 downloads, 0 subscriptions

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python.

Changes:

-Fixed bug in save/restore. -Fixed bug in initial design.