Projects that are tagged with bandits.


Logo contextual 0.9.8.4

by robinvanemden - July 27, 2020, 16:05:32 CET [ Project Homepage BibTeX Download ] 29422 views, 7257 downloads, 2 subscriptions

About: R package facilitating the simulation and evaluation of context-free and contextual Multi-Armed Bandit policies.

Changes:

Minor update.


Logo SMPyBandits 0.9.2

by Naereen - March 20, 2018, 20:12:13 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 21296 views, 3085 downloads, 0 subscriptions

About: SMPyBandits: an Open-Source Research Framework for Single and Multi-Players Multi-Arms Bandits (MAB) Algorithms in Python

Changes:

Initial Announcement on mloss.org.


Logo Obandit 0.2

by fre - November 6, 2017, 14:33:02 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7683 views, 1776 downloads, 0 subscriptions

About: Obandit is an Ocaml module for multi-armed bandits. It supports the EXP, UCB and Epsilon-greedy family of algorithms.

Changes:

Initial Announcement on mloss.org.


Logo BayesOpt, a Bayesian Optimization toolbox 0.8.2

by rmcantin - December 9, 2015, 04:53:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 77198 views, 15831 downloads, 0 subscriptions

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python.

Changes:

-Fixed bug in save/restore. -Fixed bug in initial design.