About: This toolbox provides functions for maximizing and minimizing submodular set functions, with applications to Bayesian experimental design, inference in Markov Random Fields, clustering and others. Changes:

About: DiffSharp is a functional automatic differentiation (AD) library providing gradients, Hessians, Jacobians, directional derivatives, and matrixfree Hessian and Jacobianvector products as higherorder functions. It allows exact and efficient calculation of derivatives, with support for nesting. Changes:Fixed: Bug fix in forward AD implementation of Sigmoid and ReLU for D, DV, and DM (fixes #16, thank you @mrakgr) Improvement: Performance improvement by removing several more Parallel.For and Array.Parallel.map operations, working better with OpenBLAS multithreading Added: Operations involving incompatible dimensions of DV and DM will now throw exceptions for warning the user

About: The SUMO Toolbox is a Matlab toolbox that automatically builds accurate surrogate models (also known as metamodels or response surface models) of a given data source (e.g., simulation code, data set, script, ...) within the accuracy and time constraints set by the user. The toolbox minimizes the number of data points (which it selects automatically) since they are usually expensive. Changes:Incremental update, fixing some cosmetic issues, coincides with JMLR publication.

About: Generalized linear and additive models by likelihood based boosting Changes:Fetched by rcranrobot on 20130401 00:00:04.893311

About: C++ software for statistical classification, probability estimation and interpolation/nonlinear regression using variable bandwidth kernel estimation. Changes:New in Version 0.9.8:

About: Breiman and Cutler's random forests for classification and regression Changes:Fetched by rcranrobot on 20130401 00:00:07.638240

About: xgboost: eXtreme Gradient Boosting It is an efficient and scalable implementation of gradient boosting framework. The package includes efficient linear model solver and tree learning algorithm. The package can automatically do parallel computation with OpenMP, and it can be more than 10 times faster than existing gradient boosting packages such as gbm or sklearn.GBM . It supports various objective functions, including regression, classification and ranking. The package is made to be extensible, so that user are also allowed to define there own objectives easily. The newest version of xgboost now supports distributed learning on various platforms such as hadoop, mpi and scales to even larger problems Changes:

About: This package contains a python and a matlab implementation of the most widely used algorithms for multiarmed bandit problems. The purpose of this package is to provide simple environments for comparison and numerical evaluation of policies. Changes:Initial Announcement on mloss.org.

About: The Maja Machine Learning Framework (MMLF) is a general framework for problems in the domain of Reinforcement Learning (RL) written in python. It provides a set of RL related algorithms and a set of benchmark domains. Furthermore it is easily extensible and allows to automate benchmarking of different agents. Changes:

About: Pyriel is a Python system for learning classification rules from data. Unlike other rule learning systems, it is designed to learn rule lists that maximize the area under the ROC curve (AUC) instead of accuracy. Pyriel is mostly an experimental research tool, but it's robust and fast enough to be used for lightweight industrial data mining. Changes:1.5 Changed CF (confidence factor) to do LaPlace smoothing of estimates. New flag "scoreforclass C" causes scores to be computed relative to a given (positive) class. For twoclass problems. Fixed bug in example sampling code (sample n) Fixed bug keeping oldstyle example formats (terminated by dot) from working. More code restructuring.
