About: The GPML toolbox is a flexible and generic Octave 3.2.x and Matlab 7.x implementation of inference and prediction in Gaussian Process (GP) models. Changes:

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinearoptimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python. Changes:Fixed bugs and doc typos

About: Toeblitz is a MATLAB/Octave package for operations on positive definite Toeplitz matrices. It can solve Toeplitz systems Tx = b in O(n*log(n)) time and O(n) memory, compute matrix inverses T^(1) (with free log determinant) in O(n^2) time and memory, compute log determinants (without inverses) in O(n^2) time and O(n) memory, and compute traces of products A*T for any matrix A, in minimal O(n^2) time and memory. Changes:Adding a writeup in written/toeblitz.pdf describing the package.

About: The GPstuff toolbox is a versatile collection of Gaussian process models and computational tools required for inference. The tools include, among others, various inference methods, sparse approximations and model assessment methods. Changes:20140722 Version 4.5 New features
Improvements
Several minor bugfixes

About: This package is an implementation of a linear RankSVM solver with nonconvex regularization. Changes:Initial Announcement on mloss.org.

About: A mutual information library for C and Mex bindings for MATLAB. Aimed at feature selection, and provides simple methods to calculate mutual information, conditional mutual information, entropy, conditional entropy, Renyi entropy/mutual information, and weighted variants of Shannon entropies/mutual informations. Works with discrete distributions, and expects column vectors of features. Changes:Added weighted entropy functions. Fixed a few memory handling bugs.

About: ITE (Information Theoretical Estimators) is capable of estimating many different variants of entropy, mutual information, divergence, association measures, cross quantities and kernels on distributions. Thanks to its highly modular design, ITE supports additionally (i) the combinations of the estimation techniques, (ii) the easy construction and embedding of novel information theoretical estimators, and (iii) their immediate application in information theoretical optimization problems. Changes:

About: The package computes the optimal parameters for the Choquet kernel Changes:Initial Announcement on mloss.org.

About: LIBOL is an opensource library with a family of stateoftheart online learning algorithms for machine learning and big data analytics research. The current version supports 16 online algorithms for binary classification and 13 online algorithms for multiclass classification. Changes:In contrast to our last version (V0.2.3), the new version (V0.3.0) has made some important changes as follows: • Add a template and guide for adding new algorithms; • Improve parameter settings and make documentation clear; • Improve documentation on data formats and key functions; • Amend the "OGD" function to use different loss types; • Fixed some name inconsistency and other minor bugs.

About: The glmie toolbox contains scalable estimation routines for GLMs (generalised linear models) and SLMs (sparse linear models) as well as an implementation of a scalable convex variational Bayesian inference relaxation. We designed the glmie package to be simple, generic and easily expansible. Most of the code is written in Matlab including some MEX files. The code is fully compatible to both Matlab 7.x and GNU Octave 3.2.x. Probabilistic classification, sparse linear modelling and logistic regression are covered in a common algorithmical framework allowing for both MAP estimation and approximate Bayesian inference. Changes:added factorial mean field inference as a third algorithm complementing expectation propagation and variational Bayes generalised nonGaussian potentials so that affine instead of linear functions of the latent variables can be used

About: ALgebraic COmbinatorial COmpletion of MAtrices. A collection of algorithms to impute or denoise single entries in an incomplete rank one matrix, to determine for which entries this is possible with any algorithm, and to provide algorithmindependent error estimates. Includes demo scripts. Changes:Initial Announcement on mloss.org.

About: GPgrid toolkit for fast GP analysis on grid input Changes:Initial Announcement on mloss.org.

About: Fast Multidimensional GP Inference using Projected Additive Approximation Changes:Initial Announcement on mloss.org.

About: Stochastic neighbor embedding originally aims at the reconstruction of given distance relations in a lowdimensional Euclidean space. This can be regarded as general approach to multidimensional scaling, but the reconstruction is based on the definition of input (and output) neighborhood probability alone. The present implementation also allows for handling dissimilarity or scoreinduced neighborhood topologies and makes use of quasi 2nd order gradientbased (l)BFGS optimization. Changes:

About: Approximate Rank One FACtorization of tensors. An algorithm for factorization of threewaytensors and determination of their rank, includes example applications. Changes:Initial Announcement on mloss.org.

About: The aim is to embed a given data relationship matrix into a lowdimensional Euclidean space such that the point distances / distance ranks correlate best with the original input relationships. Input relationships may be given as (sparse) (asymmetric) distance, dissimilarity, or (negative!) score matrices. Inputoutput relations are modeled as lowconditioned. (Weighted) Pearson and soft Spearman rank correlation, and unweighted soft Kendall correlation are supported correlation measures for input/output object neighborhood relationships. Changes:

About: libDAI provides free & open source implementations of various (approximate) inference methods for graphical models with discrete variables, including Bayesian networks and Markov Random Fields. Changes:Release 0.3.1 fixes various bugs. The issues on 64bit Windows platforms have been fixed and libDAI now offers full 64bit support on all supported platforms (Linux, Mac OSX, Windows).

About: The package provides a Lagrangian approach to the posterior regularization of given linear mappings. This is important in two cases, (a) when systems are underdetermined and (b) when the external model for calculating the mapping is invariant to properties such as scaling. The software may be applied in cases when the external model does not provide its own regularization strategy. In addition, the package allows to rank attributes according to their distortion potential to a given linear mapping. Changes:Version 1.1 (May 23, 2012) memory and time optimizations distderivrel.m now supports assessing the relevance of attribute pairs Version 1.0 (Nov 9, 2011) * Initial Announcement on mloss.org.

About: Message passing for topic modeling Changes:

About: This local and parallel computation toolbox is the Octave and Matlab implementation of several localized Gaussian process regression methods: the domain decomposition method (Park et al., 2011, DDM), partial independent conditional (Snelson and Ghahramani, 2007, PIC), localized probabilistic regression (Urtasun and Darrell, 2008, LPR), and bagging for Gaussian process regression (Chen and Ren, 2009, BGP). Most of the localized regression methods can be applied for general machine learning problems although DDM is only applicable for spatial datasets. In addition, the GPLP provides two parallel computation versions of the domain decomposition method. The easiness of being parallelized is one of the advantages of the localized regression, and the two parallel implementations will provide a good guidance about how to materialize this advantage as software. Changes:Initial Announcement on mloss.org.
