About: The Accord.NET Framework is a .NET machine learning framework combined with audio and image processing libraries completely written in C#. It is a complete framework for building productiongrade computer vision, computer audition, signal processing and statistics applications even for commercial use. A comprehensive set of sample applications provide a fast start to get up and running quickly, and an extensive online documentation helps fill in the details. Changes:Adding a large number of new distributions, such as AndersonDaring, ShapiroWilk, Inverse ChiSquare, Lévy, Folded Normal, Shifted LogLogistic, Kumaraswamy, Trapezoidal, Uquadratic and BetaPrime distributions, BirnbaumSaunders, Generalized Normal, Gumbel, Power Lognormal, Power Normal, Triangular, Tukey Lambda, Logistic, Hyperbolic Secant, Degenerate and General Continuous distributions. Other additions include new statistical hypothesis tests such as AndersonDaring and ShapiroWilk; as well as support for all of LIBLINEAR's support vector machine algorithms; and format reading support for MATLAB/Octave matrices, LibSVM models, sparse LibSVM data files, and many others. For a complete list of changes, please see the full release notes at the release details page at: https://github.com/accordnet/framework/releases

About: The GPML toolbox is a flexible and generic Octave 3.2.x and Matlab 7.x implementation of inference and prediction in Gaussian Process (GP) models. Changes:

About: STK++: A Statistical Toolkit Framework in C++ Changes:Inegrating openmp to the current release. Many enhancement in the clustering project. bug fix

About: RLPy is a framework for performing reinforcement learning (RL) experiments in Python. RLPy provides a large library of agent and domain components, and a suite of tools to aid in experiments (parallelization, hyperparameter optimization, code profiling, and plotting). Changes:

About: Toeblitz is a MATLAB/Octave package for operations on positive definite Toeplitz matrices. It can solve Toeplitz systems Tx = b in O(n*log(n)) time and O(n) memory, compute matrix inverses T^(1) (with free log determinant) in O(n^2) time and memory, compute log determinants (without inverses) in O(n^2) time and O(n) memory, and compute traces of products A*T for any matrix A, in minimal O(n^2) time and memory. Changes:Adding a writeup in written/toeblitz.pdf describing the package.

About: The GPstuff toolbox is a versatile collection of Gaussian process models and computational tools required for inference. The tools include, among others, various inference methods, sparse approximations and model assessment methods. Changes:20140722 Version 4.5 New features
Improvements
Several minor bugfixes

About: Crino: a neuralnetwork library based on Theano Changes:1.0.0 (7 july 2014) :  Initial release of crino  Implements a torchlike library to build artificial neural networks (ANN)  Provides standard implementations for : * autoencoders * multilayer perceptrons (MLP) * deep neural networks (DNN) * input output deep architecture (IODA)  Provides a batchgradient backpropagation algorithm, with adaptative learning rate

About: A Matlab benchmarking toolbox for online and adaptive regression with kernels. Changes:

About: RLLib is a lightweight C++ template library that implements incremental, standard, and gradient temporaldifference learning algorithms in Reinforcement Learning. It is an optimized library for robotic applications and embedded devices that operates under fast duty cycles (e.g., < 30 ms). RLLib has been tested and evaluated on RoboCup 3D soccer simulation agents, physical NAO V4 humanoid robots, and Tiva C series launchpad microcontrollers to predict, control, learn behaviors, and represent learnable knowledge. The implementation of the RLLib library is inspired by the RLPark API, which is a library of temporaldifference learning algorithms written in Java. Changes:Current release version is v2.0.

About: Loglinear analysis for highdimensional data Changes:Initial Announcement on mloss.org.

About: MOSIS is a modularized framework for signal processing, stream analysis, machine learning and stream mining applications. Changes:

About: The package computes the optimal parameters for the Choquet kernel Changes:Initial Announcement on mloss.org.

About: "Ordinal Choquistic Regression" model using the maximum likelihood Changes:Initial Announcement on mloss.org.

About: ELKI is a framework for implementing datamining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods. Changes:Additions and Improvements from ELKI 0.5.5: Algorithms Clustering:
Outlier detection
Distances
Database Layer and Data Types Projection layer * Parser for simple textual data (for use with Levenshtein distance) Various random projection families (including Feature Bagging, Achlioptas, and pstable) Latitude+Longitude to ECEF Sparse vector improvements and bug fixes New filter: remove NaN values and missing values New filter: add histogrambased jitter New filter: normalize using statistical distributions New filter: robust standardization using Median and MAD New filter: Linear discriminant analysis (LDA) Index Layer
Mathematics and Statistics
Visualization
Other

About: minFunc is a Matlab function for unconstrained optimization of differentiable realvalued multivariate functions using linesearch methods. It uses an interface very similar to the Matlab Optimization Toolbox function fminunc, and can be called as a replacement for this function. On many problems, minFunc requires fewer function evaluations to converge than fminunc (or minimize.m). Further it can optimize problems with a much larger number of variables (fminunc is restricted to several thousand variables), and uses a line search that is robust to several common function pathologies. Changes:Initial Announcement on mloss.org.

About: LIBOL is an opensource library with a family of stateoftheart online learning algorithms for machine learning and big data analytics research. The current version supports 16 online algorithms for binary classification and 13 online algorithms for multiclass classification. Changes:In contrast to our last version (V0.2.3), the new version (V0.3.0) has made some important changes as follows: • Add a template and guide for adding new algorithms; • Improve parameter settings and make documentation clear; • Improve documentation on data formats and key functions; • Amend the "OGD" function to use different loss types; • Fixed some name inconsistency and other minor bugs.

About: The glmie toolbox contains scalable estimation routines for GLMs (generalised linear models) and SLMs (sparse linear models) as well as an implementation of a scalable convex variational Bayesian inference relaxation. We designed the glmie package to be simple, generic and easily expansible. Most of the code is written in Matlab including some MEX files. The code is fully compatible to both Matlab 7.x and GNU Octave 3.2.x. Probabilistic classification, sparse linear modelling and logistic regression are covered in a common algorithmical framework allowing for both MAP estimation and approximate Bayesian inference. Changes:added factorial mean field inference as a third algorithm complementing expectation propagation and variational Bayes generalised nonGaussian potentials so that affine instead of linear functions of the latent variables can be used

About: FABIA is a biclustering algorithm that clusters rows and columns of a matrix simultaneously. Consequently, members of a row cluster are similar to each other on a subset of columns and, analogously, members of a column cluster are similar to each other on a subset of rows. Biclusters are found by factor analysis where both the factors and the loading matrix are sparse. FABIA is a multiplicative model that extracts linear dependencies between samples and feature patterns. Applications include detection of transcriptional modules in gene expression data and identification of haplotypes/>identity by descent< consisting of rare variants obtained by next generation sequencing. Changes:CHANGES IN VERSION 2.8.0NEW FEATURES
CHANGES IN VERSION 2.4.0
CHANGES IN VERSION 2.3.1NEW FEATURES
2.0.0:
1.4.0:

About: MLlib provides a distributed machine learning (ML) library to address the growing need for scalable ML. MLlib is developed in Spark (http://spark.incubator.apache.org/), a cluster computing system designed for iterative computation. Moreover, it is a component of a larger system called MLbase (www.mlbase.org) that aims to provide userfriendly distributed ML functionality both for ML researchers and domain experts. MLlib currently consists of scalable implementations of algorithms for classification, regression, collaborative filtering and clustering. Changes:Initial Announcement on mloss.org.

About: This package includes implementations of the CCM, DMV and DMV+CCM parsers from Klein and Manning (2004), and code for testing them with the WSJ, Negra and Cast3LB corpuses (English, German and Spanish respectively). A detailed description of the parsers can be found in Klein (2005). Changes:Initial Announcement on mloss.org.
