About: This project is a C++ toolkit containing machine learning algorithms and tools for creating complex software in C++ to solve real world problems. Changes:This release adds support for cuDNN 5.1 as well as a number of minor bug fixes and usability improvements.

About: Kernelbased Learning Platform (KeLP) is Java framework that supports the implementation of kernelbased learning algorithms, as well as an agile definition of kernel functions over generic data representation, e.g. vectorial data or discrete structures. The framework has been designed to decouple kernel functions and learning algorithms, through the definition of specific interfaces. Once a new kernel function has been implemented, it can be automatically adopted in all the available kernelmachine algorithms. KeLP includes different Online and Batch Learning algorithms for Classification, Regression and Clustering, as well as several Kernel functions, ranging from vectorbased to structural kernels. It allows to build complex kernel machine based systems, leveraging on JSON/XML interfaces to instantiate prediction models without writing a single line of code. Changes:In addition to minor bug fixes, this release includes:
Check out this new version from our repositories. API Javadoc is already available. Your suggestions will be very precious for us, so download and try KeLP 2.1.0!

About: MLweb is an open source project that aims at bringing machine learning capabilities into web pages and web applications, while maintaining all computations on the client side. It includes (i) a javascript library to enable scientific computing within web pages, (ii) a javascript library implementing machine learning algorithms for classification, regression, clustering and dimensionality reduction, (iii) a web application providing a matlablike development environment. Changes:

About: ADENINE (A Data ExploratioN pIpeliNE) is a machine learning framework for data exploration that encompasses stateoftheart techniques for missing values imputing, data preprocessing, dimensionality reduction and clustering tasks. Changes:Initial Announcement on mloss.org.

About: The Weka workbench contains a collection of visualization tools and algorithms for data analysis and predictive modelling, together with graphical user interfaces for easy access to this [...] Changes:In core weka:
In packages:

About: ELKI is a framework for implementing datamining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods. Changes:Additions and improvements from ELKI 0.7.0 to 0.7.1: Algorithm additions:
Important bug fixes:
UI improvements:
Smaller changes:

About: An extensible C++ library of Hierarchical Bayesian clustering algorithms, such as Bayesian Gaussian mixture models, variational Dirichlet processes, Gaussian latent Dirichlet allocation and more. Changes:New maximum cluster argument for all algorithms. Also no more matlab interface since it seemed no one was using it, and I cannot support it any longer.

About: The apcluster package implements Frey's and Dueck's Affinity Propagation clustering in R. The package further provides leveraged affinity propagation, exemplarbased agglomerative clustering, and various tools for visual analysis of clustering results. Changes:

About: Apache Mahout is an Apache Software Foundation project with the goal of creating both a community of users and a scalable, Javabased framework consisting of many machine learning algorithm [...] Changes:Apache Mahout introduces a new math environment we call Samsara, for its theme of universal renewal. It reflects a fundamental rethinking of how scalable machine learning algorithms are built and customized. MahoutSamsara is here to help people create their own math while providing some offtheshelf algorithm implementations. At its core are general linear algebra and statistical operations along with the data structures to support them. You can use is as a library or customize it in Scala with Mahoutspecific extensions that look something like R. MahoutSamsara comes with an interactive shell that runs distributed operations on a Spark cluster. This make prototyping or task submission much easier and allows users to customize algorithms with a whole new degree of freedom. Mahout Algorithms include many new implementations built for speed on MahoutSamsara. They run on Spark 1.3+ and some on H2O, which means as much as a 10x speed increase. You’ll find robust matrix decomposition algorithms as well as a Naive Bayes classifier and collaborative filtering. The new sparkitemsimilarity enables the next generation of cooccurrence recommenders that can use entire user click streams and context in making recommendations.

About: The Cognitive Foundry is a modular Java software library of machine learning components and algorithms designed for research and applications. Changes:

About: SALSA (Software lab for Advanced machine Learning with Stochastic Algorithms) is an implementation of the wellknown stochastic algorithms for Machine Learning developed in the highlevel technical computing language Julia. The SALSA software package is designed to address challenges in sparse linear modelling, linear and nonlinear Support Vector Machines applied to large data samples with usercentric and userfriendly emphasis. Changes:Initial Announcement on mloss.org.

About: Learning MWay Tree  Web Scale Clustering  EMtree, Ktree, kmeans, TSVQ, repeated kmeans, clustering, random projections, random indexing, hashing, bit signatures Changes:Initial Announcement on mloss.org.

About: Cluster quality Evaluation software. Implements cluster quality metrics based on ground truths such as Purity, Entropy, Negentropy, F1 and NMI. It includes a novel approach to correct for pathological or ineffective clusterings called 'Divergence from a Random Baseline'. Changes:Moved project to GitHub.

About: The autoencoder based data clustering toolkit provides a quick start of clustering based on deep autoencoder nets. This toolkit can cluster data in feature space with a deep nonlinear nets. Changes:Initial Announcement on mloss.org.

About: Hubnessaware Machine Learning for Highdimensional Data Changes:

About: The Accord.NET Framework is a .NET machine learning framework combined with audio and image processing libraries completely written in C#. It is a complete framework for building productiongrade computer vision, computer audition, signal processing and statistics applications even for commercial use. A comprehensive set of sample applications provide a fast start to get up and running quickly, and an extensive online documentation helps fill in the details. Changes:Adding a large number of new distributions, such as AndersonDaring, ShapiroWilk, Inverse ChiSquare, Lévy, Folded Normal, Shifted LogLogistic, Kumaraswamy, Trapezoidal, Uquadratic and BetaPrime distributions, BirnbaumSaunders, Generalized Normal, Gumbel, Power Lognormal, Power Normal, Triangular, Tukey Lambda, Logistic, Hyperbolic Secant, Degenerate and General Continuous distributions. Other additions include new statistical hypothesis tests such as AndersonDaring and ShapiroWilk; as well as support for all of LIBLINEAR's support vector machine algorithms; and format reading support for MATLAB/Octave matrices, LibSVM models, sparse LibSVM data files, and many others. For a complete list of changes, please see the full release notes at the release details page at: https://github.com/accordnet/framework/releases

About: C++ software for statistical classification, probability estimation and interpolation/nonlinear regression using variable bandwidth kernel estimation. Changes:New in Version 0.9.8:

About: STK++: A Statistical Toolkit Framework in C++ Changes:Inegrating openmp to the current release. Many enhancement in the clustering project. bug fix

About: A generalized version of spectral clustering using the graph pLaplacian. Changes:

About: DRVQ is a C++ library implementation of dimensionalityrecursive vector quantization, a fast vector quantization method in highdimensional Euclidean spaces under arbitrary data distributions. It is an approximation of kmeans that is practically constant in data size and applies to arbitrarily high dimensions but can only scale to a few thousands of centroids. As a byproduct of training, a tree structure performs either exact or approximate quantization on trained centroids, the latter being not very precise but extremely fast. Changes:Initial Announcement on mloss.org.
