About: Somoclu is a massively parallel implementation of self-organizing maps. It relies on OpenMP for multicore execution, MPI for distributing the workload, and it can be accelerated by CUDA on a GPU cluster. A sparse kernel is also included, which is useful for training maps on vector spaces generated in text mining processes. Apart from a command line interface, Python, Julia, R, and MATLAB are supported. Changes:
|
About: A non-iterative, incremental and hyperparameter-free learning method for one-layer feedforward neural networks without hidden layers. This method efficiently obtains the optimal parameters of the network, regardless of whether the data contains a greater number of samples than variables or vice versa. It does this by using a square loss function that measures errors before the output activation functions and scales them by the slope of these functions at each data point. The outcome is a system of linear equations that obtain the network's weights and that is further transformed using Singular Value Decomposition. Changes:Initial Announcement on mloss.org.
|
About: DIANNE is a modular software framework for designing, training and evaluating artificial neural networks on heterogeneous, distributed infrastructure . It is built on top of OSGi and AIOLOS and can transparently deploy and redeploy (parts of) a neural network on multiple machines, as well as scale up training on a compute cluster. Changes:Initial Announcement on mloss.org.
|
About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python. Changes:-Fixed bug in save/restore. -Fixed bug in initial design.
|
About: The DL-Learner framework contains several algorithms for supervised concept learning in Description Logics (DLs) and OWL. Changes:See http://dl-learner.org/development/changelog/.
|
About: Universal Python-written numerical optimization toolbox. Problems: NLP, LP, QP, NSP, MILP, LSP, LLSP, MMP, GLP, SLE, MOP etc; general logical constraints, categorical variables, automatic differentiation, stochastic programming, interval analysis, many other goodies Changes:http://openopt.org/Changelog
|
About: MultiBoost is a multi-purpose boosting package implemented in C++. It is based on the multi-class/multi-task AdaBoost.MH algorithm [Schapire-Singer, 1999]. Basic base learners (stumps, trees, products, Haar filters for image processing) can be easily complemented by new data representations and the corresponding base learners, without interfering with the main boosting engine. Changes:Major changes :
Minor fixes:
|
About: Embarrassingly Parallel Array Computing: EPAC is a machine learning workflow builder. Changes:Initial Announcement on mloss.org.
|
About: A library for artificial neural networks. Changes:Added algorithms:
|
About: The Rchemcpp package implements the marginalized graph kernel and extensions, Tanimoto kernels, graph kernels, pharmacophore and 3D kernels suggested for measuring the similarity of molecules. Changes:Moved from CRAN to Bioconductor. Improved handling of molecules, visualization and examples.
|
About: LIBSVM is an integrated software for support vector classification, (C-SVC, nu-SVC ), regression (epsilon-SVR, nu-SVR) and distribution estimation (one-class SVM). It supports multi-class [...] Changes:Initial Announcement on mloss.org.
|
About: A MATLAB spectral clustering package to deal with large data sets. Our tool can handle large data sets (200,000 RCV1 data) on a 4GB memory general machine. Spectral clustering algorithm has been [...] Changes:
|
About: A Java library to create, process and manage mixtures of exponential families. Changes:Initial Announcement on mloss.org.
|
About: CRFSuite is a speed-oriented implementation of Conditional Random Fields (CRFs). This software features: parameter estimation using SGD and L-BFGS, l1/l2 regularization, simple data I/O format, etc. Changes:Initial Announcement on mloss.org.
|
About: LIBLINEAR is a linear classifier for data with millions of instances and features. It supports L2-regularized logistic regression (LR), L2-loss linear SVM, L1-loss linear SVM, and multi-class SVM Changes:Initial Announcement on mloss.org.
|
About: We provide some preliminary code for multiclass multiple kernel learning in Matlab using CPLEX as a base solver. Changes:Initial Announcement on mloss.org.
|
About: FlexCRFs is a conditional random field toolkit for segmenting and labeling sequence data written in C/C++ using STL library. It was implemented based on the theoretic model presented in (Lafferty et [...] Changes:Initial Announcement on mloss.org.
|
About: GibbsLDA++: A C/C++ Implementation of Latent Dirichlet Allocation (LDA) using Gibbs Sampling for parameter estimation and inference. GibbsLDA++ is fast and is designed to analyze hidden/latent topic [...] Changes:Initial Announcement on mloss.org.
|
About: BSVM solves support vector machines (SVM) for the solution of large classification and regression problems. It includes three methods Changes:Initial Announcement on mloss.org.
|
About: Code for automatically selecting the kernel parameters of an SVM. It is based on a gradient descent minimization of either the radius/margin bound, the leave-one-out error, a validation error or the [...] Changes:Initial Announcement on mloss.org.
|