About: A method to optimize the hyperparameters for machine learning methods implemented in Scikit-learn based on Derivative Free Optimization Changes:Initial Announcement on mloss.org.
|
About: A non-iterative, incremental and hyperparameter-free learning method for one-layer feedforward neural networks without hidden layers. This method efficiently obtains the optimal parameters of the network, regardless of whether the data contains a greater number of samples than variables or vice versa. It does this by using a square loss function that measures errors before the output activation functions and scales them by the slope of these functions at each data point. The outcome is a system of linear equations that obtain the network's weights and that is further transformed using Singular Value Decomposition. Changes:Initial Announcement on mloss.org.
|
About: A non-iterative learning method for one-layer (no hidden layer) neural networks, where the weights can be calculated in a closed-form manner, thereby avoiding low convergence rate and also hyperparameter tuning. The proposed learning method, LANN-SVD in short, presents a good computational efficiency for large-scale data analytic. Changes:Initial Announcement on mloss.org.
|
About: A Matlab benchmarking toolbox for online and adaptive regression with kernels. Changes:
|
About: Scalable learning of global, multi-task and local metrics from data Changes:Minor bug fix in multi-task objective computation (thanks to Junjie Hu).
|
About: An extensible C++ library of Hierarchical Bayesian clustering algorithms, such as Bayesian Gaussian mixture models, variational Dirichlet processes, Gaussian latent Dirichlet allocation and more. Changes:New maximum cluster argument for all algorithms. Also no more matlab interface since it seemed no one was using it, and I cannot support it any longer.
|
About: SALSA (Software lab for Advanced machine Learning with Stochastic Algorithms) is an implementation of the well-known stochastic algorithms for Machine Learning developed in the high-level technical computing language Julia. The SALSA software package is designed to address challenges in sparse linear modelling, linear and non-linear Support Vector Machines applied to large data samples with user-centric and user-friendly emphasis. Changes:Initial Announcement on mloss.org.
|
About: Simple and hopefully clean and easy to follow implementation of the Generalized Learning Vector Quantizer (GLVQ) with variants for metric adaptation (RGLVQ, GMLVQ, LiRaM). Changes:Initial Announcement on mloss.org.
|
About: Distributed optimization: Support Vector Machines and LASSO regression on distributed data Changes:Initial Upload
|
About: Low-Rank and Sparse tools for Background Modeling and Subtraction in Videos. Changes:Initial Announcement on mloss.org.
|
About: A MATLAB toolkit for performing generalized regression with equality/inequality constraints on the function value/gradient. Changes:Initial Announcement on mloss.org.
|
About: LIBOL is an open-source library with a family of state-of-the-art online learning algorithms for machine learning and big data analytics research. The current version supports 16 online algorithms for binary classification and 13 online algorithms for multiclass classification. Changes:In contrast to our last version (V0.2.3), the new version (V0.3.0) has made some important changes as follows: • Add a template and guide for adding new algorithms; • Improve parameter settings and make documentation clear; • Improve documentation on data formats and key functions; • Amend the "OGD" function to use different loss types; • Fixed some name inconsistency and other minor bugs.
|
About: Probabilistic performance evaluation for multiclass classification using the posterior balanced accuracy Changes:Added bibtex information.
|
About: GPgrid toolkit for fast GP analysis on grid input Changes:Initial Announcement on mloss.org.
|
About: Fast Multidimensional GP Inference using Projected Additive Approximation Changes:Initial Announcement on mloss.org.
|
About: Soltion developed by team Turtle Tamers in the ChaLearn Gesture Challenge (http://www.kaggle.com/c/GestureChallenge2) Changes:Initial Announcement on mloss.org.
|
About: PLEASD: A Matlab Toolbox for Structured Learning Changes:Initial Announcement on mloss.org.
|
About: Message passing for topic modeling Changes:
|
About: This local and parallel computation toolbox is the Octave and Matlab implementation of several localized Gaussian process regression methods: the domain decomposition method (Park et al., 2011, DDM), partial independent conditional (Snelson and Ghahramani, 2007, PIC), localized probabilistic regression (Urtasun and Darrell, 2008, LPR), and bagging for Gaussian process regression (Chen and Ren, 2009, BGP). Most of the localized regression methods can be applied for general machine learning problems although DDM is only applicable for spatial datasets. In addition, the GPLP provides two parallel computation versions of the domain decomposition method. The easiness of being parallelized is one of the advantages of the localized regression, and the two parallel implementations will provide a good guidance about how to materialize this advantage as software. Changes:Initial Announcement on mloss.org.
|