About: A MATLAB toolbox for defining complex machine learning comparisons Changes:Initial Announcement on mloss.org.
|
About: Soltion developed by team Turtle Tamers in the ChaLearn Gesture Challenge (http://www.kaggle.com/c/GestureChallenge2) Changes:Initial Announcement on mloss.org.
|
About: This program is a C++ implementation of Naive Bayes Classifier, which is a well-known generative classification algorithm for the application such as text classification. The Naive Bayes algorithm requires the probabilistic distribution to be discrete. The program uses the multinomial event model for representation, the maximum likelihood estimate with a Laplace smoothing technique for learning parameters. A sparse-data structure is defined to represent the feature vector in the program to seek higher computational speed. Changes:Initial Announcement on mloss.org.
|
About: An implementation of MROGH descriptor. For more information, please refer to: “Bin Fan, Fuchao Wu and Zhanyi Hu, Aggregating Gradient Distributions into Intensity Orders: A Novel Local Image Descriptor, CVPR 2011, pp.2377-2384.” The most up-to-date information can be found at : http://vision.ia.ac.cn/Students/bfan/index.htm Changes:Initial Announcement on mloss.org.
|
About: Regression forests, Random Forests for regression. Original implementation by Leo Breiman. Changes:Initial Announcement on mloss.org.
|
About: A library for calculating and accessing generalized Stirling numbers of the second kind, which are used for inference in Poisson-Dirichlet processes. Changes:Initial Announcement on mloss.org.
|
About: FsAlg is a linear algebra library that supports generic types. Changes:Initial Announcement on mloss.org.
|
About: ALgebraic COmbinatorial COmpletion of MAtrices. A collection of algorithms to impute or denoise single entries in an incomplete rank one matrix, to determine for which entries this is possible with any algorithm, and to provide algorithm-independent error estimates. Includes demo scripts. Changes:Initial Announcement on mloss.org.
|
About: A Theano framework for building and training neural networks Changes:Initial Announcement on mloss.org.
|
About: A method to optimize the hyperparameters for machine learning methods implemented in Scikit-learn based on Derivative Free Optimization Changes:Initial Announcement on mloss.org.
|
About: Code for Calibrated AdaMEC for binary cost-sensitive classification. The method is just AdaBoost that properly calibrates its probability estimates and uses a cost-sensitive (i.e. risk-minimizing) decision threshold to classify new data. Changes:Updated license information
|
About: The original Random Forests implementation by Breiman and Cutler. Changes:Initial Announcement on mloss.org.
|
About: Recur is a collection of Gstreamer plugins and language modelling tools based on recurrent neural networks. Changes:Initial Announcement on mloss.org.
|
About: This is the core MCMC sampler for the nonparametric sparse factor analysis model presented in David A. Knowles and Zoubin Ghahramani (2011). Nonparametric Bayesian Sparse Factor Models with application to Gene Expression modelling. Annals of Applied Statistics Changes:Initial Announcement on mloss.org.
|
About: This library implements the Optimum-Path Forest classifier for unsupervised and supervised learning. Changes:Initial Announcement on mloss.org.
|
About: Python toolbox for manifold optimization with support for automatic differentiation Changes:Initial Announcement on mloss.org.
|
About: In this paper, we propose an improved principal component analysis based on maximum entropy (MaxEnt) preservation, called MaxEnt-PCA, which is derived from a Parzen window estimation of Renyi’s quadratic entropy. Instead of minimizing the reconstruction error either based on L2-norm or L1-norm, the MaxEnt-PCA attempts to preserve as much as possible the uncertainty information of the data measured by entropy. The optimal solution of MaxEnt-PCA consists of the eigenvectors of a Laplacian probability matrix corresponding to the MaxEnt distribution. MaxEnt-PCA (1) is rotation invariant, (2) is free from any distribution assumption, and (3) is robust to outliers. Extensive experiments on real-world datasets demonstrate the effectiveness of the proposed linear method as compared to other related robust PCA methods. Changes:Initial Announcement on mloss.org.
|
About: SALSA (Software lab for Advanced machine Learning with Stochastic Algorithms) is an implementation of the well-known stochastic algorithms for Machine Learning developed in the high-level technical computing language Julia. The SALSA software package is designed to address challenges in sparse linear modelling, linear and non-linear Support Vector Machines applied to large data samples with user-centric and user-friendly emphasis. Changes:Initial Announcement on mloss.org.
|
About: Log-linear analysis for high-dimensional data Changes:Initial Announcement on mloss.org.
|
About: Robust sparse representation has shown significant potential in solving challenging problems in computer vision such as biometrics and visual surveillance. Although several robust sparse models have been proposed and promising results have been obtained, they are either for error correction or for error detection, and learning a general framework that systematically unifies these two aspects and explore their relation is still an open problem. In this paper, we develop a half-quadratic (HQ) framework to solve the robust sparse representation problem. By defining different kinds of half-quadratic functions, the proposed HQ framework is applicable to performing both error correction and error detection. More specifically, by using the additive form of HQ, we propose an L1-regularized error correction method by iteratively recovering corrupted data from errors incurred by noises and outliers; by using the multiplicative form of HQ, we propose an L1-regularized error detection method by learning from uncorrupted data iteratively. We also show that the L1-regularization solved by soft-thresholding function has a dual relationship to Huber M-estimator, which theoretically guarantees the performance of robust sparse representation in terms of M-estimation. Experiments on robust face recognition under severe occlusion and corruption validate our framework and findings. Changes:Initial Announcement on mloss.org.
|