Showing Items 281290 of 550 on page 29 of 55: First Previous 24 25 26 27 28 29 30 31 32 33 34 Next Last
About: Nonnegative Sparse Coding, Discriminative Semisupervised Learning, sparse probability graph Changes:Initial Announcement on mloss.org.

About: The KernelMachine Library is a free (released under the LGPL) C++ library to promote the use of and progress of kernel machines. Changes:Updated mloss entry (minor fixes).

About: Python module to ease pattern classification analyses of large datasets. It provides highlevel abstraction of typical processing steps (e.g. data preparation, classification, feature selection, [...] Changes:
This release aggregates all the changes occurred between official
releases in 0.4 series and various snapshot releases (in 0.5 and 0.6
series). To get better overview of high level changes see
:ref:
Also adapts changes from 0.4.6 and 0.4.7 (see corresponding changelogs).
This is a special release, because it has never seen the general public.
A summary of fundamental changes introduced in this development version
can be seen in the :ref: Most notably, this version was to first to come with a comprehensive twoday workshop/tutorial.
A bugfix release
A bugfix release

About: An annotated java framework for machine learning, aimed at making it really easy to access analytically functions. Changes:Now supports OLS and GLS regression and NaiveBayes classification

About: Pam Changes:Fetched by rcranrobot on 20130401 00:00:06.709586

About: In this paper, we propose an improved principal component analysis based on maximum entropy (MaxEnt) preservation, called MaxEntPCA, which is derived from a Parzen window estimation of Renyi’s quadratic entropy. Instead of minimizing the reconstruction error either based on L2norm or L1norm, the MaxEntPCA attempts to preserve as much as possible the uncertainty information of the data measured by entropy. The optimal solution of MaxEntPCA consists of the eigenvectors of a Laplacian probability matrix corresponding to the MaxEnt distribution. MaxEntPCA (1) is rotation invariant, (2) is free from any distribution assumption, and (3) is robust to outliers. Extensive experiments on realworld datasets demonstrate the effectiveness of the proposed linear method as compared to other related robust PCA methods. Changes:Initial Announcement on mloss.org.

About: MetropolisHastings alogrithm is a Markov chain Monte Carlo method for obtaining a sequence of random samples from a probability distribution for which direct sampling is difficult. Thi sequence can be used to approximate the distribution. Changes:Initial Announcement on mloss.org.

About: This code is developed based on Uriel Roque's active set algorithm for the linear least squares problem with nonnegative variables in: Portugal, L.; Judice, J.; and Vicente, L. 1994. A comparison of block pivoting and interiorpoint algorithms for linear least squares problems with nonnegative variables. Mathematics of Computation 63(208):625643.Ran He, WeiShi Zheng and Baogang Hu, "Maximum Correntropy Criterion for Robust Face Recognition," IEEE TPAMI, in press, 2011. Changes:Initial Announcement on mloss.org.

About: Urheen is a toolkit for Chinese word segmentation, Chinese pos tagging, English tokenize, and English pos tagging. The Chinese word segmentation and pos tagging modules are trained with the Chinese Tree Bank 7.0. The English pos tagging module is trained with the WSJ English treebank(0223). Changes:Initial Announcement on mloss.org.

About: OpenPRNBEM is an C++ implementation of Naive Bayes Classifier, which is a wellknown generative classification algorithm for the application such as text classification. The Naive Bayes algorithm requires the probabilistic distribution to be discrete. OpenPRNBEM uses the multinomial event model for representation. The maximum likelihood estimate is used for supervised learning, and the expectationmaximization estimate is used for semisupervised and unsupervised learning. Changes:Initial Announcement on mloss.org.
