All entries.
Showing Items 411-420 of 537 on page 42 of 54: First Previous 37 38 39 40 41 42 43 44 45 46 47 Next Last

Logo Jubatus 0.5.0

by hido - November 30, 2013, 17:41:50 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1982 views, 352 downloads, 1 subscription

About: Jubatus is a general framework library for online and distributed machine learning. It currently supports classification, regression, clustering, recommendation, nearest neighbors, anomaly detection, and graph analysis. Loose model sharing provides higher scalability, better performance, and real-time capabilities, by combining online learning with distributed computations.

Changes:

0.5.0 add new supports for clustering and nearest neighbors. For more detail, see http://t.co/flMcTcYZVs


Logo Efficient Nonnegative Sparse Coding Algorithm 1.0

by openpr_nlpr - January 4, 2012, 09:44:18 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1977 views, 422 downloads, 1 subscription

About: Nonnegative Sparse Coding, Discriminative Semi-supervised Learning, sparse probability graph

Changes:

Initial Announcement on mloss.org.


Logo PLEASD 0.1

by heroesneverdie - September 10, 2012, 03:53:26 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1954 views, 459 downloads, 1 subscription

About: PLEASD: A Matlab Toolbox for Structured Learning

Changes:

Initial Announcement on mloss.org.


Logo Naive Bayes Classifier 1.0.0

by openpr_nlpr - December 2, 2011, 05:25:44 CET [ Project Homepage BibTeX Download ] 1934 views, 496 downloads, 1 subscription

About: This program is a C++ implementation of Naive Bayes Classifier, which is a well-known generative classification algorithm for the application such as text classification. The Naive Bayes algorithm requires the probabilistic distribution to be discrete. The program uses the multinomial event model for representation, the maximum likelihood estimate with a Laplace smoothing technique for learning parameters. A sparse-data structure is defined to represent the feature vector in the program to seek higher computational speed.

Changes:

Initial Announcement on mloss.org.


Logo Ankus 0.0.1

by suhyunjeon - September 13, 2013, 06:47:46 CET [ Project Homepage BibTeX Download ] 1894 views, 236 downloads, 1 subscription

About: Ankus is an open source data mining / machine learning based MapReduce that supports a variety of advanced algorithms.

Changes:

Initial Announcement on mloss.org.


Logo r-cran-quantregForest 0.2-3

by r-cran-robot - June 1, 2012, 00:00:00 CET [ Project Homepage BibTeX Download ] 1888 views, 446 downloads, 0 subscriptions

About: Quantile Regression Forests

Changes:

Fetched by r-cran-robot on 2013-04-01 00:00:07.576421


Logo MLFlex 02-21-2012-00-12

by srp33 - April 3, 2012, 16:44:43 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1874 views, 378 downloads, 1 subscription

About: Motivated by a need to classify high-dimensional, heterogeneous data from the bioinformatics domain, we developed ML-Flex, a machine-learning toolbox that enables users to perform two-class and multi-class classification analyses in a systematic yet flexible manner. ML-Flex was written in Java but is capable of interfacing with third-party packages written in other programming languages. It can handle multiple input-data formats and supports a variety of customizations. MLFlex provides implementations of various validation strategies, which can be executed in parallel across multiple computing cores, processors, and nodes. Additionally, ML-Flex supports aggregating evidence across multiple algorithms and data sets via ensemble learning. (See http://jmlr.csail.mit.edu/papers/volume13/piccolo12a/piccolo12a.pdf.)

Changes:

Initial Announcement on mloss.org.


Logo r-cran-LogicForest 2.0.0

by r-cran-robot - September 20, 2011, 00:00:00 CET [ Project Homepage BibTeX Download ] 1862 views, 417 downloads, 0 subscriptions

About: Logic Forest

Changes:

Fetched by r-cran-robot on 2013-04-01 00:00:06.077571


Logo pyGPs 1.2

by mn - July 17, 2014, 10:28:55 CET [ Project Homepage BibTeX Download ] 1839 views, 441 downloads, 2 subscriptions

About: pyGPs is a Python package for Gaussian process (GP) regression and classification for machine learning.

Changes:

Changelog pyGPs v1.2

June 30th 2014

structural updates:

  • input target now can either be in 2-d array with size (n,1) or in 1-d array with size (n,)
  • setup.py updated
  • "import pyGPs" instead of "from pyGPs.Core import gp"
  • rename ".train()" to ".optimize()"
  • rename "Graph-stuff" to "graphExtension"
  • rename kernelOnGraph to "nodeKernels" and graphKernel to "graphKernels"
  • redundancy removed for model.setData(x,y)
  • rewrite "mean.proceed()" to "getMean()" and "getDerMatrix()"
  • rewrite "cov.proceed()" to "getCovMatrix()" and "getDerMatrix()"
  • rename cov.LIN to cov.Linear (to be consistent with mean.Linear)
  • rename module "valid" to "validation"
  • add graph dataset Mutag in python file. (.npz and .mat)
  • add graphUtil.nomalizeKernel()
  • fix number of iteration problem in graphKernels "PropagationKernel"
  • add unit testing for covariance, mean functions

bug fixes:

  • derivatives for cov.LINard
  • derivative of the scalar for cov.covScale
  • demo_GPR_FITC.py missing pyGPs.mean

July 8th 2014

structural updates:

  • add hyperparameter(signal variance s2) for linear covariance
  • add unit testing for inference,likelihood functions as well as models
  • NOT show(print) "maximum number of sweep warning in inference EP" any more
  • documentation updated

bug fixes:

  • typos in lik.Laplace
  • derivative in lik.Laplace

July 14th 2014

documentation updates:

  • online docs updated
  • API file updated

structural updates:

  • made private for methods that users don't need to call

About: In this paper, we propose an improved principal component analysis based on maximum entropy (MaxEnt) preservation, called MaxEnt-PCA, which is derived from a Parzen window estimation of Renyi’s quadratic entropy. Instead of minimizing the reconstruction error either based on L2-norm or L1-norm, the MaxEnt-PCA attempts to preserve as much as possible the uncertainty information of the data measured by entropy. The optimal solution of MaxEnt-PCA consists of the eigenvectors of a Laplacian probability matrix corresponding to the MaxEnt distribution. MaxEnt-PCA (1) is rotation invariant, (2) is free from any distribution assumption, and (3) is robust to outliers. Extensive experiments on real-world datasets demonstrate the effectiveness of the proposed linear method as compared to other related robust PCA methods.

Changes:

Initial Announcement on mloss.org.


Showing Items 411-420 of 537 on page 42 of 54: First Previous 37 38 39 40 41 42 43 44 45 46 47 Next Last