All entries.
Showing Items 301-310 of 674 on page 31 of 68: First Previous 26 27 28 29 30 31 32 33 34 35 36 Next Last

Logo AISAIC 1.0.0610

by fydennis - June 13, 2013, 21:54:55 CET [ BibTeX Download ] 4554 views, 1773 downloads, 1 subscription

About: AISAIC software for analyzing human DNA copy numbers and detecting significant copy number alterations

Changes:

Initial Announcement on mloss.org.


About: This letter proposes a new multiple linear regression model using regularized correntropy for robust pattern recognition. First, we motivate the use of correntropy to improve the robustness of the classicalmean square error (MSE) criterion that is sensitive to outliers. Then an l1 regularization scheme is imposed on the correntropy to learn robust and sparse representations. Based on the half-quadratic optimization technique, we propose a novel algorithm to solve the nonlinear optimization problem. Second, we develop a new correntropy-based classifier based on the learned regularization scheme for robust object recognition. Extensive experiments over several applications confirm that the correntropy-based l1 regularization can improve recognition accuracy and receiver operator characteristic curves under noise corruption and occlusion.

Changes:

Initial Announcement on mloss.org.


About: Robust sparse representation has shown significant potential in solving challenging problems in computer vision such as biometrics and visual surveillance. Although several robust sparse models have been proposed and promising results have been obtained, they are either for error correction or for error detection, and learning a general framework that systematically unifies these two aspects and explore their relation is still an open problem. In this paper, we develop a half-quadratic (HQ) framework to solve the robust sparse representation problem. By defining different kinds of half-quadratic functions, the proposed HQ framework is applicable to performing both error correction and error detection. More specifically, by using the additive form of HQ, we propose an L1-regularized error correction method by iteratively recovering corrupted data from errors incurred by noises and outliers; by using the multiplicative form of HQ, we propose an L1-regularized error detection method by learning from uncorrupted data iteratively. We also show that the L1-regularization solved by soft-thresholding function has a dual relationship to Huber M-estimator, which theoretically guarantees the performance of robust sparse representation in terms of M-estimation. Experiments on robust face recognition under severe occlusion and corruption validate our framework and findings.

Changes:

Initial Announcement on mloss.org.


Logo r-cran-ahaz 1.14

by r-cran-robot - June 3, 2013, 00:00:00 CET [ Project Homepage BibTeX Download ] 14495 views, 2993 downloads, 0 subscriptions

About: Regularization for semiparametric additive hazards regression

Changes:

Fetched by r-cran-robot on 2018-05-01 00:00:04.091406


About: A fast and robust learning of Bayesian networks

Changes:

Initial Announcement on mloss.org.


Logo HLearn 1.0

by mikeizbicki - May 9, 2013, 05:58:18 CET [ Project Homepage BibTeX Download ] 8259 views, 1999 downloads, 1 subscription

About: HLearn makes simple machine learning routines available in Haskell by expressing them according to their algebraic structure

Changes:

Updated to version 1.0


Logo OptWok 0.3.1

by ong - May 2, 2013, 10:46:11 CET [ Project Homepage BibTeX Download ] 15422 views, 3046 downloads, 1 subscription

About: A collection of python code to perform research in optimization. The aim is to provide reusable components that can be quickly applied to machine learning problems. Used in: - Ellipsoidal multiple instance learning - difference of convex functions algorithms for sparse classfication - Contextual bandits upper confidence bound algorithm (using GP) - learning output kernels, that is kernels between the labels of a classifier.

Changes:
  • minor bugfix

Logo KNIME 2.7.4

by toldo - April 29, 2013, 09:14:39 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8814 views, 1469 downloads, 1 subscription

About: A comprehensive data mining environment, with a variety of machine learning components.

Changes:

Modifications following feedback from Knime main Author.


Logo Intelligent Parameter Utilization Tool 0.4

by feldob - April 28, 2013, 18:05:45 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 4712 views, 1190 downloads, 1 subscription

About: A descriptive and programming language independent format and API for the simplified configuration, documentation, and design of computer experiments.

Changes:

Initial Announcement on mloss.org.


Logo HDDM 0.5

by Wiecki - April 24, 2013, 02:53:07 CET [ Project Homepage BibTeX Download ] 9293 views, 2146 downloads, 1 subscription

About: HDDM is a python toolbox for hierarchical Bayesian parameter estimation of the Drift Diffusion Model (via PyMC). Drift Diffusion Models are used widely in psychology and cognitive neuroscience to study decision making.

Changes:
  • New and improved HDDM model with the following changes:
    • Priors: by default model will use informative priors (see http://ski.clps.brown.edu/hddm_docs/methods.html#hierarchical-drift-diffusion-models-used-in-hddm) If you want uninformative priors, set informative=False.
    • Sampling: This model uses slice sampling which leads to faster convergence while being slower to generate an individual sample. In our experiments, burnin of 20 is often good enough.
    • Inter-trial variablity parameters are only estimated at the group level, not for individual subjects.
    • The old model has been renamed to HDDMTransformed.
    • HDDMRegression and HDDMStimCoding are also using this model.
  • HDDMRegression takes patsy model specification strings. See http://ski.clps.brown.edu/hddm_docs/howto.html#estimate-a-regression-model and http://ski.clps.brown.edu/hddm_docs/tutorial_regression_stimcoding.html#chap-tutorial-hddm-regression
  • Improved online documentation at http://ski.clps.brown.edu/hddm_docs
  • A new HDDM demo at http://ski.clps.brown.edu/hddm_docs/demo.html
  • Ratcliff's quantile optimization method for single subjects and groups using the .optimize() method
  • Maximum likelihood optimization.
  • Many bugfixes and better test coverage.
  • hddm_fit.py command line utility is depracated.

Showing Items 301-310 of 674 on page 31 of 68: First Previous 26 27 28 29 30 31 32 33 34 35 36 Next Last