20 projects found that use matlab as the programming language.
Showing Items 1-20 of 156 on page 1 of 8: 1 2 3 4 5 6 Next Last

Logo Somoclu 1.7.2

by peterwittek - November 24, 2016, 22:43:48 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 21587 views, 3920 downloads, 3 subscriptions

About: Somoclu is a massively parallel implementation of self-organizing maps. It relies on OpenMP for multicore execution, MPI for distributing the workload, and it can be accelerated by CUDA on a GPU cluster. A sparse kernel is also included, which is useful for training maps on vector spaces generated in text mining processes. Apart from a command line interface, Python, Julia, R, and MATLAB are supported.

Changes:
  • New: The coefficient of the Gaussian neighborhood function exp(-||x-y||^2/(2(coeffradius)^2)) is now exposed in all interfaces as a parameter.
  • New: get_bmu function in the Python interface to get the best matching units given an activation map.
  • Changed: Updated PCA initialization in the Python interface to work with sk-learn 0.18 onwards.
  • Changed: Radii can be float values.
  • Fixed: Only positive values were written back to codebook during update.
  • Fixed: Sparse data is read correctly when there are class labels.

Logo JMLR GPML Gaussian Processes for Machine Learning Toolbox 4.0

by hn - October 19, 2016, 10:15:05 CET [ Project Homepage BibTeX Download ] 38756 views, 8692 downloads, 5 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: The GPML toolbox is a flexible and generic Octave/Matlab implementation of inference and prediction with Gaussian process models. The toolbox offers exact inference, approximate inference for non-Gaussian likelihoods (Laplace's Method, Expectation Propagation, Variational Bayes) as well for large datasets (FITC, VFE, KISS-GP). A wide range of covariance, likelihood, mean and hyperprior functions allows to create very complex GP models.

Changes:

A major code restructuring effort did take place in the current release unifying certain inference functions and allowing more flexibility in covariance function composition. We also redesigned the whole derivative computation pipeline to strongly improve the overall runtime. We finally include grid-based covariance approximations natively.

More generic sparse approximation using Power EP

  • unified treatment of FITC approximation, variational approaches VFE and hybrids

  • inducing input optimisation for all (compositions of) covariance functions dropping the previous limitation to a few standard examples

  • infFITC is now covered by the more generic infGaussLik function

Approximate covariance object unifying sparse approximations, grid-based approximations and exact covariance computations

  • implementation in cov/apx, cov/apxGrid, cov/apxSparse

  • generic infGaussLik unifies infExact, infFITC and infGrid

  • generic infLaplace unifies infLaplace, infFITC_Laplace and infGrid_Laplace

Hiearchical structure of covariance functions

  • clear hierachical compositional implementation

  • no more code duplication as present in covSEiso and covSEard pairs

  • two mother covariance functions

    • covDot for dot-product-based covariances and

    • covMaha for Mahalanobis-distance-based covariances

  • a variety of modifiers: eye, iso, ard, proj, fact, vlen

  • more flexibility as more variants are available and possible

  • all covariance functions offer derivatives w.r.t. inputs

Faster derivative computations for mean and cov functions

  • switched from partial derivatives to directional derivatives

  • simpler and more concise interface of mean and cov functions

  • much faster marginal likelihood derivative computations

  • simpler and more compact code

New mean functions

  • new mean/meanWSPC (Weighted Sum of Projected Cosines or Random Kitchen Sink features) following a suggestion by William Herlands

  • new mean/meanWarp for constructing a new mean from an existing one by means of a warping function adapted from William Herlands

New optimizer

  • added a new minimize_minfunc, contributed by Truong X. Nghiem

New GLM link function

  • added the twice logistic link function util/glm_invlink_logistic2

Smaller fixes

  • two-fold speedup of util/elsympol used by covADD by Truong X. Nghiem

  • bugfix in util/logphi as reported by John Darby


Logo slim for matlab 0.2

by ustunb - August 23, 2016, 20:27:00 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1712 views, 256 downloads, 3 subscriptions

About: learn optimized scoring systems using MATLAB and the CPLEX Optimization Studio

Changes:

Initial Announcement on mloss.org.


Logo NaN toolbox 3.0.3

by schloegl - August 19, 2016, 11:08:57 CET [ Project Homepage BibTeX Download ] 52891 views, 10687 downloads, 3 subscriptions

About: NaN-toolbox is a statistics and machine learning toolbox for handling data with and without missing values.

Changes:

Changes in v.3.0.3 - improve compatibility for Octave on Windows

Changes in v.3.0.1 - fix packaging for octave

Changes in v.2.8.5 - bug fix: trimmean - compiler support for gcc-5 and clang - fix typos

For details see the CHANGELOG at http://pub.ist.ac.at/~schloegl/matlab/NaN/CHANGELOG


Logo Social Impact theory based Optimizer library 1.1

by rishem - July 29, 2016, 13:19:47 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8933 views, 1787 downloads, 2 subscriptions

About: This is an optimization library based on Social Impact Theory(SITO). The optimizer works in the same way as PSO and GA.

Changes:

bug removed


About: Nowadays this is very popular to use the deep architectures in machine learning. Deep Belief Networks (DBNs) are deep architectures that use a stack of Restricted Boltzmann Machines (RBM) to create a powerful generative model using training data. DBNs have many abilities such as feature extraction and classification that are used in many applications including image processing, speech processing, text categorization, etc. This paper introduces a new object oriented toolbox with the most important abilities needed for the implementation of DBNs. According to the results of the experiments conducted on the MNIST (image), ISOLET (speech), and the 20 Newsgroups (text) datasets, it was shown that the toolbox can learn automatically a good representation of the input from unlabeled data with better discrimination between different classes. Also on all the aforementioned datasets, the obtained classification errors are comparable to those of the state of the art classifiers. In addition, the toolbox supports different sampling methods (e.g. Gibbs, CD, PCD and our new FEPCD method), different sparsity methods (quadratic, rate distortion and our new normal method), different RBM types (generative and discriminative), GPU based, etc. The toolbox is a user-friendly open source software in MATLAB and Octave and is freely available on the website.

Changes:

New in toolbox

  • Using GPU in Backpropagation
  • Revision of some demo scripts
  • Function approximation with multiple outputs
  • Feature extraction with GRBM in first layer

cardinal


Logo JMLR Information Theoretical Estimators 0.63

by szzoli - June 9, 2016, 23:42:14 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 117255 views, 21744 downloads, 3 subscriptions

About: ITE (Information Theoretical Estimators) is capable of estimating many different variants of entropy, mutual information, divergence, association measures, cross quantities and kernels on distributions. Thanks to its highly modular design, ITE supports additionally (i) the combinations of the estimation techniques, (ii) the easy construction and embedding of novel information theoretical estimators, and (iii) their immediate application in information theoretical optimization problems.

Changes:
  • Conditional Shannon entropy estimation: added.

  • Conditional Shannon mutual information estimation: included.


Logo JMLR GPstuff 4.7

by avehtari - June 9, 2016, 17:45:15 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 38220 views, 9202 downloads, 3 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 1 vote)

About: The GPstuff toolbox is a versatile collection of Gaussian process models and computational tools required for inference. The tools include, among others, various inference methods, sparse approximations and model assessment methods.

Changes:

2016-06-09 Version 4.7

Development and release branches available at https://github.com/gpstuff-dev/gpstuff

New features

  • Simple Bayesian Optimization demo

Improvements

  • Improved use of PSIS
  • More options added to gp_monotonic
  • Monotonicity now works for additive covariance functions with selected variables
  • Possibility to use gpcf_squared.m-covariance function with derivative observations/monotonicity
  • Default behaviour made more robust by changing default jitter from 1e-9 to 1e-6
  • LA-LOO uses the cavity method as the default (see Vehtari et al (2016). Bayesian leave-one-out cross-validation approximations for Gaussian latent variable models. JMLR, accpeted for publication)
  • Selected variables -option works now better with monotonicity

Bugfixes

  • small error in derivative observation computation fixed
  • several minor bug fixes

Logo FEAST 1.1.4

by apocock - March 12, 2016, 18:35:08 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 39134 views, 7093 downloads, 2 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: FEAST provides implementations of common mutual information based filter feature selection algorithms (mim, mifs, mrmr, cmim, icap, jmi, disr, fcbf, etc), and an implementation of RELIEF. Written for C/C++ & Matlab.

Changes:
  • Fixed an issue where zero MI values would cause it to segfault.
  • Fixes to documentation and comments.
  • Updated internal version of MIToolbox.

Logo MDLText 1

by renatoms88 - March 3, 2016, 19:31:25 CET [ BibTeX Download ] 739 views, 287 downloads, 2 subscriptions

About: testing mloss.org

Changes:

Initial Announcement on mloss.org.


Logo Local high order regularization 1.0

by kkim - March 2, 2016, 13:46:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1261 views, 321 downloads, 2 subscriptions

Rating Empty StarEmpty StarEmpty StarEmpty StarEmpty Star
(based on 1 vote)

About: Local high-order regularization for semi-supervised learning

Changes:

Initial Announcement on mloss.org.


Logo MIToolbox 2.1.2

by apocock - January 10, 2016, 22:19:30 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 28978 views, 4931 downloads, 2 subscriptions

About: A mutual information library for C and Mex bindings for MATLAB. Aimed at feature selection, and provides simple methods to calculate mutual information, conditional mutual information, entropy, conditional entropy, Renyi entropy/mutual information, and weighted variants of Shannon entropies/mutual informations. Works with discrete distributions, and expects column vectors of features.

Changes:

Relicensed as BSD. Added checks to catch MATLAB inputs that aren't doubles.


Logo Sparse PCA 2.0

by tbuehler - December 31, 2015, 16:24:42 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 6624 views, 1513 downloads, 3 subscriptions

About: A Matlab implementation of Sparse PCA using the inverse power method for nonlinear eigenproblems.

Changes:
  • Added deflation scheme to compute multiple principal components
  • Several internal runtime and memory optimizations
  • API change: sparsePCA.m is now used to compute multiple components; use computeTradeOffCurve.m to reproduce the examples in the NIPS paper

Logo BayesOpt, a Bayesian Optimization toolbox 0.8.2

by rmcantin - December 9, 2015, 04:53:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 27276 views, 5097 downloads, 4 subscriptions

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python.

Changes:

-Fixed bug in save/restore. -Fixed bug in initial design.


About: This MATLAB package provides the MLAPG algorithm proposed in our ICCV 2015 paper. It is efficient for PSD constrained metric learning, and also effective for person re-identification. For more details, please visit http://www.cbsr.ia.ac.cn/users/scliao/projects/mlapg/.

Changes:

Initial Announcement on mloss.org.


Logo A Library for Online Streaming Feature Selection 1.0

by ykui713 - November 25, 2015, 13:23:01 CET [ BibTeX Download ] 1149 views, 462 downloads, 1 subscription

About: LOFS is a software toolbox for online streaming feature selection

Changes:

Initial Announcement on mloss.org.


Logo NPD Face Detector Training 1.0

by openpr_nlpr - October 8, 2015, 04:22:36 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2490 views, 386 downloads, 2 subscriptions

About: This MATLAB package provides the Deep Quadratic Tree (DQT) and the Normalized Pixel Difference (NPD) based face detector training method proposed in our PAMI 2015 paper. It is fast, and effective for unconstrained face detection. For more details, please visit http://www.cbsr.ia.ac.cn/users/scliao/projects/npdface/.

Changes:

Initial Announcement on mloss.org.


Logo Optunity 1.1.1

by claesenm - September 30, 2015, 07:06:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7817 views, 1807 downloads, 3 subscriptions

About: Optunity is a library containing various optimizers for hyperparameter tuning. Hyperparameter tuning is a recurrent problem in many machine learning tasks, both supervised and unsupervised.This package provides several distinct approaches to solve such problems including some helpful facilities such as cross-validation and a plethora of score functions.

Changes:

This minor release has the same feature set as Optunity 1.1.0, but incorporates several bug fixes, mostly related to the specification of structured search spaces.


Logo Chalearn gesture challenge code by jun wan 2.0

by joewan - September 29, 2015, 08:50:22 CET [ BibTeX BibTeX for corresponding Paper Download ] 6852 views, 1610 downloads, 2 subscriptions

About: This code is provided by Jun Wan. It is used in the Chalearn one-shot learning gesture challenge (round 2). This code includes: bag of features, 3D MoSIFT-based features (i.e. 3D MoSIFT, 3D EMoSIFT and 3D SMoSIFT), and the MFSK feature.

Changes:

Initial Announcement on mloss.org.


Logo JMLR libDAI 0.3.2

by jorism - July 17, 2015, 15:59:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 53025 views, 9885 downloads, 4 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 1 vote)

About: libDAI provides free & open source implementations of various (approximate) inference methods for graphical models with discrete variables, including Bayesian networks and Markov Random Fields.

Changes:

Release 0.3.2 fixes various bugs and adds GLC (Generalized Loop Corrections) written by Siamak Ravanbakhsh.


Showing Items 1-20 of 156 on page 1 of 8: 1 2 3 4 5 6 Next Last