All entries.
Showing Items 41-50 of 640 on page 5 of 64: Previous 1 2 3 4 5 6 7 8 9 10 Next Last

Logo DIANNE 0.5.0

by sbohez - October 25, 2016, 19:51:07 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1408 views, 248 downloads, 3 subscriptions

About: DIANNE is a modular software framework for designing, training and evaluating artificial neural networks on heterogeneous, distributed infrastructure . It is built on top of OSGi and AIOLOS and can transparently deploy and redeploy (parts of) a neural network on multiple machines, as well as scale up training on a compute cluster.

Changes:

Initial Announcement on mloss.org.


Logo JMLR GPML Gaussian Processes for Machine Learning Toolbox 4.0

by hn - October 19, 2016, 10:15:05 CET [ Project Homepage BibTeX Download ] 42705 views, 9474 downloads, 5 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: The GPML toolbox is a flexible and generic Octave/Matlab implementation of inference and prediction with Gaussian process models. The toolbox offers exact inference, approximate inference for non-Gaussian likelihoods (Laplace's Method, Expectation Propagation, Variational Bayes) as well for large datasets (FITC, VFE, KISS-GP). A wide range of covariance, likelihood, mean and hyperprior functions allows to create very complex GP models.

Changes:

A major code restructuring effort did take place in the current release unifying certain inference functions and allowing more flexibility in covariance function composition. We also redesigned the whole derivative computation pipeline to strongly improve the overall runtime. We finally include grid-based covariance approximations natively.

More generic sparse approximation using Power EP

  • unified treatment of FITC approximation, variational approaches VFE and hybrids

  • inducing input optimisation for all (compositions of) covariance functions dropping the previous limitation to a few standard examples

  • infFITC is now covered by the more generic infGaussLik function

Approximate covariance object unifying sparse approximations, grid-based approximations and exact covariance computations

  • implementation in cov/apx, cov/apxGrid, cov/apxSparse

  • generic infGaussLik unifies infExact, infFITC and infGrid

  • generic infLaplace unifies infLaplace, infFITC_Laplace and infGrid_Laplace

Hiearchical structure of covariance functions

  • clear hierachical compositional implementation

  • no more code duplication as present in covSEiso and covSEard pairs

  • two mother covariance functions

    • covDot for dot-product-based covariances and

    • covMaha for Mahalanobis-distance-based covariances

  • a variety of modifiers: eye, iso, ard, proj, fact, vlen

  • more flexibility as more variants are available and possible

  • all covariance functions offer derivatives w.r.t. inputs

Faster derivative computations for mean and cov functions

  • switched from partial derivatives to directional derivatives

  • simpler and more concise interface of mean and cov functions

  • much faster marginal likelihood derivative computations

  • simpler and more compact code

New mean functions

  • new mean/meanWSPC (Weighted Sum of Projected Cosines or Random Kitchen Sink features) following a suggestion by William Herlands

  • new mean/meanWarp for constructing a new mean from an existing one by means of a warping function adapted from William Herlands

New optimizer

  • added a new minimize_minfunc, contributed by Truong X. Nghiem

New GLM link function

  • added the twice logistic link function util/glm_invlink_logistic2

Smaller fixes

  • two-fold speedup of util/elsympol used by covADD by Truong X. Nghiem

  • bugfix in util/logphi as reported by John Darby


Logo AMIDST Toolbox 0.6.0

by ana - October 14, 2016, 19:35:27 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 6354 views, 1126 downloads, 4 subscriptions

About: A Java Toolbox for Scalable Probabilistic Machine Learning.

Changes:
  • Added sparklink module implementing the integration with Apache Spark. More information here.
  • Fluent pattern in latent-variable-models
  • Predefined model implementing the concept drift detection

Detailed information can be found in the toolbox's web page


Logo RLScore 0.7

by aatapa - September 20, 2016, 09:51:25 CET [ Project Homepage BibTeX Download ] 1318 views, 296 downloads, 3 subscriptions

About: RLScore - regularized least-squares machine learning algorithms package

Changes:

Initial Announcement on mloss.org.


Logo r-cran-bst 0.3-14

by r-cran-robot - September 12, 2016, 00:00:00 CET [ Project Homepage BibTeX Download ] 7595 views, 1758 downloads, 2 subscriptions

About: Gradient Boosting

Changes:

Fetched by r-cran-robot on 2017-05-01 00:00:03.015285


Logo slim for matlab 0.2

by ustunb - August 23, 2016, 20:27:00 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2298 views, 384 downloads, 3 subscriptions

About: learn optimized scoring systems using MATLAB and the CPLEX Optimization Studio

Changes:

Initial Announcement on mloss.org.


Logo Sparse Compositional Metric Learning v1.11

by bellet - August 2, 2016, 11:43:03 CET [ BibTeX BibTeX for corresponding Paper Download ] 6079 views, 1807 downloads, 3 subscriptions

About: Scalable learning of global, multi-task and local metrics from data

Changes:

Minor bug fix in multi-task objective computation (thanks to Junjie Hu).


Logo Social Impact theory based Optimizer library 1.1

by rishem - July 29, 2016, 13:19:47 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 10104 views, 1999 downloads, 2 subscriptions

About: This is an optimization library based on Social Impact Theory(SITO). The optimizer works in the same way as PSO and GA.

Changes:

bug removed


Logo r-cran-darch 0.12.0

by r-cran-robot - July 19, 2016, 00:00:00 CET [ Project Homepage BibTeX Download ] 3 views, 1 download, 0 subscriptions

About: Package for Deep Architectures and Restricted Boltzmann Machines

Changes:

Initial Announcement on mloss.org by r-cran-robot


About: Nowadays this is very popular to use the deep architectures in machine learning. Deep Belief Networks (DBNs) are deep architectures that use a stack of Restricted Boltzmann Machines (RBM) to create a powerful generative model using training data. DBNs have many abilities such as feature extraction and classification that are used in many applications including image processing, speech processing, text categorization, etc. This paper introduces a new object oriented toolbox with the most important abilities needed for the implementation of DBNs. According to the results of the experiments conducted on the MNIST (image), ISOLET (speech), and the 20 Newsgroups (text) datasets, it was shown that the toolbox can learn automatically a good representation of the input from unlabeled data with better discrimination between different classes. Also on all the aforementioned datasets, the obtained classification errors are comparable to those of the state of the art classifiers. In addition, the toolbox supports different sampling methods (e.g. Gibbs, CD, PCD and our new FEPCD method), different sparsity methods (quadratic, rate distortion and our new normal method), different RBM types (generative and discriminative), GPU based, etc. The toolbox is a user-friendly open source software in MATLAB and Octave and is freely available on the website.

Changes:

New in toolbox

  • Using GPU in Backpropagation
  • Revision of some demo scripts
  • Function approximation with multiple outputs
  • Feature extraction with GRBM in first layer

cardinal


Showing Items 41-50 of 640 on page 5 of 64: Previous 1 2 3 4 5 6 7 8 9 10 Next Last