Projects running under agnostic.
Showing Items 21-40 of 232 on page 2 of 12: Previous 1 2 3 4 5 6 7 Next Last

Logo Tools for Regression and Classification 1.0.0

by matloff - October 29, 2016, 08:22:40 CET [ Project Homepage BibTeX Download ] 2128 views, 410 downloads, 3 subscriptions

About: Toolkit for parametric and nonparametric regression and classification.

Changes:

Initial Announcement on mloss.org.


Logo rectools a Novel Toolbox for Recommender Systems 1.0.0

by matloff - October 29, 2016, 07:41:58 CET [ Project Homepage BibTeX Download ] 1977 views, 425 downloads, 2 subscriptions

Rating Empty StarEmpty StarEmpty StarEmpty StarEmpty Star
(based on 1 vote)

About: Novel R toolbox for collaborative filtering recommender systems.

Changes:

Initial Announcement on mloss.org.


Logo JMLR GPML Gaussian Processes for Machine Learning Toolbox 4.0

by hn - October 19, 2016, 10:15:05 CET [ Project Homepage BibTeX Download ] 47762 views, 10492 downloads, 5 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 2 votes)

About: The GPML toolbox is a flexible and generic Octave/Matlab implementation of inference and prediction with Gaussian process models. The toolbox offers exact inference, approximate inference for non-Gaussian likelihoods (Laplace's Method, Expectation Propagation, Variational Bayes) as well for large datasets (FITC, VFE, KISS-GP). A wide range of covariance, likelihood, mean and hyperprior functions allows to create very complex GP models.

Changes:

A major code restructuring effort did take place in the current release unifying certain inference functions and allowing more flexibility in covariance function composition. We also redesigned the whole derivative computation pipeline to strongly improve the overall runtime. We finally include grid-based covariance approximations natively.

More generic sparse approximation using Power EP

  • unified treatment of FITC approximation, variational approaches VFE and hybrids

  • inducing input optimisation for all (compositions of) covariance functions dropping the previous limitation to a few standard examples

  • infFITC is now covered by the more generic infGaussLik function

Approximate covariance object unifying sparse approximations, grid-based approximations and exact covariance computations

  • implementation in cov/apx, cov/apxGrid, cov/apxSparse

  • generic infGaussLik unifies infExact, infFITC and infGrid

  • generic infLaplace unifies infLaplace, infFITC_Laplace and infGrid_Laplace

Hiearchical structure of covariance functions

  • clear hierachical compositional implementation

  • no more code duplication as present in covSEiso and covSEard pairs

  • two mother covariance functions

    • covDot for dot-product-based covariances and

    • covMaha for Mahalanobis-distance-based covariances

  • a variety of modifiers: eye, iso, ard, proj, fact, vlen

  • more flexibility as more variants are available and possible

  • all covariance functions offer derivatives w.r.t. inputs

Faster derivative computations for mean and cov functions

  • switched from partial derivatives to directional derivatives

  • simpler and more concise interface of mean and cov functions

  • much faster marginal likelihood derivative computations

  • simpler and more compact code

New mean functions

  • new mean/meanWSPC (Weighted Sum of Projected Cosines or Random Kitchen Sink features) following a suggestion by William Herlands

  • new mean/meanWarp for constructing a new mean from an existing one by means of a warping function adapted from William Herlands

New optimizer

  • added a new minimize_minfunc, contributed by Truong X. Nghiem

New GLM link function

  • added the twice logistic link function util/glm_invlink_logistic2

Smaller fixes

  • two-fold speedup of util/elsympol used by covADD by Truong X. Nghiem

  • bugfix in util/logphi as reported by John Darby


Logo r-cran-bst 0.3-14

by r-cran-robot - September 12, 2016, 00:00:00 CET [ Project Homepage BibTeX Download ] 9169 views, 2107 downloads, 2 subscriptions

About: Gradient Boosting

Changes:

Fetched by r-cran-robot on 2017-10-01 00:00:05.765455


Logo Sparse Compositional Metric Learning v1.11

by bellet - August 2, 2016, 11:43:03 CET [ BibTeX BibTeX for corresponding Paper Download ] 7090 views, 2115 downloads, 3 subscriptions

About: Scalable learning of global, multi-task and local metrics from data

Changes:

Minor bug fix in multi-task objective computation (thanks to Junjie Hu).


Logo r-cran-darch 0.12.0

by r-cran-robot - July 19, 2016, 00:00:00 CET [ Project Homepage BibTeX Download ] 388 views, 106 downloads, 0 subscriptions

About: Package for Deep Architectures and Restricted Boltzmann Machines

Changes:

Fetched by r-cran-robot on 2017-10-01 00:00:08.434606


Logo JMLR GPstuff 4.7

by avehtari - June 9, 2016, 17:45:15 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 47770 views, 11840 downloads, 3 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarWhole Star
(based on 1 vote)

About: The GPstuff toolbox is a versatile collection of Gaussian process models and computational tools required for inference. The tools include, among others, various inference methods, sparse approximations and model assessment methods.

Changes:

2016-06-09 Version 4.7

Development and release branches available at https://github.com/gpstuff-dev/gpstuff

New features

  • Simple Bayesian Optimization demo

Improvements

  • Improved use of PSIS
  • More options added to gp_monotonic
  • Monotonicity now works for additive covariance functions with selected variables
  • Possibility to use gpcf_squared.m-covariance function with derivative observations/monotonicity
  • Default behaviour made more robust by changing default jitter from 1e-9 to 1e-6
  • LA-LOO uses the cavity method as the default (see Vehtari et al (2016). Bayesian leave-one-out cross-validation approximations for Gaussian latent variable models. JMLR, accpeted for publication)
  • Selected variables -option works now better with monotonicity

Bugfixes

  • small error in derivative observation computation fixed
  • several minor bug fixes

Logo AutoWEKA 2.0

by larsko - May 19, 2016, 19:58:41 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3029 views, 941 downloads, 3 subscriptions

About: Automatically finds the best model with its best parameter settings for a given classification or regression task.

Changes:

Initial Announcement on mloss.org.


Logo JMLR JKernelMachines 3.0

by dpicard - May 4, 2016, 17:53:28 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 42556 views, 8990 downloads, 4 subscriptions

Rating Whole StarWhole StarWhole StarWhole Star1/2 Star
(based on 4 votes)

About: machine learning library in java for easy development of new kernels and kernel algorithms

Changes:

Version 3.0

/! Warning: this version is incompatible with previous code

  • change license to BSD 3-clauses
  • change package name to net.jkernelmachines
  • change to maven build system (available through central)
  • online training interfaces to allow continuous online learning
  • add a new budget oriented kernel classifier
  • new kernel and processing especially for strings

Logo r-cran-bnclassify 0.3.2

by r-cran-robot - May 1, 2016, 00:00:04 CET [ Project Homepage BibTeX Download ] 3166 views, 779 downloads, 2 subscriptions

About: Learning Discrete Bayesian Network Classifiers from Data

Changes:

Fetched by r-cran-robot on 2016-05-01 00:00:04.546512


Logo pymanopt 0.1

by j_towns - April 7, 2016, 14:44:27 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2402 views, 464 downloads, 3 subscriptions

About: Python toolbox for manifold optimization with support for automatic differentiation

Changes:

Initial Announcement on mloss.org.


Logo JaTeCS 1.0.0

by aesuli - April 5, 2016, 17:23:12 CET [ Project Homepage BibTeX Download ] 3225 views, 707 downloads, 2 subscriptions

About: Jatecs is an open source Java library focused on automatic text categorization.

Changes:

Initial Announcement on mloss.org.


Logo Local high order regularization 1.0

by kkim - March 2, 2016, 13:46:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2097 views, 533 downloads, 2 subscriptions

Rating Empty StarEmpty StarEmpty StarEmpty StarEmpty Star
(based on 1 vote)

About: Local high-order regularization for semi-supervised learning

Changes:

Initial Announcement on mloss.org.


Logo r-cran-BayesTree 0.3-1.4

by r-cran-robot - February 21, 2016, 00:00:00 CET [ Project Homepage BibTeX Download ] 12221 views, 2722 downloads, 1 subscription

About: Bayesian Additive Regression Trees

Changes:

Fetched by r-cran-robot on 2017-10-01 00:00:05.277645


About: Collection of algorithms for Gaussian Processes. Regression, Classification, Multi task, Multi output, Hierarchical, Sparse

Changes:

Initial Announcement on mloss.org.


Logo Sparse PCA 2.0

by tbuehler - December 31, 2015, 16:24:42 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8667 views, 1825 downloads, 3 subscriptions

About: A Matlab implementation of Sparse PCA using the inverse power method for nonlinear eigenproblems.

Changes:
  • Added deflation scheme to compute multiple principal components
  • Several internal runtime and memory optimizations
  • API change: sparsePCA.m is now used to compute multiple components; use computeTradeOffCurve.m to reproduce the examples in the NIPS paper

About: Efficient and Flexible Distributed/Mobile Deep Learning Framework, for python, R, Julia and more

Changes:

This version comes with Distributed and Mobile Examples


Logo Apache Mahout 0.11.1

by gsingers - November 9, 2015, 16:12:06 CET [ Project Homepage BibTeX Download ] 26761 views, 6625 downloads, 3 subscriptions

About: Apache Mahout is an Apache Software Foundation project with the goal of creating both a community of users and a scalable, Java-based framework consisting of many machine learning algorithm [...]

Changes:

Apache Mahout introduces a new math environment we call Samsara, for its theme of universal renewal. It reflects a fundamental rethinking of how scalable machine learning algorithms are built and customized. Mahout-Samsara is here to help people create their own math while providing some off-the-shelf algorithm implementations. At its core are general linear algebra and statistical operations along with the data structures to support them. You can use is as a library or customize it in Scala with Mahout-specific extensions that look something like R. Mahout-Samsara comes with an interactive shell that runs distributed operations on a Spark cluster. This make prototyping or task submission much easier and allows users to customize algorithms with a whole new degree of freedom. Mahout Algorithms include many new implementations built for speed on Mahout-Samsara. They run on Spark 1.3+ and some on H2O, which means as much as a 10x speed increase. You’ll find robust matrix decomposition algorithms as well as a Naive Bayes classifier and collaborative filtering. The new spark-itemsimilarity enables the next generation of cooccurrence recommenders that can use entire user click streams and context in making recommendations.


Logo Cognitive Foundry 3.4.2

by Baz - October 30, 2015, 06:53:03 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 35558 views, 5994 downloads, 4 subscriptions

About: The Cognitive Foundry is a modular Java software library of machine learning components and algorithms designed for research and applications.

Changes:
  • General:
    • Upgraded MTJ to 1.0.3.
  • Common:
    • Added package for hash function computation including Eva, FNV-1a, MD5, Murmur2, Prime, SHA1, SHA2
    • Added callback-based forEach implementations to Vector and InfiniteVector, which can be faster for iterating through some vector types.
    • Optimized DenseVector by removing a layer of indirection.
    • Added method to compute set of percentiles in UnivariateStatisticsUtil and fixed issue with percentile interpolation.
    • Added utility class for enumerating combinations.
    • Adjusted ScalarMap implementation hierarchy.
    • Added method for copying a map to VectorFactory and moved createVectorCapacity up from SparseVectorFactory.
    • Added method for creating square identity matrix to MatrixFactory.
    • Added Random implementation that uses a cached set of values.
  • Learning:
    • Implemented feature hashing.
    • Added factory for random forests.
    • Implemented uniform distribution over integer values.
    • Added Chi-squared similarity.
    • Added KL divergence.
    • Added general conditional probability distribution.
    • Added interfaces for Regression, UnivariateRegression, and MultivariateRegression.
    • Fixed null pointer exception that can happen in K-means with an empty cluster.
    • Fixed name of maxClusters property on AgglomerativeClusterer (was called maxMinDistance).
  • Text:
    • Improvements to LDA Gibbs sampler.

Logo Optunity 1.1.1

by claesenm - September 30, 2015, 07:06:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 10358 views, 2240 downloads, 3 subscriptions

About: Optunity is a library containing various optimizers for hyperparameter tuning. Hyperparameter tuning is a recurrent problem in many machine learning tasks, both supervised and unsupervised.This package provides several distinct approaches to solve such problems including some helpful facilities such as cross-validation and a plethora of score functions.

Changes:

This minor release has the same feature set as Optunity 1.1.0, but incorporates several bug fixes, mostly related to the specification of structured search spaces.


Showing Items 21-40 of 232 on page 2 of 12: Previous 1 2 3 4 5 6 7 Next Last