Projects running under linux.
Showing Items 41-60 of 260 on page 3 of 13: Previous 1 2 3 4 5 6 7 8 Next Last

Logo Hivemall 0.3

by myui - March 13, 2015, 17:08:22 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8193 views, 1367 downloads, 3 subscriptions

About: Hivemall is a scalable machine learning library running on Hive/Hadoop.

Changes:
  • Supported Matrix Factorization
  • Added a support for TF-IDF computation
  • Supported AdaGrad/AdaDelta
  • Supported AdaGradRDA classification
  • Added normalization scheme

Logo libcmaes 0.9.5

by beniz - March 9, 2015, 09:05:22 CET [ Project Homepage BibTeX Download ] 7832 views, 1581 downloads, 3 subscriptions

About: Libcmaes is a multithreaded C++11 library (with Python bindings) for high performance blackbox stochastic optimization of difficult, possibly non-linear and non-convex functions, using the CMA-ES algorithm for Covariance Matrix Adaptation Evolution Strategy. Libcmaes is useful to minimize / maximize any function, without information regarding gradient or derivability.

Changes:

This is a major release, with several novelties, improvements and fixes, among which:

  • step-size two-point adaptaion scheme for improved performances in some settings, ref #88

  • important bug fixes to the ACM surrogate scheme, ref #57, #106

  • simple high-level workflow under Python, ref #116

  • improved performances in high dimensions, ref #97

  • improved profile likelihood and contour computations, including under geno/pheno transforms, ref #30, #31, #48

  • elitist mechanism for forcing best solutions during evolution, ref 103

  • new legacy plotting function, ref #110

  • optional initial function value, ref #100

  • improved C++ API, ref #89

  • Python bindings support with Anaconda, ref #111

  • configure script now tries to detect numpy when building Python bindings, ref #113

  • Python bindings now have embedded documentation, ref #114

  • support for Travis continuous integration, ref #122

  • lower resolution random seed initialization


Logo CN24 Convolutional Neural Networks for Semantic Segmentation 1.0

by erik - February 23, 2015, 09:02:06 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1958 views, 396 downloads, 1 subscription

About: CN24 is a complete semantic segmentation framework using fully convolutional networks.

Changes:

Initial Announcement on mloss.org.


Logo JMLR DLLearner 1.0

by Jens - February 13, 2015, 11:39:46 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 19354 views, 4618 downloads, 6 subscriptions

Rating Whole StarWhole StarWhole StarWhole Star1/2 Star
(based on 3 votes)

About: The DL-Learner framework contains several algorithms for supervised concept learning in Description Logics (DLs) and OWL.

Changes:

See http://dl-learner.org/development/changelog/.


Logo Auto encoder Based Data Clustering Toolkit 1.0

by openpr_nlpr - February 10, 2015, 08:30:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1858 views, 354 downloads, 2 subscriptions

About: The auto-encoder based data clustering toolkit provides a quick start of clustering based on deep auto-encoder nets. This toolkit can cluster data in feature space with a deep nonlinear nets.

Changes:

Initial Announcement on mloss.org.


Logo Histogram of Oriented Gradient 1.0

by openpr_nlpr - February 10, 2015, 08:27:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1497 views, 296 downloads, 2 subscriptions

About: This is an exact implementation of Histogram of Oriented Gradient as mentioned in the paper by Dalal.

Changes:

Initial Announcement on mloss.org.


Logo JMLR Information Theoretical Estimators 0.61

by szzoli - February 8, 2015, 14:04:27 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 89600 views, 17522 downloads, 2 subscriptions

About: ITE (Information Theoretical Estimators) is capable of estimating many different variants of entropy, mutual information, divergence, association measures, cross quantities and kernels on distributions. Thanks to its highly modular design, ITE supports additionally (i) the combinations of the estimation techniques, (ii) the easy construction and embedding of novel information theoretical estimators, and (iii) their immediate application in information theoretical optimization problems.

Changes:
  • Explicit additive constant computation in generalized kNN based Renyi entropy estimators: enhancement suggestion has been added.

  • Analytical value computation of the exponentiated Jensen-Renyi kernel-2: simplified.


Logo JMLR SHOGUN 4.0.0

by sonne - February 5, 2015, 09:09:37 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 104846 views, 14869 downloads, 6 subscriptions

Rating Whole StarWhole StarWhole StarEmpty StarEmpty Star
(based on 6 votes)

About: The SHOGUN machine learning toolbox's focus is on large scale learning methods with focus on Support Vector Machines (SVM), providing interfaces to python, octave, matlab, r and the command line.

Changes:

This release features the work of our 8 GSoC 2014 students [student; mentors]:

  • OpenCV Integration and Computer Vision Applications [Abhijeet Kislay; Kevin Hughes]
  • Large-Scale Multi-Label Classification [Abinash Panda; Thoralf Klein]
  • Large-scale structured prediction with approximate inference [Jiaolong Xu; Shell Hu]
  • Essential Deep Learning Modules [Khaled Nasr; Sergey Lisitsyn, Theofanis Karaletsos]
  • Fundamental Machine Learning: decision trees, kernel density estimation [Parijat Mazumdar ; Fernando Iglesias]
  • Shogun Missionary & Shogun in Education [Saurabh Mahindre; Heiko Strathmann]
  • Testing and Measuring Variable Interactions With Kernels [Soumyajit De; Dino Sejdinovic, Heiko Strathmann]
  • Variational Learning for Gaussian Processes [Wu Lin; Heiko Strathmann, Emtiyaz Khan]

It also contains several cleanups and bugfixes:

Features

  • New Shogun project description [Heiko Strathmann]
  • ID3 algorithm for decision tree learning [Parijat Mazumdar]
  • New modes for PCA matrix factorizations: SVD & EVD, in-place or reallocating [Parijat Mazumdar]
  • Add Neural Networks with linear, logistic and softmax neurons [Khaled Nasr]
  • Add kernel multiclass strategy examples in multiclass notebook [Saurabh Mahindre]
  • Add decision trees notebook containing examples for ID3 algorithm [Parijat Mazumdar]
  • Add sudoku recognizer ipython notebook [Alejandro Hernandez]
  • Add in-place subsets on features, labels, and custom kernels [Heiko Strathmann]
  • Add Principal Component Analysis notebook [Abhijeet Kislay]
  • Add Multiple Kernel Learning notebook [Saurabh Mahindre]
  • Add Multi-Label classes to enable Multi-Label classification [Thoralf Klein]
  • Add rectified linear neurons, dropout and max-norm regularization to neural networks [Khaled Nasr]
  • Add C4.5 algorithm for multiclass classification using decision trees [Parijat Mazumdar]
  • Add support for arbitrary acyclic graph-structured neural networks [Khaled Nasr]
  • Add CART algorithm for classification and regression using decision trees [Parijat Mazumdar]
  • Add CHAID algorithm for multiclass classification and regression using decision trees [Parijat Mazumdar]
  • Add Convolutional Neural Networks [Khaled Nasr]
  • Add Random Forests algorithm for ensemble learning using CART [Parijat Mazumdar]
  • Add Restricted Botlzmann Machines [Khaled Nasr]
  • Add Stochastic Gradient Boosting algorithm for ensemble learning [Parijat Mazumdar]
  • Add Deep contractive and denoising autoencoders [Khaled Nasr]
  • Add Deep belief networks [Khaled Nasr]

Bugfixes

  • Fix reference counting bugs in CList when reference counting is on [Heiko Strathmann, Thoralf Klein, lambday]
  • Fix memory problem in PCA::apply_to_feature_matrix [Parijat Mazumdar]
  • Fix crash in LeastAngleRegression for the case D greater than N [Parijat Mazumdar]
  • Fix memory violations in bundle method solvers [Thoralf Klein]
  • Fix fail in library_mldatahdf5.cpp example when http://mldata.org is not working properly [Parijat Mazumdar]
  • Fix memory leaks in Vowpal Wabbit, LibSVMFile and KernelPCA [Thoralf Klein]
  • Fix memory and control flow issues discovered by Coverity [Thoralf Klein]
  • Fix R modular interface SWIG typemap (Requires SWIG >= 2.0.5) [Matt Huska]

Cleanup and API Changes

  • PCA now depends on Eigen3 instead of LAPACK [Parijat Mazumdar]
  • Removing redundant and fixing implicit imports [Thoralf Klein]
  • Hide many methods from SWIG, reducing compile memory by 500MiB [Heiko Strathmann, Fernando Iglesias, Thoralf Klein]

Logo Distributed Frank Wolfe Algorithm 0.02

by alirezabagheri - January 28, 2015, 00:35:03 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1753 views, 481 downloads, 2 subscriptions

About: Distributed optimization: Support Vector Machines and LASSO regression on distributed data

Changes:

Initial Upload


Logo fertilized forests 1.0beta

by Chrisl_S - January 23, 2015, 16:04:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1669 views, 424 downloads, 1 subscription

About: The fertilized forests project has the aim to provide an easy to use, easy to extend, yet fast library for decision forests. It summarizes the research in this field and provides a solid platform to extend it. Offering consistent interfaces to C++, Python and Matlab and being available for all major compilers gives the user high flexibility for using the library.

Changes:

Initial Announcement on mloss.org.


Logo Hub Miner 1.1

by nenadtomasev - January 22, 2015, 16:33:51 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2998 views, 589 downloads, 2 subscriptions

About: Hubness-aware Machine Learning for High-dimensional Data

Changes:
  • BibTex support for all algorithm implementations, making all of them easy to reference (via algref package).

  • Two more hubness-aware approaches (meta-metric-learning and feature construction)

  • An implementation of Hit-Miss networks for analysis.

  • Several minor bug fixes.

  • The following instance selection methods were added: HMScore, Carving, Iterative Case Filtering, ENRBF.

  • The following clustering quality indexes were added: Folkes-Mallows, Calinski-Harabasz, PBM, G+, Tau, Point-Biserial, Hubert's statistic, McClain-Rao, C-root-k.

  • Some more experimental scripts have been included.

  • Extensions in the estimation of hubness risk.

  • Alias and weighted reservoir methods for weight-proportional random selection.


Logo JMLR RL library 3.00.00

by frezza - January 13, 2015, 04:15:16 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2478 views, 677 downloads, 2 subscriptions

About: A template based C++ reinforcement learning library

Changes:

Initial Announcement on mloss.org.


Logo gaml 1.10

by frezza - January 8, 2015, 14:06:58 CET [ Project Homepage BibTeX Download ] 1427 views, 397 downloads, 2 subscriptions

About: C++ generic programming tools for machine learning

Changes:

Initial Announcement on mloss.org.


Logo JEMLA 1.0

by bathaeian - January 4, 2015, 08:34:49 CET [ Project Homepage BibTeX Download ] 1310 views, 416 downloads, 3 subscriptions

About: Java package for calculating Entropy for Machine Learning Applications. It has implemented several methods of handling missing values. So it can be used as a lab for examining missing values.

Changes:

Discretizing numerical values is added to calculate mode of values and fractional replacement of missing ones. class diagram is on the web http://profs.basu.ac.ir/bathaeian/free_space/jemla.rar


Logo Accord.NET Framework 2.14.0

by cesarsouza - December 9, 2014, 23:04:04 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 26397 views, 4965 downloads, 2 subscriptions

About: The Accord.NET Framework is a .NET machine learning framework combined with audio and image processing libraries completely written in C#. It is a complete framework for building production-grade computer vision, computer audition, signal processing and statistics applications even for commercial use. A comprehensive set of sample applications provide a fast start to get up and running quickly, and an extensive online documentation helps fill in the details.

Changes:

Adding a large number of new distributions, such as Anderson-Daring, Shapiro-Wilk, Inverse Chi-Square, Lévy, Folded Normal, Shifted Log-Logistic, Kumaraswamy, Trapezoidal, U-quadratic and BetaPrime distributions, Birnbaum-Saunders, Generalized Normal, Gumbel, Power Lognormal, Power Normal, Triangular, Tukey Lambda, Logistic, Hyperbolic Secant, Degenerate and General Continuous distributions.

Other additions include new statistical hypothesis tests such as Anderson-Daring and Shapiro-Wilk; as well as support for all of LIBLINEAR's support vector machine algorithms; and format reading support for MATLAB/Octave matrices, LibSVM models, sparse LibSVM data files, and many others.

For a complete list of changes, please see the full release notes at the release details page at:

https://github.com/accord-net/framework/releases


About: a parallel LDA learning toolbox in Multi-Core Systems for big topic modeling.

Changes:

Initial Announcement on mloss.org.


Logo linearizedGP 1.0

by dsteinberg - November 28, 2014, 07:02:54 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1628 views, 375 downloads, 1 subscription

About: Gaussian processes with general nonlinear likelihoods using the unscented transform or Taylor series linearisation.

Changes:

Initial Announcement on mloss.org.


Logo LogRegCrowds, Logistic Regression from Crowds 1.0

by fmpr - October 30, 2014, 19:10:23 CET [ Project Homepage BibTeX Download ] 1486 views, 435 downloads, 2 subscriptions

About: LogReg-Crowds is a collection of Julia implementations of various approaches for learning a logistic regression model multiple annotators and crowds, namely the works of Raykar et al. (2010), Rodrigues et al. (2013) and Dawid and Skene (1979).

Changes:

Initial Announcement on mloss.org. Added GitHub page.


About: This library implements the Optimum-Path Forest classifier for unsupervised and supervised learning.

Changes:

Initial Announcement on mloss.org.


Logo pySPACE 1.2

by krell84 - October 29, 2014, 15:36:28 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 4575 views, 940 downloads, 1 subscription

About: pySPACE is the abbreviation for "Signal Processing and Classification Environment in Python using YAML and supporting parallelization". It is a modular software for processing of large data streams that has been specifically designed to enable distributed execution and empirical evaluation of signal processing chains. Various signal processing algorithms (so called nodes) are available within the software, from finite impulse response filters over data-dependent spatial filters (e.g. CSP, xDAWN) to established classifiers (e.g. SVM, LDA). pySPACE incorporates the concept of node and node chains of the MDP framework. Due to its modular architecture, the software can easily be extended with new processing nodes and more general operations. Large scale empirical investigations can be configured using simple text- configuration files in the YAML format, executed on different (distributed) computing modalities, and evaluated using an interactive graphical user interface.

Changes:

improved testing, improved documentation, windows compatibility, more algorithms


Showing Items 41-60 of 260 on page 3 of 13: Previous 1 2 3 4 5 6 7 8 Next Last