All entries.
Showing Items 101-110 of 537 on page 11 of 54: First Previous 6 7 8 9 10 11 12 13 14 15 16 Next Last

Logo ELF Ensemble Learning Framework 0.1

by mjahrer - May 10, 2010, 23:54:53 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 4714 views, 778 downloads, 1 subscription

About: ELF provides many well implemented supervised learners for classification and regression tasks with an opportunity of ensemble learning.

Changes:

Initial Announcement on mloss.org.


Logo ELKI 0.6.0

by erich - January 10, 2014, 18:32:28 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 9577 views, 1756 downloads, 3 subscriptions

About: ELKI is a framework for implementing data-mining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods.

Changes:

Additions and Improvements from ELKI 0.5.5:

Algorithms

Clustering:

  • Hierarchical Clustering - the slower naive variants were added, and the code was refactored
  • Partition extraction from hierarchical clusterings - different linkage strategies (e.g. Ward)
  • Canopy pre-Clustering
  • Naive Mean-Shift Clustering
  • Affinity propagation clustering (both with distances and similarities / kernel functions)
  • K-means variations: Best-of-multiple-runs, bisecting k-means
  • New k-means initialization: farthest points, sample initialization
  • Cheng and Church Biclustering
  • P3C Subspace Clustering
  • One-dimensional clustering algorithm based on kernel density estimation

Outlier detection

  • COP - correlation outlier probabilities
  • LDF - a kernel density based LOF variant
  • Simplified LOF - a simpler version of LOF (not using reachability distance)
  • Simple Kernel Density LOF - a simple LOF using kernel density (more consistent than LDF)
  • Simple outlier ensemble algorithm
  • PINN - projection indexed nearest neighbors, via projected indexes.
  • ODIN - kNN graph based outlier detection
  • DWOF - Dynamic-Window Outlier Factor (contributed by Omar Yousry)
  • ABOD refactored, into ABOD, FastABOD and LBABOD

Distances

  • Geodetic distances now support different world models (WGS84 etc.) and are subtantially faster.
  • Levenshtein distances for processing strings, e.g. for analyzing phonemes (contributed code, see "Word segmentation through cross-lingual word-to-phoneme alignment", SLT2013, Stahlberg et al.)
  • Bray-Curtis, Clark, Kulczynski1 and Lorentzian distances with R-tree indexing support
  • Histogram matching distances
  • Probabilistic divergence distances (Jeffrey, Jensen-Shannon, Chi2, Kullback-Leibler)
  • Kulczynski2 similarity
  • Kernel similarity code has been refactored, and additional kernel functions have been added

Database Layer and Data Types

Projection layer * Parser for simple textual data (for use with Levenshtein distance) Various random projection families (including Feature Bagging, Achlioptas, and p-stable) Latitude+Longitude to ECEF Sparse vector improvements and bug fixes New filter: remove NaN values and missing values New filter: add histogram-based jitter New filter: normalize using statistical distributions New filter: robust standardization using Median and MAD New filter: Linear discriminant analysis (LDA)

Index Layer

  • Another speed up in R-trees
  • Refactoring of M- and R-trees: Support for different strategies in M-tree New strategies for M-tree splits Speedups in M-tree
  • New index structure: in-memory k-d-tree
  • New index structure: in-memory Locality Sensitive Hashing (LSH)
  • New index structure: approximate projected indexes, such as PINN
  • Index support for geodetic data - (Details: Geodetic Distance Queries on R-Trees for Indexing Geographic Data, SSTD13)
  • Sampled k nearest neighbors: reference KDD13 "Subsampling for Efficient and Effective Unsupervised Outlier Detection Ensembles"
  • Cached (precomputed) k-nearest neighbors to share across multiple runs
  • Benchmarking "algorithms" for indexes

Mathematics and Statistics

  • Many new distributions have been added, now 28 different distributions are supported
  • Additional estimation methods (using advanced statistics such as L-Moments), now 44 estimators are available
  • Trimming and Winsorizing
  • Automatic best-fit distribution estimation
  • Preprocessor using these distributions for rescaling data sets
  • API changes related to the new distributions support
  • More kernel density functions
  • RANSAC covariance matrix builder (unfortunately rather slow)

Visualization

  • 3D projected coordinates (Details: Interactive Data Mining with 3D-Parallel-Coordinate-Trees, SIGMOD2013)
  • Convex hulls now also include nested hierarchical clusters

Other

  • Parser speedups
  • Sparse vector bug fixes and improvements
  • Various bug fixes
  • PCA, MDS and LDA filters
  • Text output was slightly improved (but still needs to be redesigned from scratch - please contribute!)
  • Refactoring of hierarchy classes
  • New heap classes and infrastructure enhancements
  • Classes can have aliases, e.g. "l2" for euclidean distance.
  • Some error messages were made more informative.
  • Benchmarking classes, also for approximate nearest neighbor search.

Logo Encog Machine Learning Framework 3.2

by jeffheaton - July 5, 2014, 23:47:06 CET [ Project Homepage BibTeX Download ] 2603 views, 575 downloads, 1 subscription

About: Encog is a Machine Learning framework for Java, C#, Javascript and C/C++ that supports SVM's, Genetic Programming, Bayesian Networks, Hidden Markov Models and other algorithms.

Changes:

Changes for Encog 3.2:

Issue #53: Fix Out Of Range Bug In BasicMLSequenceSet. Issue #52: Unhandled exception in Encog.Util.File.ResourceLoader.CreateStream (ResourceLoader.cs) Issue #50: Concurrency bugs in PruneIncremental Issue #48: Unit Tests Failing - TestHessian Issue #46: Couple of small fixes - Temporal DataSet and SCG training Issue #45: Fixed EndMinutesStrategy to correctly evaluate ShouldStop after the specified number of minutes have elapsed. Issue #44: Encog.ML.Data.Basic.BasicMLDataPairCentroid.Add() & .Remove() Issue #43: Unit Tests Failing - Matrix not full rank Issue #42: Nuget - NuSpec Issue #36: Load Examples easier


Logo JMLR EnsembleSVM 2.0

by claesenm - March 31, 2014, 08:06:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 4557 views, 1603 downloads, 2 subscriptions

About: The EnsembleSVM library offers functionality to perform ensemble learning using Support Vector Machine (SVM) base models. In particular, we offer routines for binary ensemble models using SVM base classifiers. Experimental results have shown the predictive performance to be comparable with standard SVM models but with drastically reduced training time. Ensemble learning with SVM models is particularly useful for semi-supervised tasks.

Changes:

The library has been updated and features a variety of new functionality as well as more efficient implementations of original features. The following key improvements have been made:

  1. Support for multithreading in training and prediction with ensemble models. Since both of these are embarassingly parallel, this has induced a significant speedup (3-fold on quad-core).
  2. Extensive programming framework for aggregation of base model predictions which allows highly efficient prototyping of new aggregation approaches. Additionally we provide several predefined strategies, including (weighted) majority voting, logistic regression and nonlinear SVMs of your choice -- be sure to check out the esvm-edit tool! The provided framework also allows you to efficiently program your own, novel aggregation schemes.
  3. Full code transition to C++11, the latest C++ standard, which enabled various performance improvements. The new release requires moderately recent compilers, such as gcc 4.7.2+ or clang 3.2+.
  4. Generic implementations of convenient facilities have been added, such as thread pools, deserialization factories and more.

The API and ABI have undergone significant changes, many of which are due to the transition to C++11.


Logo epac 0.10

by jinpengli - October 9, 2013, 14:00:15 CET [ Project Homepage BibTeX Download ] 1201 views, 320 downloads, 1 subscription

About: Embarrassingly Parallel Array Computing: EPAC is a machine learning workflow builder.

Changes:

Initial Announcement on mloss.org.


Logo Epistatic MAP Imputation 1.1

by colm - November 25, 2010, 21:01:10 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2611 views, 625 downloads, 1 subscription

About: Epistatic miniarray profiles (E-MAPs) are a high-throughput approach capable of quantifying aggravating or alleviating genetic interactions between gene pairs. The datasets resulting from E-MAP experiments typically take the form of a symmetric pairwise matrix of interaction scores. These datasets have a significant number of missing values - up to 35% - that can reduce the effectiveness of some data analysis techniques and prevent the use of others. This project contains nearest neighbor based tools for the imputation and prediction of these missing values. The code is implemented in Python and uses a nearest neighbor based approach. Two variants are used - a simple weighted nearest neighbors, and a local least squares based regression.

Changes:

Initial Announcement on mloss.org.


Logo JMLR Error Correcting Output Codes Library 0.1

by sescalera - March 5, 2010, 16:49:12 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8370 views, 1071 downloads, 1 subscription

About: The open source Error-Correcting Output Codes (ECOC) library contains both state-of-the-art coding and decoding designs, as well as the option to include your own coding, decoding, and base classifier.

Changes:

Initial Announcement on mloss.org.


Logo Evaluation toolkit 1.0

by openpr_nlpr - August 13, 2013, 08:58:25 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 924 views, 182 downloads, 1 subscription

About: This evaluation toolkit provides a unified framework for evaluating bag-of-words based encoding methods over several standard image classification datasets.

Changes:

Initial Announcement on mloss.org.


About: This page contains the implementation used in the paper „Experimental Design for Efficient Identification of Gene Regulatory Networks using Sparse Bayesian Models“ by Florian Steinke, Matthias [...]

Changes:

Initial Announcement on mloss.org.


Logo Experiment Databases for Machine Learning 0.1

by JoaquinVanschoren - October 7, 2008, 18:06:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 6069 views, 965 downloads, 1 subscription

About: Experiment Databases for Machine Learning is a large public database of machine learning experiments as well as a framework for producing similar databases for specific goals. It provides a way to [...]

Changes:

Initial Announcement on mloss.org.


Showing Items 101-110 of 537 on page 11 of 54: First Previous 6 7 8 9 10 11 12 13 14 15 16 Next Last