All entries.
Showing Items 121-130 of 609 on page 13 of 61: First Previous 8 9 10 11 12 13 14 15 16 17 18 Next Last

Logo ELKI 0.7.0

by erich - November 27, 2015, 18:23:16 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 17382 views, 3185 downloads, 4 subscriptions

About: ELKI is a framework for implementing data-mining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods.

Changes:

Additions and Improvements from ELKI 0.6.0:

ELKI is now available on Maven: https://search.maven.org/#artifactdetails|de.lmu.ifi.dbs.elki|elki|0.7.0|jar

Please clone https://github.com/elki-project/example-elki-project for a minimal project example.

Uncertain data types, and clustering algorithms for uncertain data.

Major refactoring of distances - removal of Distance values and removed support for non-double-valued distance functions (in particular DoubleDistance was removed). While this reduces the generality of ELKI, we could remove about 2.5% of the codebase by not having to have optimized codepaths for double-distance anymore. Generics for distances were present in almost any distance-based algorithm, and we were also happy to reduce the use of generics this way. Support for non-double-valued distances can trivially be added again, e.g. by adding the specialization one level higher: at the query instead of the distance level, for example. In this process, we also removed the Generics from NumberVector. The object-based get was deprecated for a good reason long ago, and e.g. doubleValue are more efficient (even for non-DoubleVectors).

Dropped some long-deprecated classes.

K-means:

  • speedups for some initialization heuristics.

  • K-means++ initialization no longer squares distances (again).

  • farthest-point heuristics now uses minimum instead of sum (renamed).

  • additional evaluation criteria.

  • Elkan's and Hamerly's faster k-means variants.

CLARA clustering.

X-means.

Hierarchical clustering:

  • Renamed naive algorithm to AGNES.

  • Anderbergs algorithm (faster than AGNES, slower than SLINK).

  • CLINK for complete linkage clustering in O(n²) time, O(n) memory.

  • Simple extraction from HDBSCAN.

  • "Optimal" extraction from HDBSCAN.

  • HDBSCAN, in two variants.

LSDBC clustering.

EM clustering was refactored and moved into its own package. The new version is much more extensible.

OPTICS clustering:

  • Added a list-based variant of OPTICS to our heap-based.

  • FastOPTICS (contributed by Johannes Schneider).

  • Improved OPTICS Xi cluster extraction.

Outlier detection:

  • KDEOS outlier detection (SDM14).

  • k-means based outlier detection (distance to centroid) and Silhouette coefficient based approach (which does not work too well on the toy data sets - the lowest silhouette are usually where two clusters touch).

  • bug fix in kNN weight, when distances are tied and kNN yields more than k results.

  • kNN and kNN weight outlier have their k parameter changed: old 2NN outlier is now 1NN outlier, as commonly understood in classification literature (1 nearest neighbor other than the query object; whereas in database literature the 1NN is usually the query object itself). You can get the old result back by decreasing k by one easily.

  • LOCI implementation is now only O(n^3 log n) instead of O(n^4).

  • Local Isolation Coefficient (LIC).

  • IDOS outlier detection with intrinsic dimensionality.

  • Baseline intrinsic dimensionality outlier detection.

  • Variance-of-Volumes outlier detection (VOV).

Parallel computation framework, and some parallelized algorithms

  • Parallel k-means.

  • Parallel LOF and variants.

LibSVM format parser.

kNN classification (with index acceleration).

Internal cluster evaluation:

  • Silhouette index.

  • Simplified Silhouette index (faster).

  • Davis-Bouldin index.

  • PBM index.

  • Variance-Ratio-Criteria.

  • Sum of squared errors.

  • C-Index.

  • Concordant pair indexes (Gamma, Tau).

  • Different noise handling strategies for internal indexes.

Statistical dependence measures:

  • Distance correlation dCor.

  • Hoeffings D.

  • Some divergence / mutual information measures.

Distance functions:

  • Big refactoring.

  • Time series distances refactored, allow variable length series now.

  • Hellinger distance and kernel function.

Preprocessing:

  • Faster MDS implementation using power iterations.

Indexing improvements:

  • Precomputed distance matrix "index".

  • iDistance index (static only).

  • Inverted-list index for sparse data and cosine/arccosine distance.

  • Cover tree index (static only).

  • Additional LSH hash functions.

Frequent Itemset Mining:

  • Improved APRIORI implementation.

  • FP-Growth added.

  • Eclat (basic version only) added.

Uncertain clustering:

  • Discrete and continuous data models.

  • FDBSCAN clustering.

  • UKMeans clustering.

  • CKMeans clustering.

  • Representative Uncertain Clustering (Meta-algorithm).

  • Center-of-mass meta Clustering (allows using other clustering algorithms on uncertain objects).

Mathematics:

  • Several estimators for intrinsic dimensionality.

MiniGUI has two "secret" new options: -minigui.last -minigui.autorun to load the last saved configuration and run it, for convenience.

Logging API has been extended, to make logging more convenient in a number of places (saving some lines for progress logging and timing).


Logo Encog Machine Learning Framework 3.2

by jeffheaton - July 5, 2014, 23:47:06 CET [ Project Homepage BibTeX Download ] 5518 views, 2097 downloads, 1 subscription

About: Encog is a Machine Learning framework for Java, C#, Javascript and C/C++ that supports SVM's, Genetic Programming, Bayesian Networks, Hidden Markov Models and other algorithms.

Changes:

Changes for Encog 3.2:

Issue #53: Fix Out Of Range Bug In BasicMLSequenceSet. Issue #52: Unhandled exception in Encog.Util.File.ResourceLoader.CreateStream (ResourceLoader.cs) Issue #50: Concurrency bugs in PruneIncremental Issue #48: Unit Tests Failing - TestHessian Issue #46: Couple of small fixes - Temporal DataSet and SCG training Issue #45: Fixed EndMinutesStrategy to correctly evaluate ShouldStop after the specified number of minutes have elapsed. Issue #44: Encog.ML.Data.Basic.BasicMLDataPairCentroid.Add() & .Remove() Issue #43: Unit Tests Failing - Matrix not full rank Issue #42: Nuget - NuSpec Issue #36: Load Examples easier


Logo JMLR EnsembleSVM 2.0

by claesenm - March 31, 2014, 08:06:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 10393 views, 3620 downloads, 2 subscriptions

About: The EnsembleSVM library offers functionality to perform ensemble learning using Support Vector Machine (SVM) base models. In particular, we offer routines for binary ensemble models using SVM base classifiers. Experimental results have shown the predictive performance to be comparable with standard SVM models but with drastically reduced training time. Ensemble learning with SVM models is particularly useful for semi-supervised tasks.

Changes:

The library has been updated and features a variety of new functionality as well as more efficient implementations of original features. The following key improvements have been made:

  1. Support for multithreading in training and prediction with ensemble models. Since both of these are embarassingly parallel, this has induced a significant speedup (3-fold on quad-core).
  2. Extensive programming framework for aggregation of base model predictions which allows highly efficient prototyping of new aggregation approaches. Additionally we provide several predefined strategies, including (weighted) majority voting, logistic regression and nonlinear SVMs of your choice -- be sure to check out the esvm-edit tool! The provided framework also allows you to efficiently program your own, novel aggregation schemes.
  3. Full code transition to C++11, the latest C++ standard, which enabled various performance improvements. The new release requires moderately recent compilers, such as gcc 4.7.2+ or clang 3.2+.
  4. Generic implementations of convenient facilities have been added, such as thread pools, deserialization factories and more.

The API and ABI have undergone significant changes, many of which are due to the transition to C++11.


Logo epac 0.10

by jinpengli - October 9, 2013, 14:00:15 CET [ Project Homepage BibTeX Download ] 2496 views, 659 downloads, 1 subscription

About: Embarrassingly Parallel Array Computing: EPAC is a machine learning workflow builder.

Changes:

Initial Announcement on mloss.org.


Logo Epistatic MAP Imputation 1.1

by colm - November 25, 2010, 21:01:10 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3845 views, 913 downloads, 1 subscription

About: Epistatic miniarray profiles (E-MAPs) are a high-throughput approach capable of quantifying aggravating or alleviating genetic interactions between gene pairs. The datasets resulting from E-MAP experiments typically take the form of a symmetric pairwise matrix of interaction scores. These datasets have a significant number of missing values - up to 35% - that can reduce the effectiveness of some data analysis techniques and prevent the use of others. This project contains nearest neighbor based tools for the imputation and prediction of these missing values. The code is implemented in Python and uses a nearest neighbor based approach. Two variants are used - a simple weighted nearest neighbors, and a local least squares based regression.

Changes:

Initial Announcement on mloss.org.


Logo JMLR Error Correcting Output Codes Library 0.1

by sescalera - March 5, 2010, 16:49:12 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 10871 views, 1450 downloads, 1 subscription

About: The open source Error-Correcting Output Codes (ECOC) library contains both state-of-the-art coding and decoding designs, as well as the option to include your own coding, decoding, and base classifier.

Changes:

Initial Announcement on mloss.org.


Logo Evaluation toolkit 1.0

by openpr_nlpr - August 13, 2013, 08:58:25 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2364 views, 528 downloads, 1 subscription

About: This evaluation toolkit provides a unified framework for evaluating bag-of-words based encoding methods over several standard image classification datasets.

Changes:

Initial Announcement on mloss.org.


About: This page contains the implementation used in the paper „Experimental Design for Efficient Identification of Gene Regulatory Networks using Sparse Bayesian Models“ by Florian Steinke, Matthias [...]

Changes:

Initial Announcement on mloss.org.


Logo Experiment Databases for Machine Learning 0.1

by JoaquinVanschoren - October 7, 2008, 18:06:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7377 views, 1217 downloads, 1 subscription

About: Experiment Databases for Machine Learning is a large public database of machine learning experiments as well as a framework for producing similar databases for specific goals. It provides a way to [...]

Changes:

Initial Announcement on mloss.org.


Logo ExtRESCAL 0.7.2

by nzhiltsov - January 20, 2015, 00:35:15 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 6601 views, 1274 downloads, 2 subscriptions

About: Scalable tensor factorization

Changes:
  • Improve (speed up) initialization of A by summation

Showing Items 121-130 of 609 on page 13 of 61: First Previous 8 9 10 11 12 13 14 15 16 17 18 Next Last