Projects running under platform independent.
Showing Items 21-40 of 83 on page 2 of 5: Previous 1 2 3 4 5 Next

About: RLLib is a lightweight C++ template library that implements incremental, standard, and gradient temporal-difference learning algorithms in Reinforcement Learning. It is an optimized library for robotic applications and embedded devices that operates under fast duty cycles (e.g., < 30 ms). RLLib has been tested and evaluated on RoboCup 3D soccer simulation agents, physical NAO V4 humanoid robots, and Tiva C series launchpad microcontrollers to predict, control, learn behaviors, and represent learnable knowledge. The implementation of the RLLib library is inspired by the RLPark API, which is a library of temporal-difference learning algorithms written in Java.

Changes:

Current release version is v2.0.


Logo Chordalysis 1.0

by fpetitjean - March 24, 2014, 01:22:06 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1263 views, 333 downloads, 1 subscription

About: Log-linear analysis for high-dimensional data

Changes:

Initial Announcement on mloss.org.


Logo MOSIS 0.55

by claasahl - March 9, 2014, 17:35:40 CET [ BibTeX Download ] 3399 views, 1075 downloads, 2 subscriptions

About: MOSIS is a modularized framework for signal processing, stream analysis, machine learning and stream mining applications.

Changes:
  • Move "flow"-related classes into package "de.claas.mosis.flow" (e.g. Node and Link).
  • Refined and improved "flow"-related tests (e.g. Iterator and Node tests).
  • Refactored tests for data formats (e.g. PlainText and JSON tests).
  • Added visitor design pattern for graph-based functions (e.g. initialization and processing).
  • Documented parameters of Processor implementations.

Logo The Choquet Kernel 1.00

by AliFall - February 11, 2014, 16:21:15 CET [ BibTeX BibTeX for corresponding Paper Download ] 1058 views, 291 downloads, 1 subscription

About: The package computes the optimal parameters for the Choquet kernel

Changes:

Initial Announcement on mloss.org.


Logo Ordinal Choquistic Regression 1.00

by AliFall - January 30, 2014, 15:42:34 CET [ BibTeX BibTeX for corresponding Paper Download ] 1231 views, 297 downloads, 1 subscription

About: "Ordinal Choquistic Regression" model using the maximum likelihood

Changes:

Initial Announcement on mloss.org.


Logo ELKI 0.6.0

by erich - January 10, 2014, 18:32:28 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 11771 views, 2114 downloads, 3 subscriptions

About: ELKI is a framework for implementing data-mining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods.

Changes:

Additions and Improvements from ELKI 0.5.5:

Algorithms

Clustering:

  • Hierarchical Clustering - the slower naive variants were added, and the code was refactored
  • Partition extraction from hierarchical clusterings - different linkage strategies (e.g. Ward)
  • Canopy pre-Clustering
  • Naive Mean-Shift Clustering
  • Affinity propagation clustering (both with distances and similarities / kernel functions)
  • K-means variations: Best-of-multiple-runs, bisecting k-means
  • New k-means initialization: farthest points, sample initialization
  • Cheng and Church Biclustering
  • P3C Subspace Clustering
  • One-dimensional clustering algorithm based on kernel density estimation

Outlier detection

  • COP - correlation outlier probabilities
  • LDF - a kernel density based LOF variant
  • Simplified LOF - a simpler version of LOF (not using reachability distance)
  • Simple Kernel Density LOF - a simple LOF using kernel density (more consistent than LDF)
  • Simple outlier ensemble algorithm
  • PINN - projection indexed nearest neighbors, via projected indexes.
  • ODIN - kNN graph based outlier detection
  • DWOF - Dynamic-Window Outlier Factor (contributed by Omar Yousry)
  • ABOD refactored, into ABOD, FastABOD and LBABOD

Distances

  • Geodetic distances now support different world models (WGS84 etc.) and are subtantially faster.
  • Levenshtein distances for processing strings, e.g. for analyzing phonemes (contributed code, see "Word segmentation through cross-lingual word-to-phoneme alignment", SLT2013, Stahlberg et al.)
  • Bray-Curtis, Clark, Kulczynski1 and Lorentzian distances with R-tree indexing support
  • Histogram matching distances
  • Probabilistic divergence distances (Jeffrey, Jensen-Shannon, Chi2, Kullback-Leibler)
  • Kulczynski2 similarity
  • Kernel similarity code has been refactored, and additional kernel functions have been added

Database Layer and Data Types

Projection layer * Parser for simple textual data (for use with Levenshtein distance) Various random projection families (including Feature Bagging, Achlioptas, and p-stable) Latitude+Longitude to ECEF Sparse vector improvements and bug fixes New filter: remove NaN values and missing values New filter: add histogram-based jitter New filter: normalize using statistical distributions New filter: robust standardization using Median and MAD New filter: Linear discriminant analysis (LDA)

Index Layer

  • Another speed up in R-trees
  • Refactoring of M- and R-trees: Support for different strategies in M-tree New strategies for M-tree splits Speedups in M-tree
  • New index structure: in-memory k-d-tree
  • New index structure: in-memory Locality Sensitive Hashing (LSH)
  • New index structure: approximate projected indexes, such as PINN
  • Index support for geodetic data - (Details: Geodetic Distance Queries on R-Trees for Indexing Geographic Data, SSTD13)
  • Sampled k nearest neighbors: reference KDD13 "Subsampling for Efficient and Effective Unsupervised Outlier Detection Ensembles"
  • Cached (precomputed) k-nearest neighbors to share across multiple runs
  • Benchmarking "algorithms" for indexes

Mathematics and Statistics

  • Many new distributions have been added, now 28 different distributions are supported
  • Additional estimation methods (using advanced statistics such as L-Moments), now 44 estimators are available
  • Trimming and Winsorizing
  • Automatic best-fit distribution estimation
  • Preprocessor using these distributions for rescaling data sets
  • API changes related to the new distributions support
  • More kernel density functions
  • RANSAC covariance matrix builder (unfortunately rather slow)

Visualization

  • 3D projected coordinates (Details: Interactive Data Mining with 3D-Parallel-Coordinate-Trees, SIGMOD2013)
  • Convex hulls now also include nested hierarchical clusters

Other

  • Parser speedups
  • Sparse vector bug fixes and improvements
  • Various bug fixes
  • PCA, MDS and LDA filters
  • Text output was slightly improved (but still needs to be redesigned from scratch - please contribute!)
  • Refactoring of hierarchy classes
  • New heap classes and infrastructure enhancements
  • Classes can have aliases, e.g. "l2" for euclidean distance.
  • Some error messages were made more informative.
  • Benchmarking classes, also for approximate nearest neighbor search.

Logo minFunc 2012

by markSchmidt - December 18, 2013, 01:07:07 CET [ Project Homepage BibTeX Download ] 1429 views, 353 downloads, 1 subscription

About: minFunc is a Matlab function for unconstrained optimization of differentiable real-valued multivariate functions using line-search methods. It uses an interface very similar to the Matlab Optimization Toolbox function fminunc, and can be called as a replacement for this function. On many problems, minFunc requires fewer function evaluations to converge than fminunc (or minimize.m). Further it can optimize problems with a much larger number of variables (fminunc is restricted to several thousand variables), and uses a line search that is robust to several common function pathologies.

Changes:

Initial Announcement on mloss.org.


Logo LIBOL 0.3.0

by stevenhoi - December 12, 2013, 15:26:14 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 9208 views, 3135 downloads, 2 subscriptions

About: LIBOL is an open-source library with a family of state-of-the-art online learning algorithms for machine learning and big data analytics research. The current version supports 16 online algorithms for binary classification and 13 online algorithms for multiclass classification.

Changes:

In contrast to our last version (V0.2.3), the new version (V0.3.0) has made some important changes as follows:

• Add a template and guide for adding new algorithms;

• Improve parameter settings and make documentation clear;

• Improve documentation on data formats and key functions;

• Amend the "OGD" function to use different loss types;

• Fixed some name inconsistency and other minor bugs.


About: The glm-ie toolbox contains scalable estimation routines for GLMs (generalised linear models) and SLMs (sparse linear models) as well as an implementation of a scalable convex variational Bayesian inference relaxation. We designed the glm-ie package to be simple, generic and easily expansible. Most of the code is written in Matlab including some MEX files. The code is fully compatible to both Matlab 7.x and GNU Octave 3.2.x. Probabilistic classification, sparse linear modelling and logistic regression are covered in a common algorithmical framework allowing for both MAP estimation and approximate Bayesian inference.

Changes:

added factorial mean field inference as a third algorithm complementing expectation propagation and variational Bayes

generalised non-Gaussian potentials so that affine instead of linear functions of the latent variables can be used


Logo FABIA 2.8.0

by hochreit - October 18, 2013, 10:14:57 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 10096 views, 2120 downloads, 1 subscription

Rating Whole StarWhole StarWhole StarWhole Star1/2 Star
(based on 1 vote)

About: FABIA is a biclustering algorithm that clusters rows and columns of a matrix simultaneously. Consequently, members of a row cluster are similar to each other on a subset of columns and, analogously, members of a column cluster are similar to each other on a subset of rows. Biclusters are found by factor analysis where both the factors and the loading matrix are sparse. FABIA is a multiplicative model that extracts linear dependencies between samples and feature patterns. Applications include detection of transcriptional modules in gene expression data and identification of haplotypes/>identity by descent< consisting of rare variants obtained by next generation sequencing.

Changes:

CHANGES IN VERSION 2.8.0

NEW FEATURES

o rescaling of lapla
o extractPlot does not plot sorted matrices

CHANGES IN VERSION 2.4.0

o spfabia bugfixes

CHANGES IN VERSION 2.3.1

NEW FEATURES

o Getters and setters for class Factorization

2.0.0:

  • spfabia: fabia for a sparse data matrix (in sparse matrix format) and sparse vector/matrix computations in the code to speed up computations. spfabia applications: (a) detecting >identity by descent< in next generation sequencing data with rare variants, (b) detecting >shared haplotypes< in disease studies based on next generation sequencing data with rare variants;
  • fabia for non-negative factorization (parameter: non_negative);
  • changed to C and removed dependencies to Rcpp;
  • improved update for lambda (alpha should be smaller, e.g. 0.03);
  • introduced maximal number of row elements (lL);
  • introduced cycle bL when upper bounds nL or lL are effective;
  • reduced computational complexity;
  • bug fixes: (a) update formula for lambda: tighter approximation, (b) corrected inverse of the conditional covariance matrix of z;

1.4.0:

  • New option nL: maximal number of biclusters per row element;
  • Sort biclusters according to information content;
  • Improved and extended preprocessing;
  • Update to R2.13

Logo MLlib 0.8

by atalwalkar - October 10, 2013, 00:56:25 CET [ Project Homepage BibTeX Download ] 2031 views, 419 downloads, 1 subscription

About: MLlib provides a distributed machine learning (ML) library to address the growing need for scalable ML. MLlib is developed in Spark (http://spark.incubator.apache.org/), a cluster computing system designed for iterative computation. Moreover, it is a component of a larger system called MLbase (www.mlbase.org) that aims to provide user-friendly distributed ML functionality both for ML researchers and domain experts. MLlib currently consists of scalable implementations of algorithms for classification, regression, collaborative filtering and clustering.

Changes:

Initial Announcement on mloss.org.


Logo Implementation of the DMV and CCM Parsers 0.2.0

by francolq - September 24, 2013, 07:06:46 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1326 views, 299 downloads, 1 subscription

About: This package includes implementations of the CCM, DMV and DMV+CCM parsers from Klein and Manning (2004), and code for testing them with the WSJ, Negra and Cast3LB corpuses (English, German and Spanish respectively). A detailed description of the parsers can be found in Klein (2005).

Changes:

Initial Announcement on mloss.org.


Logo Test 1.0

by willyie - August 23, 2013, 23:05:22 CET [ BibTeX Download ] 788 views, 268 downloads, 1 subscription

About: Test submission. Is MLOSS working?

Changes:

Initial Announcement on mloss.org.


Logo CIlib Computational Intelligence Library 0.8

by gpampara - August 22, 2013, 08:34:21 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1647 views, 500 downloads, 1 subscription

About: CIlib is a library of computational intelligence algorithms and supporting components that allows simple extension and experimentation. The library is peer reviewed and is backed by a leading research group in the field. The library is under active development.

Changes:

Initial Announcement on mloss.org.


About: Stochastic neighbor embedding originally aims at the reconstruction of given distance relations in a low-dimensional Euclidean space. This can be regarded as general approach to multi-dimensional scaling, but the reconstruction is based on the definition of input (and output) neighborhood probability alone. The present implementation also allows for handling dissimilarity or score-induced neighborhood topologies and makes use of quasi 2nd order gradient-based (l-)BFGS optimization.

Changes:
  • gradient in xsne_fun.m fixed! (constant factor m was missing)

  • symmetry option re-introduced allowing for enabling symmetric and asymmetric versions of SNE and t-SNE


Logo cbMDS Correlation Based Multi Dimensional Scaling 1.2

by emstrick - July 27, 2013, 14:35:36 CET [ BibTeX BibTeX for corresponding Paper Download ] 3805 views, 936 downloads, 1 subscription

About: The aim is to embed a given data relationship matrix into a low-dimensional Euclidean space such that the point distances / distance ranks correlate best with the original input relationships. Input relationships may be given as (sparse) (asymmetric) distance, dissimilarity, or (negative!) score matrices. Input-output relations are modeled as low-conditioned. (Weighted) Pearson and soft Spearman rank correlation, and unweighted soft Kendall correlation are supported correlation measures for input/output object neighborhood relationships.

Changes:
  • Initial release (Ver 1.0): Weighted Pearson and correlation and soft Spearman rank correlation, Tue Dec 4 16:14:51 CET 2012

  • Ver 1.1 Added soft Kendall correlation, Fri Mar 8 08:41:09 CET 2013

  • Ver 1.2 Added reconstruction of sparse relationship matrices, Fri Jul 26 16:58:37 CEST 2013


Logo NearOED 1.0

by gabobert - July 11, 2013, 16:54:12 CET [ Project Homepage BibTeX Download ] 1247 views, 332 downloads, 1 subscription

About: The toolbox from the paper Near-optimal Experimental Design for Model Selection in Systems Biology (Busetto et al. 2013, submitted) implemented in MATLAB.

Changes:

Initial Announcement on mloss.org.


Logo BCFWstruct 1.0.0

by ppletscher - June 25, 2013, 16:19:33 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1847 views, 427 downloads, 1 subscription

About: Block-Coordinate Frank-Wolfe Optimization for Structural SVMs

Changes:

Initial Announcement on mloss.org.


Logo JMLR Jstacs 2.1

by keili - June 3, 2013, 07:32:55 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 15503 views, 3709 downloads, 2 subscriptions

About: A Java framework for statistical analysis and classification of biological sequences

Changes:

New classes:

  • MultipleIterationsCondition: Requires another TerminationCondition to fail a contiguous, specified number of times
  • ClassifierFactory: Allows for creating standard classifiers
  • SeqLogoPlotter: Plot PNG sequence logos from within Jstacs
  • MultivariateGaussianEmission: Multivariate Gaussian emission density for a Hidden Markov Model
  • MEManager: Maximum entropy model

New features and improvements:

  • Alignment: Added free shift alignment
  • PerformanceMeasure and sub-classes: Extension to weighted test data
  • AbstractClassifier, ClassifierAssessment and sub-classes: Adaption to weighted PerformanceMeasures
  • DNAAlphabet: Parser speed-up
  • PFMComparator: Extension to PFM from other sources/databases
  • ToolBox: New convenience methods for computing several statistics (e.g., median, correlation)
  • SignificantMotifOccurrencesFinder: New methods for computing PWMs and statistics from predictions
  • SequenceScore and sub-classes: New method toString(NumberFormat)
  • DataSet: Adaption to weighted data, e.g., partitioning
  • REnvironment: Changed several methods from String to CharSequence

Restructuring:

  • changed MultiDimensionalSequenceWrapperDiffSM to MultiDimensionalSequenceWrapperDiffSS

Several minor new features, bug fixes, and code cleanups


Logo Cognitive Foundry 3.3.3

by Baz - May 21, 2013, 05:59:37 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 17734 views, 2837 downloads, 2 subscriptions

About: The Cognitive Foundry is a modular Java software library of machine learning components and algorithms designed for research and applications.

Changes:
  • General:
    • Made code able to compile under both Java 1.6 and 1.7. This required removing some potentially unsafe methods that used varargs with generics.
    • Upgraded XStream dependency to 1.4.4.
    • Improved support for regression algorithms in learning.
    • Added general-purpose adapters to make it easier to compose learning algorithms and adapt their input or output.
  • Common Core:
    • Added isSparse, toArray, dotDivide, and dotDivideEquals methods for Vector and Matrix.
    • Added scaledPlus, scaledPlusEquals, scaledMinus, and scaledMinusEquals to Ring (and thus Vector and Matrix) for potentially faster such operations.
    • Fixed issue where matrix and dense vector equals was not checking for equal dimensionality.
    • Added transform, transformEquals, tranformNonZeros, and transformNonZerosEquals to Vector.
    • Made LogNumber into a signed version of a log number and moved the prior unsigned implementation into UnsignedLogNumber.
    • Added EuclideanRing interface that provides methods for times, timesEquals, divide, and divideEquals. Also added Field interface that provides methods for inverse and inverseEquals. These interfaces are now implemented by the appropriate number classes such as ComplexNumber, MutableInteger, MutableLong, MutableDouble, LogNumber, and UnsignedLogNumber.
    • Added interface for Indexer and DefaultIndexer implementation for creating a zero-based indexing of values.
    • Added interfaces for MatrixFactoryContainer and DivergenceFunctionContainer.
    • Added ReversibleEvaluator, which various identity functions implement as well as a new utility class ForwardReverseEvaluatorPair to create a reversible evaluator from a pair of other evaluators.
    • Added method to create an ArrayList from a pair of values in CollectionUtil.
    • ArgumentChecker now properly throws assertion errors for NaN values. Also added checks for long types.
    • Fixed handling of Infinity in subtraction for LogMath.
    • Fixed issue with angle method that would cause a NaN if cosine had a rounding error.
    • Added new createMatrix methods to MatrixFactory that initializes the Matrix with the given value.
    • Added copy, reverse, and isEmpty methods for several array types to ArrayUtil.
    • Added utility methods for creating a HashMap, LinkedHashMap, HashSet, or LinkedHashSet with an expected size to CollectionUtil.
    • Added getFirst and getLast methods for List types to CollectionUtil.
    • Removed some calls to System.out and Exception.printStackTrace.
  • Common Data:
    • Added create method for IdentityDataConverter.
    • ReversibleDataConverter now is an extension of ReversibleEvaluator.
  • Learning Core:
    • Added general learner transformation capability to make it easier to adapt and compose algorithms. InputOutputTransformedBatchLearner provides this capability for supervised learning algorithms by composing together a triplet. CompositeBatchLearnerPair does it for a pair of algorithms.
    • Added a constant and identity learners.
    • Added Chebyshev, Identity, and Minkowski distance metrics.
    • Added methods to DatasetUtil to get the output values for a dataset and to compute the sum of weights.
    • Made generics more permissive for supervised cost functions.
    • Added ClusterDistanceEvaluator for taking a clustering that encodes the distance from an input value to all clusters and returns the result as a vector.
    • Fixed potential round-off issue in decision tree splitter.
    • Added random subspace technique, implemented in RandomSubspace.
    • Separated functionality from LinearFunction into IdentityScalarFunction. LinearFunction by default is the same, but has parameters that can change the slope and offset of the function.
    • Default squashing function for GeneralizedLinearModel and DifferentiableGeneralizedLinearModel is now a linear function instead of an atan function.
    • Added a weighted estimator for the Poisson distribution.
    • Added Regressor interface for evaluators that are the output of (single-output) regression learning algorithms. Existing such evaluators have been updated to implement this interface.
    • Added support for regression ensembles including additive and averaging ensembles with and without weights. Added a learner for regression bagging in BaggingRegressionLearner.
    • Added a simple univariate regression class in UnivariateLinearRegression.
    • MultivariateDecorrelator now is a VectorInputEvaluator and VectorOutputEvaluator.
    • Added bias term to PrimalEstimatedSubGradient.
  • Text Core:
    • Fixed issue with the start position for tokens from LetterNumberTokenizer being off by one except for the first one.

Showing Items 21-40 of 83 on page 2 of 5: Previous 1 2 3 4 5 Next