Projects that are tagged with decision tree learning.


Logo Cognitive Foundry 3.4.0

by Baz - April 3, 2015, 08:28:14 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 18672 views, 3036 downloads, 2 subscriptions

About: The Cognitive Foundry is a modular Java software library of machine learning components and algorithms designed for research and applications.

Changes:
  • General:
    • Now requires Java 1.7 or higher.
    • Improved compatibility with Java 1.8 functions by removing ClonableSerializable requirement from many function-style interfaces.
  • Common Core:
    • Improved iteration speed over sparse MTJ vectors.
    • Added utility methods for more stable log(1+x), exp(1-x), log(1 - exp(x)), and log(1 + exp(x)) to LogMath.
    • Added method for creating a partial permutations to Permutation.
    • Added methods for computing standard deviation to UnivariateStatisticsUtil.
    • Added increment, decrement, and list view methods to Vector and Matrix.
    • Added shorter versions of get and set for Vector and Matrix getElement and setElement.
    • Added aliases of dot for dotProduct in VectorSpace.
    • Added utility methods for divideByNorm2 to VectorUtil.
  • Learning:
    • Added a learner for a Factorization Machine using SGD.
    • Added a iterative reporter for validation set performance.
    • Added new methods to statistical distribution classes to allow for faster sampling without boxing, in batches, or without creating extra memory.
    • Made generics for performance evaluators more permissive.
    • ParameterGradientEvaluator changed to not require input, output, and gradient types to be the same. This allows more sane gradient definitions for scalar functions.
    • Added parameter to enforce a minimum size in a leaf node for decision tree learning. It is configured through the splitting function.
    • Added ability to filter which dimensions to use in the random subspace and variance tree node splitter.
    • Added ReLU, leaky ReLU, and soft plus activation functions for neural networks.
    • Added IntegerDistribution interface for distributions over natural numbers.
    • Added a method to get the mean of a numeric distribution without boxing.
    • Fixed an issue in DefaultDataDistribution that caused the total to be off when a value was set to less than or equal to 0.
    • Added property for rate to GammaDistribution.
    • Added method to get standard deviation from a UnivariateGaussian.
    • Added clone operations for decision tree classes.
    • Fixed issue TukeyKramerConfidence interval computation.
    • Fixed serialization issue with SMO output.

Logo fertilized forests 1.0beta

by Chrisl_S - January 23, 2015, 16:04:31 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 664 views, 137 downloads, 1 subscription

About: The fertilized forests project has the aim to provide an easy to use, easy to extend, yet fast library for decision forests. It summarizes the research in this field and provides a solid platform to extend it. Offering consistent interfaces to C++, Python and Matlab and being available for all major compilers gives the user high flexibility for using the library.

Changes:

Initial Announcement on mloss.org.


Logo Neural network designer 1.1.1

by bragi - December 28, 2012, 11:38:10 CET [ Project Homepage BibTeX Download ] 4164 views, 998 downloads, 1 subscription

About: a dbms for resonating neural networks. Create and use different types of machine learning algorithms.

Changes:

AIML compatible (AIML files can be imported); new 'Grid channel' for developing board games; improved topics editor; new demo project: ALice (from AIML); lots of bug-fixes and speed improvements


Logo Milk 0.5

by luispedro - November 7, 2012, 13:08:28 CET [ Project Homepage BibTeX Download ] 23457 views, 5700 downloads, 1 subscription

Rating Whole StarWhole StarWhole StarEmpty StarEmpty Star
(based on 2 votes)

About: Python Machine Learning Toolkit

Changes:

Added LASSO (using coordinate descent optimization). Made SVM classification (learning and applying) much faster: 2.5x speedup on yeast UCI dataset.


Logo treelearn 1

by iskander - September 21, 2011, 16:12:27 CET [ Project Homepage BibTeX Download ] 2635 views, 648 downloads, 1 subscription

About: A python implementation of Breiman's Random Forests.

Changes:

Initial Announcement on mloss.org.


Logo QuickDT 0.1

by sanity - September 21, 2011, 13:43:37 CET [ Project Homepage BibTeX Download ] 2602 views, 768 downloads, 1 subscription

About: A decision tree learner that is designed to be reasonably fast, but the primary goal is ease of use

Changes:

Initial Announcement on mloss.org.


Logo Cubist 2.07

by zenog - September 2, 2011, 14:52:17 CET [ Project Homepage BibTeX Download ] 2664 views, 707 downloads, 1 subscription

About: Cubist is the regression counterpart to the C5.0 decision tree tool.

Changes:

Initial Announcement on mloss.org.


Logo C5.0 2.07

by zenog - September 2, 2011, 14:49:04 CET [ Project Homepage BibTeX Download ] 2877 views, 728 downloads, 1 subscription

About: C5.0 is the successor of the C4.5 decision tree algorithm/tool. In particular, it is faster and more memory-efficient.

Changes:

Initial Announcement on mloss.org.


Logo TiMBL 6.1

by antalvdb - January 11, 2008, 09:20:57 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 5761 views, 1306 downloads, 0 comments, 0 subscriptions

About: The TiMBL software package is a fast, decision-tree-based implementation of k-nearest neighbor classification. The package includes the IB1, IB2, TRIBL, TRIBL2, and IGTree algorithms, and offers [...]

Changes:

Initial Announcement on mloss.org.