All entries.
Showing Items 71-80 of 542 on page 8 of 55: First Previous 3 4 5 6 7 8 9 10 11 12 13 Next Last

Logo WEKA 3.7.11

by mhall - April 24, 2014, 10:13:12 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 40171 views, 5850 downloads, 2 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarEmpty Star
(based on 6 votes)

About: The Weka workbench contains a collection of visualization tools and algorithms for data analysis and predictive modelling, together with graphical user interfaces for easy access to this [...]

Changes:

In core weka:

  • Bagging and RandomForest are now faster if the base learner is a WeightedInstancesHandler
  • Speed-ups for REPTree and other classes that use entropy calculations
  • Many other code improvements and speed-ups
  • Additional statistics available in the output of LinearRegression and SimpleLinearRegression. Contributed by Chris Meyer
  • Reduced memory consumption in BayesNet
  • Improvements to the package manager: load status of individual packages can now be toggled to prevent a package from loading; "Available" button now displays the latest version of all available packages that are compatible with the base version of Weka
  • RandomizableFilteredClassifier
  • Canopy clusterer
  • ImageViewer KnowledgeFlow component
  • PMML export support for Logistic. Infrastructure and changes contributed by David Person
  • Extensive tool-tips now displayed in the Explorer's scheme selector tree lists
  • Join KnowledgeFlow component for performing an inner join on two incoming streams/data sets

In packages:

  • IWSSembeded package, contributed by Pablo Bermejo
  • CVAttributeEval package, contributed by Justin Liang
  • distributedWeka package for Hadoop
  • Improvements to multiLayerPerceptrons and addtion of MLPAutoencoder
  • Code clean-up in many packages

Logo libstb 1.8

by wbuntine - April 24, 2014, 09:02:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 5067 views, 969 downloads, 1 subscription

About: Generalised Stirling Numbers for Pitman-Yor Processes: this library provides ways of computing generalised 2nd-order Stirling numbers for Pitman-Yor and Dirichlet processes. Included is a tester and parameter optimiser. This accompanies Buntine and Hutter's article: http://arxiv.org/abs/1007.0296, and a series of papers by Buntine and students at NICTA and ANU.

Changes:

Moved repository to GitHub, and added thread support to use the main table lookups in multi-threaded code.


Logo libAGF 0.9.7

by Petey - April 15, 2014, 04:55:41 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 8462 views, 1689 downloads, 1 subscription

About: C++ software for statistical classification, probability estimation and interpolation/non-linear regression using variable bandwidth kernel estimation.

Changes:

New in Version 0.9.7:

  • multi-class classification generalizes class-borders algorithm using a recursive control language
  • hierarchical clustering
  • improved pre-processing

Logo GradMC 2.00

by tur - April 14, 2014, 15:48:48 CET [ BibTeX Download ] 1708 views, 572 downloads, 1 subscription

About: GradMC is an algorithm for MR motion artifact removal implemented in Matlab

Changes:

Added support for multi-rigid motion correction.


Logo MShadow 1.0

by antinucleon - April 10, 2014, 02:57:54 CET [ Project Homepage BibTeX Download ] 811 views, 227 downloads, 1 subscription

About: Lightweight CPU/GPU Matrix/Tensor Template Library in C++/CUDA. Support element-wise expression expand in high performance. Code once, run smoothly on both GPU and CPU

Changes:

Initial Announcement on mloss.org.


Logo CXXNET 0.1

by antinucleon - April 10, 2014, 02:47:08 CET [ Project Homepage BibTeX Download ] 917 views, 231 downloads, 1 subscription

About: CXXNET (spelled as: C plus plus net) is a neural network toolkit build on mshadow(https://github.com/tqchen/mshadow). It is yet another implementation of (convolutional) neural network. It is in C++, with about 1000 lines of network layer implementations, easily configuration via config file, and can get the state of art performance.

Changes:

Initial Announcement on mloss.org.


Logo JMLR Tapkee 1.0

by blackburn - April 10, 2014, 02:45:58 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 6104 views, 1753 downloads, 1 subscription

About: Tapkee is an efficient and flexible C++ template library for dimensionality reduction.

Changes:

Initial Announcement on mloss.org.


Logo JMLR MOA Massive Online Analysis Nov-13

by abifet - April 4, 2014, 03:50:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 11519 views, 4520 downloads, 1 subscription

About: Massive Online Analysis (MOA) is a real time analytic tool for data streams. It is a software environment for implementing algorithms and running experiments for online learning from evolving data streams. MOA includes a collection of offline and online methods as well as tools for evaluation. In particular, it implements boosting, bagging, and Hoeffding Trees, all with and without Naive Bayes classifiers at the leaves. MOA supports bi-directional interaction with WEKA, the Waikato Environment for Knowledge Analysis, and it is released under the GNU GPL license.

Changes:

New version November 2013


Logo SAMOA 0.0.1

by gdfm - April 2, 2014, 17:09:08 CET [ Project Homepage BibTeX Download ] 691 views, 200 downloads, 1 subscription

About: SAMOA is a platform for mining big data streams. It is a distributed streaming machine learning (ML) framework that contains a programing abstraction for distributed streaming ML algorithms.

Changes:

Initial Announcement on mloss.org.


Logo JMLR MultiBoost 1.2.02

by busarobi - March 31, 2014, 16:13:04 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 23639 views, 4138 downloads, 1 subscription

About: MultiBoost is a multi-purpose boosting package implemented in C++. It is based on the multi-class/multi-task AdaBoost.MH algorithm [Schapire-Singer, 1999]. Basic base learners (stumps, trees, products, Haar filters for image processing) can be easily complemented by new data representations and the corresponding base learners, without interfering with the main boosting engine.

Changes:

Major changes :

  • The “early stopping” feature can now based on any metric output with the --outputinfo command line argument.

  • Early stopping now works with --slowresume command line argument.

Minor fixes:

  • More informative output when testing.

  • Various compilation glitch with recent clang (OsX/Linux).


Showing Items 71-80 of 542 on page 8 of 55: First Previous 3 4 5 6 7 8 9 10 11 12 13 Next Last