5 projects found that use julia as the programming language.


Logo Somoclu 1.7.4

by peterwittek - June 6, 2017, 15:48:11 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 32720 views, 5892 downloads, 3 subscriptions

About: Somoclu is a massively parallel implementation of self-organizing maps. It relies on OpenMP for multicore execution, MPI for distributing the workload, and it can be accelerated by CUDA on a GPU cluster. A sparse kernel is also included, which is useful for training maps on vector spaces generated in text mining processes. Apart from a command line interface, Python, Julia, R, and MATLAB are supported.

Changes:
  • New: Verbosity parameter in the command-line, Python, MATLAB, and Julia interfaces.
  • Changed: Calculation of U-matrix parallelized.
  • Changed: Moved feeding data to train method in the Python interface.
  • Fixed: The random seed was set to 0 for testing purposes. This is now changed to a wall-time based initialization.
  • Fixed: Sparse matrix reader made more robust.
  • Fixed: Compatibility with kohonen 3 resolved.
  • Fixed: Compatibility with Matplotlib 2 resolved.

Logo LogRegCrowds, Logistic Regression from Crowds 1.0

by fmpr - January 16, 2017, 18:10:57 CET [ Project Homepage BibTeX Download ] 4365 views, 1099 downloads, 3 subscriptions

About: LogReg-Crowds is a collection of Julia implementations of various approaches for learning a logistic regression model multiple annotators and crowds, namely the works of Raykar et al. (2010), Rodrigues et al. (2013) and Dawid and Skene (1979).

Changes:

Initial Announcement on mloss.org. Added GitHub page.


About: Efficient and Flexible Distributed/Mobile Deep Learning Framework, for python, R, Julia and more

Changes:

This version comes with Distributed and Mobile Examples


Logo SALSA.jl 0.0.5

by jumutc - September 28, 2015, 17:28:56 CET [ Project Homepage BibTeX Download ] 2560 views, 557 downloads, 1 subscription

About: SALSA (Software lab for Advanced machine Learning with Stochastic Algorithms) is an implementation of the well-known stochastic algorithms for Machine Learning developed in the high-level technical computing language Julia. The SALSA software package is designed to address challenges in sparse linear modelling, linear and non-linear Support Vector Machines applied to large data samples with user-centric and user-friendly emphasis.

Changes:

Initial Announcement on mloss.org.


Logo XGBoost v0.4.0

by crowwork - May 12, 2015, 08:57:16 CET [ Project Homepage BibTeX Download ] 19355 views, 3392 downloads, 3 subscriptions

About: xgboost: eXtreme Gradient Boosting It is an efficient and scalable implementation of gradient boosting framework. The package includes efficient linear model solver and tree learning algorithm. The package can automatically do parallel computation with OpenMP, and it can be more than 10 times faster than existing gradient boosting packages such as gbm or sklearn.GBM . It supports various objective functions, including regression, classification and ranking. The package is made to be extensible, so that user are also allowed to define there own objectives easily. The newest version of xgboost now supports distributed learning on various platforms such as hadoop, mpi and scales to even larger problems

Changes:
  • Distributed version of xgboost that runs on YARN, scales to billions of examples

  • Direct save/load data and model from/to S3 and HDFS

  • Feature importance visualization in R module, by Michael Benesty

  • Predict leaf index

  • Poisson regression for counts data

  • Early stopping option in training

  • Native save load support in R and python

  • xgboost models now can be saved using save/load in R

  • xgboost python model is now pickable

  • sklearn wrapper is supported in python module

  • Experimental External memory version