All entries.
Showing Items 501-510 of 620 on page 51 of 62: First Previous 46 47 48 49 50 51 52 53 54 55 56 Next Last

About: The package provides a Lagrangian approach to the posterior regularization of given linear mappings. This is important in two cases, (a) when systems are under-determined and (b) when the external model for calculating the mapping is invariant to properties such as scaling. The software may be applied in cases when the external model does not provide its own regularization strategy. In addition, the package allows to rank attributes according to their distortion potential to a given linear mapping.

Changes:

Version 1.1 (May 23, 2012) memory and time optimizations distderivrel.m now supports assessing the relevance of attribute pairs

Version 1.0 (Nov 9, 2011) * Initial Announcement on mloss.org.


Logo Reranker Framework 1.0

by zenog - October 29, 2012, 10:05:30 CET [ Project Homepage BibTeX Download ] 2225 views, 698 downloads, 1 subscription

About: ReFr is a software architecture for specifying, training and using reranking models.

Changes:

Initial Announcement on mloss.org.


Logo reserbot alpha 1

by neuromancer - January 31, 2011, 14:27:18 CET [ Project Homepage BibTeX Download ] 6060 views, 1597 downloads, 1 subscription

About: A chatterbot that learns natural languages learning from imitation.

Changes:

Alpha 1 - Codename: Wendell Borton ("Bllluuhhhhh...!!")

Short term memory greatly improved.


Logo revrand 0.3

by dsteinberg - April 29, 2016, 07:31:27 CET [ Project Homepage BibTeX Download ] 1669 views, 353 downloads, 3 subscriptions

About: A library of scalable Bayesian generalised linear models with fancy features

Changes:
  • Simplification of all of the algorithm interfaces by using Parameter (bounded) types
  • Re-factoring of the modules in the library to make it more user friendly

Logo RFD 1.0

by openpr_nlpr - April 28, 2014, 10:34:57 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2567 views, 540 downloads, 1 subscription

About: This is an unoptimized implementation of the RFD binary descriptor, which is published in the following paper. B. Fan, et al. Receptive Fields Selection for Binary Feature Description. IEEE Transaction on Image Processing, 2014. doi: http://dx.doi.org/10.1109/TIP.2014.2317981

Changes:

Initial Announcement on mloss.org.


Logo RiVal 0.1

by alansaid - July 29, 2015, 12:39:54 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1457 views, 407 downloads, 2 subscriptions

About: Rival is an open source Java toolkit for recommender system evaluation. It provides a simple way to create evaluation results comparable across different recommendation frameworks.

Changes:

Initial Announcement on mloss.org.


Logo JMLR RL Glue and Codecs -- Glue 3.x and Codecs

by btanner - October 12, 2009, 07:50:03 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 20149 views, 2266 downloads, 1 subscription

About: RL-Glue allows agents, environments, and experiments written in Java, C/C++, Matlab, Python, and Lisp to inter operate, accelerating research by promoting software re-use in the community.

Changes:

RL-Glue paper has been published in JMLR.


Logo JMLR RL library 3.00.00

by frezza - January 13, 2015, 04:15:16 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3044 views, 871 downloads, 2 subscriptions

About: A template based C++ reinforcement learning library

Changes:

Initial Announcement on mloss.org.


About: RLLib is a lightweight C++ template library that implements incremental, standard, and gradient temporal-difference learning algorithms in Reinforcement Learning. It is an optimized library for robotic applications and embedded devices that operates under fast duty cycles (e.g., < 30 ms). RLLib has been tested and evaluated on RoboCup 3D soccer simulation agents, physical NAO V4 humanoid robots, and Tiva C series launchpad microcontrollers to predict, control, learn behaviors, and represent learnable knowledge. The implementation of the RLLib library is inspired by the RLPark API, which is a library of temporal-difference learning algorithms written in Java.

Changes:

Current release version is v2.0.


Logo RLPy 1.3a

by bobklein2 - August 28, 2014, 14:34:35 CET [ Project Homepage BibTeX Download ] 4774 views, 1023 downloads, 1 subscription

About: RLPy is a framework for performing reinforcement learning (RL) experiments in Python. RLPy provides a large library of agent and domain components, and a suite of tools to aid in experiments (parallelization, hyperparameter optimization, code profiling, and plotting).

Changes:
  • Fixed bug where results using same random seed were different with visualization turned on/off
  • Created RLPy package on pypi (Available at https://pypi.python.org/pypi/rlpy)
  • Switched from custom logger class to python default
  • Added unit tests
  • Code readability improvements (formatting, variable names/ordering)
  • Restructured TD Learning heirarchy
  • Updated tutorials

Showing Items 501-510 of 620 on page 51 of 62: First Previous 46 47 48 49 50 51 52 53 54 55 56 Next Last