Projects running under linux.
Showing Items 1-20 of 215 on page 1 of 11: 1 2 3 4 5 6 Next Last

About: RLLib is a lightweight C++ template library that implements incremental, standard, and gradient temporal-difference learning algorithms in Reinforcement Learning. It is an optimized library for robotic applications and embedded devices that operates under fast duty cycles (e.g., < 30 ms). RLLib has been tested and evaluated on RoboCup 3D soccer simulation agents, physical NAO V4 humanoid robots, and Tiva C series launchpad microcontrollers to predict, control, learn behaviors, and represent learnable knowledge. The implementation of the RLLib library is inspired by the RLPark API, which is a library of temporal-difference learning algorithms written in Java.

Changes:

Current release version is v2.0.


Logo WEKA 3.7.11

by mhall - April 24, 2014, 10:13:12 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 35659 views, 5069 downloads, 2 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarEmpty Star
(based on 6 votes)

About: The Weka workbench contains a collection of visualization tools and algorithms for data analysis and predictive modelling, together with graphical user interfaces for easy access to this [...]

Changes:

In core weka:

  • Bagging and RandomForest are now faster if the base learner is a WeightedInstancesHandler
  • Speed-ups for REPTree and other classes that use entropy calculations
  • Many other code improvements and speed-ups
  • Additional statistics available in the output of LinearRegression and SimpleLinearRegression. Contributed by Chris Meyer
  • Reduced memory consumption in BayesNet
  • Improvements to the package manager: load status of individual packages can now be toggled to prevent a package from loading; "Available" button now displays the latest version of all available packages that are compatible with the base version of Weka
  • RandomizableFilteredClassifier
  • Canopy clusterer
  • ImageViewer KnowledgeFlow component
  • PMML export support for Logistic. Infrastructure and changes contributed by David Person
  • Extensive tool-tips now displayed in the Explorer's scheme selector tree lists
  • Join KnowledgeFlow component for performing an inner join on two incoming streams/data sets

In packages:

  • IWSSembeded package, contributed by Pablo Bermejo
  • CVAttributeEval package, contributed by Justin Liang
  • distributedWeka package for Hadoop
  • Improvements to multiLayerPerceptrons and addtion of MLPAutoencoder
  • Code clean-up in many packages

Logo libstb 1.8

by wbuntine - April 24, 2014, 09:02:17 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3270 views, 609 downloads, 1 subscription

About: Generalised Stirling Numbers for Pitman-Yor Processes: this library provides ways of computing generalised 2nd-order Stirling numbers for Pitman-Yor and Dirichlet processes. Included is a tester and parameter optimiser. This accompanies Buntine and Hutter's article: http://arxiv.org/abs/1007.0296, and a series of papers by Buntine and students at NICTA and ANU.

Changes:

Moved repository to GitHub, and added thread support to use the main table lookups in multi-threaded code.


Logo GradMC 2.00

by tur - April 14, 2014, 15:48:48 CET [ BibTeX Download ] 1006 views, 362 downloads, 1 subscription

About: GradMC is an algorithm for MR motion artifact removal implemented in Matlab

Changes:

Added support for multi-rigid motion correction.


Logo JMLR Information Theoretical Estimators 0.57

by szzoli - April 10, 2014, 18:35:22 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 32518 views, 7011 downloads, 2 subscriptions

About: ITE (Information Theoretical Estimators) is capable of estimating many different variants of entropy, mutual information, divergence, association measures, cross quantities and kernels on distributions. Thanks to its highly modular design, ITE supports additionally (i) the combinations of the estimation techniques, (ii) the easy construction and embedding of novel information theoretical estimators, and (iii) their immediate application in information theoretical optimization problems.

Changes:
  • Kullback-Leibler divergence estimation based on maximum likelihood estimation + analytical formula in the chosen exponential family: added.

  • A new sampling based entropy estimator with KDE correction on the left/right sides: added.

  • Quick tests: updated with the new estimators.


Logo Somoclu 1.3.1

by peterwittek - April 10, 2014, 06:41:38 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 2201 views, 407 downloads, 2 subscriptions

About: Somoclu is a massively parallel implementation of self-organizing maps. It relies on OpenMP for multicore execution, MPI for distributing the workload, and it can be accelerated by CUDA on a GPU cluster. A sparse kernel is also included, which is useful for training maps on vector spaces generated in text mining processes.

Changes:
  • Initial Windows support through GCC on Windows.
  • Better I/O separation for the Python, R, and MATLAB interfaces.
  • Bug fixes: major MPI initialization bug fixed.

Logo MShadow 1.0

by antinucleon - April 10, 2014, 02:57:54 CET [ Project Homepage BibTeX Download ] 216 views, 34 downloads, 1 subscription

About: Lightweight CPU/GPU Matrix/Tensor Template Library in C++/CUDA. Support element-wise expression expand in high performance. Code once, run smoothly on both GPU and CPU

Changes:

Initial Announcement on mloss.org.


Logo CXXNET 0.1

by antinucleon - April 10, 2014, 02:47:08 CET [ Project Homepage BibTeX Download ] 219 views, 31 downloads, 1 subscription

About: CXXNET (spelled as: C plus plus net) is a neural network toolkit build on mshadow(https://github.com/tqchen/mshadow). It is yet another implementation of (convolutional) neural network. It is in C++, with about 1000 lines of network layer implementations, easily configuration via config file, and can get the state of art performance.

Changes:

Initial Announcement on mloss.org.


Logo JMLR Tapkee 1.0

by blackburn - April 10, 2014, 02:45:58 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 4211 views, 1125 downloads, 1 subscription

About: Tapkee is an efficient and flexible C++ template library for dimensionality reduction.

Changes:

Initial Announcement on mloss.org.


Logo JMLR dlib ml 18.7

by davis685 - April 10, 2014, 01:47:08 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 69460 views, 12152 downloads, 2 subscriptions

About: This project is a C++ toolkit containing machine learning algorithms and tools that facilitate creating complex software in C++ to solve real world problems.

Changes:

The major new feature in this release is a Python API for training histogram-of-oriented-gradient based object detectors and examples showing how to use this type of detector to perform real-time face detection. Additionally, this release also adds simpler interfaces for learning to solve assignment and multi-target tracking problems.


Logo JMLR MOA Massive Online Analysis Nov-13

by abifet - April 4, 2014, 03:50:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 9619 views, 3942 downloads, 1 subscription

About: Massive Online Analysis (MOA) is a real time analytic tool for data streams. It is a software environment for implementing algorithms and running experiments for online learning from evolving data streams. MOA includes a collection of offline and online methods as well as tools for evaluation. In particular, it implements boosting, bagging, and Hoeffding Trees, all with and without Naive Bayes classifiers at the leaves. MOA supports bi-directional interaction with WEKA, the Waikato Environment for Knowledge Analysis, and it is released under the GNU GPL license.

Changes:

New version November 2013


Logo JMLR MultiBoost 1.2.02

by busarobi - March 31, 2014, 16:13:04 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 19573 views, 3422 downloads, 1 subscription

About: MultiBoost is a multi-purpose boosting package implemented in C++. It is based on the multi-class/multi-task AdaBoost.MH algorithm [Schapire-Singer, 1999]. Basic base learners (stumps, trees, products, Haar filters for image processing) can be easily complemented by new data representations and the corresponding base learners, without interfering with the main boosting engine.

Changes:

Major changes :

  • The “early stopping” feature can now based on any metric output with the --outputinfo command line argument.

  • Early stopping now works with --slowresume command line argument.

Minor fixes:

  • More informative output when testing.

  • Various compilation glitch with recent clang (OsX/Linux).


Logo JMLR EnsembleSVM 2.0

by claesenm - March 31, 2014, 08:06:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 3154 views, 1092 downloads, 1 subscription

About: The EnsembleSVM library offers functionality to perform ensemble learning using Support Vector Machine (SVM) base models. In particular, we offer routines for binary ensemble models using SVM base classifiers. Experimental results have shown the predictive performance to be comparable with standard SVM models but with drastically reduced training time. Ensemble learning with SVM models is particularly useful for semi-supervised tasks.

Changes:

The library has been updated and features a variety of new functionality as well as more efficient implementations of original features. The following key improvements have been made:

  1. Support for multithreading in training and prediction with ensemble models. Since both of these are embarassingly parallel, this has induced a significant speedup (3-fold on quad-core).
  2. Extensive programming framework for aggregation of base model predictions which allows highly efficient prototyping of new aggregation approaches. Additionally we provide several predefined strategies, including (weighted) majority voting, logistic regression and nonlinear SVMs of your choice -- be sure to check out the esvm-edit tool! The provided framework also allows you to efficiently program your own, novel aggregation schemes.
  3. Full code transition to C++11, the latest C++ standard, which enabled various performance improvements. The new release requires moderately recent compilers, such as gcc 4.7.2+ or clang 3.2+.
  4. Generic implementations of convenient facilities have been added, such as thread pools, deserialization factories and more.

The API and ABI have undergone significant changes, many of which are due to the transition to C++11.


Logo Libra 1.0.1

by lowd - March 30, 2014, 09:42:00 CET [ Project Homepage BibTeX Download ] 8721 views, 1814 downloads, 1 subscription

About: The Libra Toolkit is a collection of algorithms for learning and inference with discrete probabilistic models, including Bayesian networks, Markov networks, dependency networks, sum-product networks, arithmetic circuits, and mixtures of trees.

Changes:

Version 1.0.1 (3/30/2014):

  • Several new algorithms -- acmn, learning ACs using MNs; idspn, SPN structure learning; mtlearn, learning mixtures of trees
  • Several new support programs -- spquery, for exact inference in SPNs; spn2ac, for converting SPNs to ACs
  • Renamed aclearnstruct to acbn
  • Replaced aclearnstruct -noac with separate bnlearn program
  • ...and many more small changes and fixes, throughout!

Logo XGBoost v0.1

by crowwork - March 27, 2014, 07:09:52 CET [ Project Homepage BibTeX Download ] 290 views, 46 downloads, 1 subscription

About: eXtreme gradient boosting (tree) library. Features: - Sparse feature format allows easy handling of missing values, and improve computation efficiency. - Efficient parallel implementation that optimizes memory and computation.

Changes:

Initial Announcement on mloss.org.


Logo BayesOpt, a Bayesian Optimization toolbox 0.6

by rmcantin - March 26, 2014, 17:48:17 CET [ Project Homepage BibTeX Download ] 4450 views, 1010 downloads, 2 subscriptions

About: BayesOpt is an efficient, C++ implementation of the Bayesian optimization methodology for nonlinear-optimization, experimental design and stochastic bandits. In the literature it is also called Sequential Kriging Optimization (SKO) or Efficient Global Optimization (EGO). There are also interfaces for C, Matlab/Octave and Python.

Changes:

-Complete refactoring of inner parts of the library. The code is easier to understand/modify and it allow simpler integration with new algorithms.

-Updated to the latest version of NLOPT (2.4.1). Wrapper code symplified.

-Error codes replaced with exceptions in C++ interface. Library is exception safe.

-API modified to support new learning methods for kernel hyperparameters (e.g: MCMC). Warning: config parameters about learning have changed. Code using previous versions might not work. Some of the learning methods (like MCMC) are not yet implemented.

-Added configuration of random numbers (can be fixed for debugging). Fixed issue with random numbers using different sources or random number with potential correlations. Now all the elements are guaranteed to use the same instance of the random engine.

-Improved numerical results (e.g.: hyperparameter optimization is done in log space)

-More examples and tests.

-Fixed bugs.

-The number of inner iterations have been increased by default, so overall optimization time using default configuration might be slower, but with improved results.


Logo OpenOpt 0.53

by Dmitrey - March 15, 2014, 13:37:23 CET [ Project Homepage BibTeX Download ] 36737 views, 7743 downloads, 3 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarEmpty Star
(based on 2 votes)

About: Universal Python-written numerical optimization toolbox. Problems: NLP, LP, QP, NSP, MILP, LSP, LLSP, MMP, GLP, SLE, MOP etc; general logical constraints, categorical variables, automatic differentiation, stochastic programming, interval analysis, many other goodies

Changes:

http://openopt.org/Changelog


Logo JMLR fastclime 1.2.3

by colin1898 - March 10, 2014, 08:54:41 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 673 views, 122 downloads, 1 subscription

About: The package "fastclime" provides a method of recover the precision matrix efficiently by applying parametric simplex method. The computation is based on a linear optimization solver. It also contains a generic LP solver and a parameterized LP solver using parametric simplex method.

Changes:

Initial Announcement on mloss.org.


Logo HierLearning 1.0

by neville - March 2, 2014, 04:24:37 CET [ BibTeX BibTeX for corresponding Paper Download ] 488 views, 94 downloads, 1 subscription

About: HierLearning is a C++11 implementation of a general-purpose, multi-agent, hierarchical reinforcement learning system for sequential decision problems.

Changes:

Initial Announcement on mloss.org.


Logo Armadillo library 4.100

by cu24gjf - February 28, 2014, 07:53:24 CET [ Project Homepage BibTeX Download ] 36593 views, 8162 downloads, 2 subscriptions

Rating Whole StarWhole StarWhole StarWhole StarEmpty Star
(based on 2 votes)

About: Armadillo is a template C++ linear algebra library aiming towards a good balance between speed and ease of use, with a function syntax similar to MATLAB. Matrix decompositions are provided through optional integration with LAPACK, or one of its high performance drop-in replacements (eg. Intel MKL, OpenBLAS).

Changes:
  • added normalise() for normalising vectors to unit p-norm
  • extended the field class to handle 3D layout
  • extended eigs_sym() and eigs_gen() to obtain eigenvalues of various forms (eg. largest or smallest magnitude)
  • automatic SIMD vectorisation of elementary expressions (eg. matrix addition) when using Clang 3.4+ with -O3 optimisation
  • faster handling of sparse submatrix views
  • workaround for a bug in LAPACK 3.4

Showing Items 1-20 of 215 on page 1 of 11: 1 2 3 4 5 6 Next Last