Projects supporting the csv data format.
Showing Items 21-40 of 44 on page 2 of 3: Previous 1 2 3 Next

Logo JMLR Waffles 2014-07-05

by mgashler - July 20, 2014, 04:53:54 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 27529 views, 7911 downloads, 2 subscriptions

About: Script-friendly command-line tools for machine learning and data mining tasks. (The command-line tools wrap functionality from a public domain C++ class library.)

Changes:

Added support for CUDA GPU-parallelized neural network layers, and several other new features. Full list of changes at http://waffles.sourceforge.net/docs/changelog.html


Logo Encog Machine Learning Framework 3.2

by jeffheaton - July 5, 2014, 23:47:06 CET [ Project Homepage BibTeX Download ] 3734 views, 1025 downloads, 1 subscription

About: Encog is a Machine Learning framework for Java, C#, Javascript and C/C++ that supports SVM's, Genetic Programming, Bayesian Networks, Hidden Markov Models and other algorithms.

Changes:

Changes for Encog 3.2:

Issue #53: Fix Out Of Range Bug In BasicMLSequenceSet. Issue #52: Unhandled exception in Encog.Util.File.ResourceLoader.CreateStream (ResourceLoader.cs) Issue #50: Concurrency bugs in PruneIncremental Issue #48: Unit Tests Failing - TestHessian Issue #46: Couple of small fixes - Temporal DataSet and SCG training Issue #45: Fixed EndMinutesStrategy to correctly evaluate ShouldStop after the specified number of minutes have elapsed. Issue #44: Encog.ML.Data.Basic.BasicMLDataPairCentroid.Add() & .Remove() Issue #43: Unit Tests Failing - Matrix not full rank Issue #42: Nuget - NuSpec Issue #36: Load Examples easier


Logo JMLR MultiBoost 1.2.02

by busarobi - March 31, 2014, 16:13:04 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 27323 views, 4747 downloads, 1 subscription

About: MultiBoost is a multi-purpose boosting package implemented in C++. It is based on the multi-class/multi-task AdaBoost.MH algorithm [Schapire-Singer, 1999]. Basic base learners (stumps, trees, products, Haar filters for image processing) can be easily complemented by new data representations and the corresponding base learners, without interfering with the main boosting engine.

Changes:

Major changes :

  • The “early stopping” feature can now based on any metric output with the --outputinfo command line argument.

  • Early stopping now works with --slowresume command line argument.

Minor fixes:

  • More informative output when testing.

  • Various compilation glitch with recent clang (OsX/Linux).


Logo JMLR EnsembleSVM 2.0

by claesenm - March 31, 2014, 08:06:20 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7063 views, 2532 downloads, 2 subscriptions

About: The EnsembleSVM library offers functionality to perform ensemble learning using Support Vector Machine (SVM) base models. In particular, we offer routines for binary ensemble models using SVM base classifiers. Experimental results have shown the predictive performance to be comparable with standard SVM models but with drastically reduced training time. Ensemble learning with SVM models is particularly useful for semi-supervised tasks.

Changes:

The library has been updated and features a variety of new functionality as well as more efficient implementations of original features. The following key improvements have been made:

  1. Support for multithreading in training and prediction with ensemble models. Since both of these are embarassingly parallel, this has induced a significant speedup (3-fold on quad-core).
  2. Extensive programming framework for aggregation of base model predictions which allows highly efficient prototyping of new aggregation approaches. Additionally we provide several predefined strategies, including (weighted) majority voting, logistic regression and nonlinear SVMs of your choice -- be sure to check out the esvm-edit tool! The provided framework also allows you to efficiently program your own, novel aggregation schemes.
  3. Full code transition to C++11, the latest C++ standard, which enabled various performance improvements. The new release requires moderately recent compilers, such as gcc 4.7.2+ or clang 3.2+.
  4. Generic implementations of convenient facilities have been added, such as thread pools, deserialization factories and more.

The API and ABI have undergone significant changes, many of which are due to the transition to C++11.


Logo Chordalysis 1.0

by fpetitjean - March 24, 2014, 01:22:06 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1412 views, 362 downloads, 1 subscription

About: Log-linear analysis for high-dimensional data

Changes:

Initial Announcement on mloss.org.


Logo MOSIS 0.55

by claasahl - March 9, 2014, 17:35:40 CET [ BibTeX Download ] 3623 views, 1153 downloads, 2 subscriptions

About: MOSIS is a modularized framework for signal processing, stream analysis, machine learning and stream mining applications.

Changes:
  • Move "flow"-related classes into package "de.claas.mosis.flow" (e.g. Node and Link).
  • Refined and improved "flow"-related tests (e.g. Iterator and Node tests).
  • Refactored tests for data formats (e.g. PlainText and JSON tests).
  • Added visitor design pattern for graph-based functions (e.g. initialization and processing).
  • Documented parameters of Processor implementations.

Logo The Choquet Kernel 1.00

by AliFall - February 11, 2014, 16:21:15 CET [ BibTeX BibTeX for corresponding Paper Download ] 1163 views, 328 downloads, 1 subscription

About: The package computes the optimal parameters for the Choquet kernel

Changes:

Initial Announcement on mloss.org.


Logo Ordinal Choquistic Regression 1.00

by AliFall - January 30, 2014, 15:42:34 CET [ BibTeX BibTeX for corresponding Paper Download ] 1340 views, 329 downloads, 1 subscription

About: "Ordinal Choquistic Regression" model using the maximum likelihood

Changes:

Initial Announcement on mloss.org.


Logo ELKI 0.6.0

by erich - January 10, 2014, 18:32:28 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 12338 views, 2256 downloads, 3 subscriptions

About: ELKI is a framework for implementing data-mining algorithms with support for index structures, that includes a wide variety of clustering and outlier detection methods.

Changes:

Additions and Improvements from ELKI 0.5.5:

Algorithms

Clustering:

  • Hierarchical Clustering - the slower naive variants were added, and the code was refactored
  • Partition extraction from hierarchical clusterings - different linkage strategies (e.g. Ward)
  • Canopy pre-Clustering
  • Naive Mean-Shift Clustering
  • Affinity propagation clustering (both with distances and similarities / kernel functions)
  • K-means variations: Best-of-multiple-runs, bisecting k-means
  • New k-means initialization: farthest points, sample initialization
  • Cheng and Church Biclustering
  • P3C Subspace Clustering
  • One-dimensional clustering algorithm based on kernel density estimation

Outlier detection

  • COP - correlation outlier probabilities
  • LDF - a kernel density based LOF variant
  • Simplified LOF - a simpler version of LOF (not using reachability distance)
  • Simple Kernel Density LOF - a simple LOF using kernel density (more consistent than LDF)
  • Simple outlier ensemble algorithm
  • PINN - projection indexed nearest neighbors, via projected indexes.
  • ODIN - kNN graph based outlier detection
  • DWOF - Dynamic-Window Outlier Factor (contributed by Omar Yousry)
  • ABOD refactored, into ABOD, FastABOD and LBABOD

Distances

  • Geodetic distances now support different world models (WGS84 etc.) and are subtantially faster.
  • Levenshtein distances for processing strings, e.g. for analyzing phonemes (contributed code, see "Word segmentation through cross-lingual word-to-phoneme alignment", SLT2013, Stahlberg et al.)
  • Bray-Curtis, Clark, Kulczynski1 and Lorentzian distances with R-tree indexing support
  • Histogram matching distances
  • Probabilistic divergence distances (Jeffrey, Jensen-Shannon, Chi2, Kullback-Leibler)
  • Kulczynski2 similarity
  • Kernel similarity code has been refactored, and additional kernel functions have been added

Database Layer and Data Types

Projection layer * Parser for simple textual data (for use with Levenshtein distance) Various random projection families (including Feature Bagging, Achlioptas, and p-stable) Latitude+Longitude to ECEF Sparse vector improvements and bug fixes New filter: remove NaN values and missing values New filter: add histogram-based jitter New filter: normalize using statistical distributions New filter: robust standardization using Median and MAD New filter: Linear discriminant analysis (LDA)

Index Layer

  • Another speed up in R-trees
  • Refactoring of M- and R-trees: Support for different strategies in M-tree New strategies for M-tree splits Speedups in M-tree
  • New index structure: in-memory k-d-tree
  • New index structure: in-memory Locality Sensitive Hashing (LSH)
  • New index structure: approximate projected indexes, such as PINN
  • Index support for geodetic data - (Details: Geodetic Distance Queries on R-Trees for Indexing Geographic Data, SSTD13)
  • Sampled k nearest neighbors: reference KDD13 "Subsampling for Efficient and Effective Unsupervised Outlier Detection Ensembles"
  • Cached (precomputed) k-nearest neighbors to share across multiple runs
  • Benchmarking "algorithms" for indexes

Mathematics and Statistics

  • Many new distributions have been added, now 28 different distributions are supported
  • Additional estimation methods (using advanced statistics such as L-Moments), now 44 estimators are available
  • Trimming and Winsorizing
  • Automatic best-fit distribution estimation
  • Preprocessor using these distributions for rescaling data sets
  • API changes related to the new distributions support
  • More kernel density functions
  • RANSAC covariance matrix builder (unfortunately rather slow)

Visualization

  • 3D projected coordinates (Details: Interactive Data Mining with 3D-Parallel-Coordinate-Trees, SIGMOD2013)
  • Convex hulls now also include nested hierarchical clusters

Other

  • Parser speedups
  • Sparse vector bug fixes and improvements
  • Various bug fixes
  • PCA, MDS and LDA filters
  • Text output was slightly improved (but still needs to be redesigned from scratch - please contribute!)
  • Refactoring of hierarchy classes
  • New heap classes and infrastructure enhancements
  • Classes can have aliases, e.g. "l2" for euclidean distance.
  • Some error messages were made more informative.
  • Benchmarking classes, also for approximate nearest neighbor search.

Logo Gesture Recogition Toolkit 0.1 Revision 289

by ngillian - December 13, 2013, 22:59:53 CET [ Project Homepage BibTeX Download ] 4546 views, 870 downloads, 1 subscription

About: The Gesture Recognition Toolkit (GRT) is a cross-platform, open-source, c++ machine learning library that has been specifically designed for real-time gesture recognition. It features a large number of machine-learning algorithms for both classification and regression in addition to a wide range of supporting algorithms for pre-processing, feature extraction and dataset management. The GRT has been designed for real-time gesture recognition, but it can also be applied to more general machine-learning tasks.

Changes:

Added Decision Tree and Random Forests.


Logo Differential Dependency Network cabig cytoscape plugin 1.0

by cbil - October 27, 2013, 17:31:58 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1784 views, 413 downloads, 1 subscription

About: DDN learns and visualize differential dependency networks from condition-specific data.

Changes:

Initial Announcement on mloss.org.


Logo JMLR CAM Java 3.1

by wangny - October 14, 2013, 22:46:03 CET [ Project Homepage BibTeX Download ] 7083 views, 3096 downloads, 1 subscription

About: The CAM R-Java software provides a noval way to solve blind source separation problem.

Changes:

In this version, we fix the problem of not working under newest R version R-3.0.


Logo MyMediaLite 3.10

by zenog - October 8, 2013, 22:29:29 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 48992 views, 9217 downloads, 1 subscription

About: MyMediaLite is a lightweight, multi-purpose library of recommender system algorithms.

Changes:

Mostly bug fixes.

For details see: https://github.com/zenogantner/MyMediaLite/blob/master/doc/Changes


Logo NuPIC 0.1

by rhyolight - August 21, 2013, 21:01:46 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 1501 views, 627 downloads, 1 subscription

About: The mission of this project is to build and support a community interested in machine learning and machine intelligence based on modeling the neocortex and the principles upon which it works.

Changes:

Initial Announcement on mloss.org.


Logo HDDM 0.5

by Wiecki - April 24, 2013, 02:53:07 CET [ Project Homepage BibTeX Download ] 4049 views, 1011 downloads, 1 subscription

About: HDDM is a python toolbox for hierarchical Bayesian parameter estimation of the Drift Diffusion Model (via PyMC). Drift Diffusion Models are used widely in psychology and cognitive neuroscience to study decision making.

Changes:
  • New and improved HDDM model with the following changes:
    • Priors: by default model will use informative priors (see http://ski.clps.brown.edu/hddm_docs/methods.html#hierarchical-drift-diffusion-models-used-in-hddm) If you want uninformative priors, set informative=False.
    • Sampling: This model uses slice sampling which leads to faster convergence while being slower to generate an individual sample. In our experiments, burnin of 20 is often good enough.
    • Inter-trial variablity parameters are only estimated at the group level, not for individual subjects.
    • The old model has been renamed to HDDMTransformed.
    • HDDMRegression and HDDMStimCoding are also using this model.
  • HDDMRegression takes patsy model specification strings. See http://ski.clps.brown.edu/hddm_docs/howto.html#estimate-a-regression-model and http://ski.clps.brown.edu/hddm_docs/tutorial_regression_stimcoding.html#chap-tutorial-hddm-regression
  • Improved online documentation at http://ski.clps.brown.edu/hddm_docs
  • A new HDDM demo at http://ski.clps.brown.edu/hddm_docs/demo.html
  • Ratcliff's quantile optimization method for single subjects and groups using the .optimize() method
  • Maximum likelihood optimization.
  • Many bugfixes and better test coverage.
  • hddm_fit.py command line utility is depracated.

Logo MLDemos 0.5.1

by basilio - March 2, 2013, 16:06:13 CET [ Project Homepage BibTeX Download ] 20624 views, 4836 downloads, 2 subscriptions

About: MLDemos is a user-friendly visualization interface for various machine learning algorithms for classification, regression, clustering, projection, dynamical systems, reward maximisation and reinforcement learning.

Changes:

New Visualization and Dataset Features Added 3D visualization of samples and classification, regression and maximization results Added Visualization panel with individual plots, correlations, density, etc. Added Editing tools to drag/magnet data, change class, increase or decrease dimensions of the dataset Added categorical dimensions (indexed dimensions with non-numerical values) Added Dataset Editing panel to swap, delete and rename dimensions, classes or categorical values Several bug-fixes for display, import/export of data, classification performance

New Algorithms and methodologies Added Projections to pre-process data (which can then be classified/regressed/clustered), with LDA, PCA, KernelPCA, ICA, CCA Added Grid-Search panel for batch-testing ranges of values for up to two parameters at a time Added One-vs-All multi-class classification for non-multi-class algorithms Trained models can now be kept and tested on new data (training on one dataset, testing on another) Added a dataset generator panel for standard toy datasets (e.g. swissroll, checkerboard,...) Added a number of clustering, regression and classification algorithms (FLAME, DBSCAN, LOWESS, CCA, KMEANS++, GP Classification, Random Forests) Added Save/Load Model option for GMMs and SVMs Added Growing Hierarchical Self Organizing Maps (original code by Michael Dittenbach) Added Automatic Relevance Determination for SVM with RBF kernel (Thanks to Ashwini Shukla!)


Logo Neural network designer 1.1.1

by bragi - December 28, 2012, 11:38:10 CET [ Project Homepage BibTeX Download ] 4163 views, 998 downloads, 1 subscription

About: a dbms for resonating neural networks. Create and use different types of machine learning algorithms.

Changes:

AIML compatible (AIML files can be imported); new 'Grid channel' for developing board games; improved topics editor; new demo project: ALice (from AIML); lots of bug-fixes and speed improvements


Logo Divvy 1.1.1

by jlewis - November 14, 2012, 20:21:29 CET [ Project Homepage BibTeX Download ] 2142 views, 1118 downloads, 1 subscription

About: Divvy is a Mac OS X application for performing dimensionality reduction, clustering, and visualization.

Changes:

Initial Announcement on mloss.org.


Logo Pattern 2.4

by tomdesmedt - August 31, 2012, 02:26:01 CET [ Project Homepage BibTeX BibTeX for corresponding Paper Download ] 7765 views, 2128 downloads, 1 subscription

About: "Pattern" is a web mining module for Python. It bundles tools for data retrieval, text analysis, clustering and classification, and data visualization.

Changes:
  • Small bug fixes in overall + performance improvements.
  • Module pattern.web: updated to the new Bing API (Bing API has is paid service now).
  • Module pattern.en: now includes Norvig's spell checking algorithm.
  • Module pattern.de: new German tagger/chunker, courtesy of Schneider & Volk (1998) who kindly agreed to release their work in Pattern under BSD.
  • Module pattern.search: the search syntax now includes { } syntax to define match groups.
  • Module pattern.vector: fast implementation of information gain for feature selection.
  • Module pattern.graph: now includes a toy semantic network of commonsense (see examples).
  • Module canvas.js: image pixel effects & editor now supports live editing

Logo MLWizard 5.2

by remat - July 26, 2012, 15:04:14 CET [ Project Homepage BibTeX Download ] 3331 views, 846 downloads, 1 subscription

About: MLwizard recommends and optimizes classification algorithms based on meta-learning and is a software wizard fully integrated into RapidMiner but can be used as library as well.

Changes:

Faster parameter optimization using genetic algorithm with predefined start population.


Showing Items 21-40 of 44 on page 2 of 3: Previous 1 2 3 Next