This library was originally implemented for the DIKU course Statistical Methods for Machine Learning. It contains the following algorithms:

Classification

  • K nearest neighbours
  • Linear discriminant analysis
  • Multilayer perceptron (with backpropagation)
  • Naive bayes (Also supporting different kernel estimator, like Epanechnikov and Gauassian kernels)

Preprocessing

  • Data normalization (zero mean, unit variance)
  • Data rescaling (data rescaled to lie between zero and one)

Regression

  • Linear regression
  • Linear regression with basis functions

Testing

  • k-fold cross validation
  • Test set validation 

 

The whole Java source code is available at github. Example usage for most classifiers is available here, and here for the multilayer perceptron.

Note that this library relies on JAMA for matrix operations.

 

 

Future plans 

  • Examples (and other documentation)
  • K-means clustering
  • Principal component analysis

 

Changelog

--- 2014/07/01: v0.1

Initial upload

Attachments:
Download this file (mllib.jar)mllib.jar[Java machine learning library]80 kB
This is the source code for an implementation of a multilayer perceptron neural network in Java, using backpropagation for learning.

Main method included for solving XOR-problem to show example of use.
 
 
This algorithm has been enhanced and included in my more general ML library.
Attachments:
Access this URL (http://thiele.nu/attachments/article/4/MultiLayerPerceptron.java)MultiLayerPerceptron.java[ ]7 kB

This is a Java implementation of the AI algorithm k-nearest neighbours (KNN) used for classification.

Constructor: NearestNeighbour(ArrayList<DataEntry> dataset, int k).


Main method included to give an example of use.

 

This algorithm has been enhanced and included in my more general ML library.

Attachments:
Access this URL (http://thiele.nu/attachments/article/3/NearestNeighbour.java)NearestNeighbour.java[ ]3 kB