Introduction

The Donders Machine Learning Toolbox (DML) is an extensible machine learning toolbox written in Matlab, tailored to the analysis of neural data. The DML package contains high-level functionality as well as implemented multivariate methods.

Contents

Input and output

Input data takes the form of an N x K or N x K x T matrix with N the number of trials, K the number of variables and T the number of time samples. The latter representation is only used in case of time series analysis methods. Output data takes the form of N x M matrix. Missing data in the input or output is indicated using NaN values. Multiple outputs where Y is an N x K matrix are handled by specialized methods (see examples section). In case of classification and regression problems, the output data is assumed to be an N x 1 vector. For classification, this vector contains class labels 1,2,..., e.g.

Y = [1 1 1 1 2 2 2 2]';

For regression this vector contains real values, e.g.

Y = [0.12 -1.23 2.4 1.1 -0.03]';

For now lets focus on a classification problem and generate some artificial training data (X1,Y1) and test data (X2,Y2):

rand('seed',1); % initialize random number generator
X1 = rand(50,100); X1(26:end,1:50) = 0.15+rand(25,50);
X2 = rand(50,100); X2(26:end,1:50) = 0.15+rand(25,50);
Y1 = [ones(25,1); 2*ones(25,1)];
Y2 = [ones(25,1); 2*ones(25,1)];

Methods

A method is a class which contains a train function and a test function. All methods inherit from the abstract dml.method class. A method's train function can either learn a transformation of X itself in which case Y is left unspecified (as in supervised learning) or learn a transformation from X to Y, as in supervised learning. A method's test function takes input data X and transforms this data to an output Y. For instance,

m = dml.standardizer;
m = m.train(X1);
Z = m.test(X1);

gives a standardized representation Z of X with mean zero and unit standard deviation. In contrast,

m = dml.naive;
m = m.train(X1,Y1);
Z = m.test(X2);

learns a Gaussian naive Bayes model and tests it on new data. Note that the output is always of the size of the input with the exception of classification methods. In that case, the returned output is a discrete probability distribution over class labels:

Z(1:5,:)
ans =

    0.9095    0.0905
    1.0000    0.0000
    1.0000    0.0000
    0.9974    0.0026
    0.9913    0.0087

While we may call methods directly, it is often convenient to encapsulate them in a multivariate analysis object. This object just provides storage for a sequence of multivariate analysis methods. For instance,

a = dml.analysis({dml.standardizer dml.naive});
a = a.train(X1,Y1);
Z = a.test(X2);

creates a multivariate analysis object, trains a sequence of methods and the applies the trained object to data. In order to test the performance, we can for example compute the classification accuracy

dml.statistic('accuracy',Y2,Z)
ans =

    0.9000

or use a binomial test for significance

dml.statistic('binomial',Y2,Z)
ans =

   2.1049e-09

Low p values indicate that the outcome is significantly different from random classification. For a more stringent significance test, consult the permutation test section. For other statistics please consult dml.statistic.

Instead of examining decoding performance, one may want to examine the parameters of the decoding method in order to be able to identify which features play an important role in the decoding. For this, we can use the model function which is implemented by most methods. This function returns a struct with various diagnostic parameters.

m = a.model;

The meaning of these parameters depends on the used method and can be consulted by invoking the help:

help dml.naive.model;
  returns
 
  m.divergence summed KL divergences between all pairs of class distributions

Here, the field m.divergence contains an nvariables x 1 vector of KL divergences between the estimated Gaussians for each class. Hence, it gives a measure of separation. Let's plot them

bar(m.divergence);

The divergences indeed indicate that the first 50 features in our data separate the classes, as is expected from the way we generated the data. Of course there are better ways to determine parameter relevance. Here we showed an example of using naive Bayes which is just one of the implemented methods. Please consult the function reference and examples for additional documentation on the implemented methods.

Multiple datasets

Sometimes we deal with multiple datasets at the same time. These are trained in the same way by using cell-arrays of data:

m = dml.naive;
m = m.train({X1 X1},{Y1 Y1});
Z = m.test({X2 X2});

This is only of use when we want to process multiple datasets in parallel or combine the data as in multitask learning. This calling convention presupposes that a method knows how to deal with multiple datasets. This is facilitated by using the dml.ndata class (see dml.naive for an example of its use).

Time series analysis

Methods who use input data of the form N x K assume that all data within a trial is acquired at one static moment in time. If input data is of the form N x K x T then we are dealing with time series data (series of different length can be modeled using NaNs). As an example of the use of a time series analysis method we will perform parameter estimation in a linear dynamical system where our observations X are noisy versions of the state variable Y:

N = 2; K = 10; T = 1000;
ncycles = 10;
Y = sin(ncycles * 2 * pi * (1:T) ./ T);
Y = repmat(reshape(Y,[1 1 numel(Y)]),[N 1 1]);
X = repmat(Y,[1 K 1]) + 0.5*randn(N,K,T);
m = dml.lds('verbose',true,'maxiter',50);
m = m.train(X,nan(size(Y))); % hidden state estimation
U = repmat(Y,[1 K 1]) + 0.5*randn(N,K,T);
Z = m.test(U);
figure
plot(squeeze(U(1,1,:))','b:');
hold on;
plot(squeeze(Y(1,:,:))','k');
plot(squeeze(Z(1,:,:))','r');
legend('noisy observation','real state','predicted state');
EM step: 0; log likelihood: -7.49181e+06
EM step: 1; log likelihood: -17665.3
EM step: 2; log likelihood: -17665.3
EM step: 3; log likelihood: -17665.3
EM step: 4; log likelihood: -17665.3
EM step: 5; log likelihood: -17665.3
EM step: 6; log likelihood: -17665.3
EM step: 7; log likelihood: -17665.2
EM step: 8; log likelihood: -17665.2
EM step: 9; log likelihood: -17665.2
EM step: 10; log likelihood: -17665.2
EM step: 11; log likelihood: -17665.2
EM step: 12; log likelihood: -17665.2
EM step: 13; log likelihood: -17665.2
EM step: 14; log likelihood: -17665.2
EM step: 15; log likelihood: -17665.2
EM step: 16; log likelihood: -17665.2
EM step: 17; log likelihood: -17665.2
EM step: 18; log likelihood: -17665.2
EM step: 19; log likelihood: -17665.2
EM step: 20; log likelihood: -17665.2
EM step: 21; log likelihood: -17665.2
EM step: 22; log likelihood: -17665.2
EM step: 23; log likelihood: -17665.2
EM step: 24; log likelihood: -17665.1
EM step: 25; log likelihood: -17665.1
EM step: 26; log likelihood: -17665.1
EM step: 27; log likelihood: -17665
EM step: 28; log likelihood: -17665
EM step: 29; log likelihood: -17664.9
EM step: 30; log likelihood: -17664.8
EM step: 31; log likelihood: -17664.6
EM step: 32; log likelihood: -17664.4
EM step: 33; log likelihood: -17664.1
EM step: 34; log likelihood: -17663.5
EM step: 35; log likelihood: -17662.5
EM step: 36; log likelihood: -17660.7
EM step: 37; log likelihood: -17656.9
EM step: 38; log likelihood: -17647.8
EM step: 39; log likelihood: -17621.1
EM step: 40; log likelihood: -17524.7
EM step: 41; log likelihood: -17179.9
EM step: 42; log likelihood: -16541.1
EM step: 43; log likelihood: -15947
EM step: 44; log likelihood: -15535.7
EM step: 45; log likelihood: -15299.1
EM step: 46; log likelihood: -15189.2
EM step: 47; log likelihood: -15146.9
EM step: 48; log likelihood: -15131.9
EM step: 49; log likelihood: -15126.2
EM step: 50; log likelihood: -15126.2

Note that we use nans to represent the state variable Y to indicate that it is unobserved. The dynamics of the predicted state is really close to the real state (necessarily sign and amplitude are not preserved).

Crossvalidation

In the previous, we use two separate datasets X1 and X2. Often, in an analysis one would want to perform crossvalidation on one dataset X. That is, the dataset is split up and each data subset is either used for training or for testing. This is facilitated by the dml.crossvalidator object. It takes a multivariate analysis object and a specification of the data subsets and automates the procedure. The stats field is then used to evaluate the results. For instance,

X = rand(50,100); X(26:end,1:50) = 0.25+rand(25,50);
Y = [ones(25,1); 2*ones(25,1)];
m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'stat','accuracy');
m = m.train(X,Y);
m.statistic
ans =

     1

by default performs a five-fold crossvalidation. The crossvalidator is trained and the statistics show that a majority of the trials was correctly classified. The stat field can also be given as an argument to the statistic function as in:

m.statistic('accuracy')
ans =

     1

Default crossvalidation behaviour can be overridden using the 'type', 'folds','proportion' and 'resample' fields. A few more examples:

m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'type','nfold','folds',10);

performs ten-fold crossvalidation.

m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'type','split','proportion',0.9);

uses 90% of the data for training and the remainder for testing

m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'resample',true);

will balance non-balanced data using resampling; it upsamples data in the training folds and downsamples data in the test folds.

One may also manually specify which samples belong to each fold using the 'trainfolds' and or 'testfolds' fields. This overrides the previous behaviours. For instance

m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'trainfolds',{1:2:50 2:2:50});

creates a crossvalidator whose training folds consist of uneven and even trials (the complement is taken as test data).

A crossvalidator also accepts multiple datasets. It will then generate results for each dataset. For instance:

m = dml.crossvalidator('mva',{dml.standardizer dml.naive},'trainfolds',{{1:2:50 2:2:50} {1:2:50 2:2:50}});
m = m.train({X X},{Y Y});
m.statistic('accuracy')
ans =

    0.9600
    0.9600

Permutation testing

In order to test whether the crossvalidation results are significant, we can use permutation testing, where we compare the results under random permutations of the output Y with the actual results:

m = dml.permutation('stat','accuracy','validator',dml.crossvalidator('mva',{dml.standardizer dml.naive}),'nperm',20,'verbose',true);
m = m.train(X,Y);
p = m.statistic
testing permutation 1 of 20: 0.52
testing permutation 2 of 20: 0.44
testing permutation 3 of 20: 0.50
testing permutation 4 of 20: 0.58
testing permutation 5 of 20: 0.52
testing permutation 6 of 20: 0.48
testing permutation 7 of 20: 0.46
testing permutation 8 of 20: 0.50
testing permutation 9 of 20: 0.42
testing permutation 10 of 20: 0.46
testing permutation 11 of 20: 0.52
testing permutation 12 of 20: 0.54
testing permutation 13 of 20: 0.66
testing permutation 14 of 20: 0.48
testing permutation 15 of 20: 0.44
testing permutation 16 of 20: 0.50
testing permutation 17 of 20: 0.46
testing permutation 18 of 20: 0.46
testing permutation 19 of 20: 0.52
testing permutation 20 of 20: 0.54
actual outcome: 1.00

p =

     0

This means that the null-hypothesis that the actual outcome is not significantly different from the permutated outcomes can be rejected. Note that, in practice, we need many more permutations to get a reliable result. The verbose option, which is available for other objects as well, gives diagnostic output.

Bootstrap testing

An important question is how stable the parameters are we estimate for our models. This question can be answered using bootstrap testing. Here, we resample data with replacement and verify how stable the parameter estimates are. For example:

m = dml.bootstrap('mva',{dml.standardizer dml.naive},'nboot',200);
m = m.train(X,Y);
[mu,se] = m.statistic;

gives the mean and standard error for all parameters which are returned by a method's model function. For instance, for naive Bayes, we can plot the divergences:

figure;
errorbar(mu.divergence,se.divergence,'ko');

Grid search

Methods may require the optimization of particular parameters. This can be achieved using the dml.gridsearch object. The gridsearch object takes a crossvalidator as argument, the parameters which needs to be optimized and the values these parameters may take on. The gridsearch can exploit warm starts; that is, for efficiency reasons, the employed multivariate analysis may start at the previous parameter settings when we test multiple values. This requires the 'restart' parameter in the used multivariate method to be set to false. Note that a gridsearch method can be used within a multivariate analysis pipeline. That is, it behaves as any other method.

In this example we show the use of the gridsearch together with the support vector machine:

m = dml.gridsearch('validator',dml.crossvalidator('type','split','stat','accuracy','mva',dml.svm('restart',false)),'vars','C','vals',fliplr(logspace(-4,1,5)),'verbose',true);
m = m.train(X,Y);
optimizing variable(s) dml.svm.C
evaluating configuration 1 of 5: dml.svm.C=10.000000
 : 1.000000
evaluating configuration 2 of 5: dml.svm.C=0.562341
 : 1.000000
evaluating configuration 3 of 5: dml.svm.C=0.031623
 : 1.000000
evaluating configuration 4 of 5: dml.svm.C=0.001778
 : 1.000000
evaluating configuration 5 of 5: dml.svm.C=0.000100
 : 1.000000
retraining optimum configuration dml.svm.C=10.000000 using all data

Customization

It is easy to write your own wrapper to a new method or even to another toolbox. This new method should be a Matlab class which inherits from the dml.method class and is required to implement the train and test functions. Optionally, it may implement the model function which returns the method's parameters in a suitable form. Comments should follow help coding conventions as shown in the following template:

classdef mymethod < dml.method
% MYMETHOD short description (max 50 characters) ending with a period.
%
% DESCRIPTION
% full description
%
% EXAMPLE (recommended)
%
% NOTE (optional)
%
% REFERENCE (optional)
%
% When using this method please refer to the following:
%
% Great paper (2012)
%
% DEVELOPER
% Marcel van Gerven (m.vangerven@donders.ru.nl)
 properties
  myproperty % description of property
 end
 methods
   function obj = mymethod(varargin)
     obj = obj@dml.method(varargin{:});
   end
   function obj = train(obj,X,Y)
     % return object trained on input data X and (optionally) output data Y
   end
   function Y = test(obj,X)
     % test the trained object on input data X and return output Y
   end
   function m = model(obj)
   % MODEL description of what this model returns
     % optional: returns parameters of this method
   end
 end

end