Bellet et al., Human-level saccade and microsaccade detection with deep neural networks, Journal of Neurophysiology 2019
- web service available
- kernel size of convolution and max pooling operations now definable by the user (thus longer or shorter time windows will be seen by U'n'Eye)
Train your own eye movement detection network in 10 minutes and label your data. Get started now
Try out our web service.
U'n'Eye is a Python 3 package, that uses PyTorch for the neural network implementation.
For a description of the algorithm, see our preprint. For any questions regarding this repository please contact [email protected] or [email protected].
We provide network weights that were learned on different datasets, described in the paper. The weights can be found in the folder training and the corresponding datasets will be available in the folder data after publication. For instructions on how to use pretrained networks, please see below.
Users can train their own network to obtain optimal performance. Please see the module description below and the example jupyter notebook UnEye.ipynb for instructions.
We provide a docker container for platform-independent use that is recommended for Windows users.
Installation: local or with docker
In the following, all commands written in boxes need to be entered into your terminal.
Local: install the python package (Mac & Linux)
1) Via the terminal, check if you have python3 installed
python3 --version
If not found, download and install python3 here.
2) Clone the GitHub repository into your local directory and make sure you are using pip3 (see below for possible error message on Mac).
git clone https://github.com/berenslab/uneye
alias pip=pip3
3) Now the last step: go into the new uneye folder in your GitHub folder and install the package. In this example on a Mac, the GitHub folder is located in the Documents folder of the user. Find out the absolute path of that folder and type the following:
cd /Users/yourusername/Documents/GitHub/uneye
Then, use one of the following commands for installation, depending on your platform:
For Mac:
pip install ./ -r ./requirements_mac.txt
For Linux:
pip install ./ -r ./requirements_lin.txt
Note: If the git command does not work under Mac OS, first run
xcode-select --install
and install the necessary tools.
Docker: the platform independent solution
1) Install docker:
2) Download the docker image that contains U'n'Eye, pytorch and all other python packages you need. This step will take some time.
docker pull mebellet/uneye:v-0.2
General:
model = uneye.DNN(sampfreq= sampling_frequency )
model.train(X,Y,Labels)
model.test(X,Y,Labels)
model.predict(X,Y)
model.crossvalidate
Generally, one first calls the network and can then apply different methods, as described in the module description below.
An example jupyter notebook is provided in this repository (UnEye.ipynb). Clone the repository as described above (or alternatively download the files). Put the file UnEye.ipynb into the folder where you want to work. Depending on whether you use U'n'Eye with the docker container or locally, enter the following into your command window to use the jupyter notebook (replace YourWorkingDirectory)
cd /YourWorkingDirectory
docker run -it --rm -p 6688:8888 --name uneye -v $(pwd)/.:/home/jovyan mebellet/uneye:v-0.2
!! Important: Now, don't just copy the output of the into your browser, but do the following:
Open your web browser. Enter "localhost:6688" as URL. Then you will be asked for a token. This token appeared when you entered the command above. It looks something like this: http://6688:8888/?token=775c758f58cdc82bf6ddf51a112228f4dd4229c5b3847bb1 . Copy the token (in this case 775c758f58cdc82bf6ddf51a112228f4dd4229c5b3847bb1) and log in. Now you should see the content of your folder in the web browser.
alias python=python3
cd /YourWorkingDirectory
jupyter notebook
To stop jupyter notebook, press Ctrl + C .
With the .py file UnEye.py you can use the package from the command line. So far it takes the following input arguments:
Input arguments (=necessary): mode: train or predict
x*: filename of the horizontal eye position (.csv or .mat file)
y*: filename of the vertical eye position (.csv or .mat file)
labels(*for training): filename of the eye movement ground truth labels (.csv or .mat file)
sampfreq*: sampling frequency of the data (Hz)
weights: ouput/input filename of trained weights
first run the following, depending on whether you use the Docker container or work locally. Note: /YourWorkingDirectory must contain the .py file UnEye.py from this repo.
cd /YourWorkingDirectory
docker run -it -p 6688:8888 --name uneye -v $(PWD)/:/home/jovyan mebellet/uneye:v-0.2 /bin/bash
cd /YourWorkingDirectory
alias python=python3
Now, stay in the command window. You can from here either train a new network or predict eye movements from new data:
python UnEye.py -m train -x data/x_name -y data/y_name -l data/labels_name -f sampfreq
Note: In this example the files are located in the directory /YourWorkingDirectory/data
The trained weights will be saves to training/weights or to training/weightsname if the argument -w weightsname is given.
python UnEye.py -m predict -x data/x_name -y data/y_name -f sampfreq
Note: This will automatically use the weights saved under training/weights unless you specify your weightsname by giving the input argument -w training/weightsname .
The predicted saccade probability and the binary prediction are saved to data/Sacc_prob and data/Sacc_pred respectively.
If you use docker, exit after usage with:
exit
The uneye module contains the DNN class
model = uneye.DNN(max_iter=500, sampfreq=1000,
lr=0.001, weights_name='weights',
min_sacc_dist=1,min_sacc_dur=6,augmentation=True,
ks=5,mp=5,inf_correction=1.5,val_samples=30)
Arguments:
max_iter: maximum number of epochs during training
sampfreq: sampling frequency of the eye tracker (Hz)
lr: learning rate of the network training
weights_name: input/output filename for trained network weights
min*_*sacc_dist: minimum distance between two saccades in ms for merging of saccades
min*_*sacc_dur: minimum saccade duration in ms for removal of small events
augmentation: whether or not to use data augmentation for training, default: True
ks: kernel size of convolution operations, default=5
mp: size of max pooling operation, default=5
inf_correction: float, value to replace Infs occuring after differential of input signal
val_samples: int, number of validation samples (for early stopping criterion)
Train the network weights with your own training data. It is recommended to label at least 500 seconds of data. Arrange the data into a matrix of samples*timebins or input them as a vector of length timebins . In case you arrange the data into a matrix, the time length of each sample should be at least 1000 ms.
During training, the current iteration and maximum number of iterations will be shown. Note that training stops earlier most of the times.
model.train(X,Y,Labels)
X,Y : horizontal and vertical eye positions, array-like, shape: {(nsamples, nbins) or (nbins)}
Labels: eye movement ground truth labels with k different values for k different classes (e.g. fixation=0, saccade=1, saccades should always be labelled with 1), array-like, shape: {(nsamples, nbins) or (nbins)}
Output: the trained weights are saved into the folder training.
Test performance of the network after training on your own test data. Arrange the data into a matrix of samples*timebins or input them as a vector of length timebins .
Prediction, Probability, Performance = model.test(X,Y,Labels)
Input parameters:
X,Y : horizontal and vertical eye positions, {array-like, shape: {(nsamples, nbins) or (nbins)}
Labels: eye movement ground truth labels, array-like, shape: {(nsamples, nbins) or (nbins)}
Output:
Prediction: eye movement class prediction for each time bin, same shape as input
Probability: softmax probability output of network, shape: {n_samples,classes,time) or (classes,time)}
Performance: numpy dictionary containing different performance measures: Cohen's Kappa, F1, onset distance, offset distance. To call these measures, use
cohens_kappa = Performance['kappa']
f1 = Performance['f1']
on_dist = Performance['on']
off_dist = Performance['off']
Predict saccades in recordings with pretrained weights. Arrange the data into a matrix of samples*timebins or input them as a vector of length timebins .
Prediction, Probability = model.predict(X,Y)
Input parameters:
X,Y : horizontal and vertical eye positions, {array-like, shape: {(nsamples, nbins) or (nbins)}
Output:
Prediction: eye movement class prediction for each time bin, same shape as input
Probability: softmax probability output of network, shape: {n_samples,classes,time) or (classes,time)}
Implementation of K-fold cross-validation.
crossvalidate(self,X,Y,Labels,X_val,Y_val,Labels_val,Labels_test=None,K=10)
Input parameters:
X,Y : horizontal and vertical eye positions, {array-like, shape: {(nsamples, nbins) or (nbins)}
Labels: array-like, shape: {(n_timepoints),(n_samples, n_timepoints)}, class labels in range [0 classes-1], fixation=0, saccades=1
X_val,Y_val: array-like, shape: {(n_timepoints),(n_samples, n_timepoints)}, additional horizontal and vertical eye positions in degree for early stopping criterion. Can be small.
Labels_val: array-like, shape: {(n_timepoints),(n_samples, n_timepoints)}, additional class labels in range [0 classes-1], fixation=0, saccades=1 for early stopping criterion. Can be small.
Labels_test: array-like, shape: {(n_timepoints),(n_samples, n_timepoints)}, if test Labels different from training labels (for training with missing labels only), optional
K: float, number of folds of cross validation
Adding other python modules to the docker container
You can add other python modules to the docker container once you pulled the image (as described above). For this, run:
docker run -it -p 6688:8888 --name uneye mebellet/uneye:v-0.2 /bin/bash
pip install module_name
This installs the python module 'module_name'. Exit the container by pressing Control+p and Control+q, then commit the changes to the image:
docker commit uneye mebellet/uneye:v-0.2