See the Workshop Webpage for the context of this tutorial.
See the Workshop Hands-on Instructions for initial installation steps.
This is a repository for the Video-Based Gaze Tracking tutorial of the Bridging the Technological Gap Summer Workshop, July 31st – August 6th, 2022, German Primate Center, Göttingen, Germany.
Made by Petr Kellnhofer, 2022.
Please take these steps before the start of the tutorial. Also please pull the latest code update (git pull
) before the tutorial starts in case of last minute changes.
The tutorial has been tested on Windows 10 x64 and on GNU/Linux Ubuntu 20.04 but it should work on any computer for which python and required packages are available (e.g. MacOS). The tutorial has been designed to run on CPU alone, so CUDA (and Nvidia GPU) is not required but it can optionally be used for a better performance. Some of the tasks utilize a webcam but alternative solutions using provided sample images are also available.
The tutorial is implemented as a jupyter notebook and utilizes python 3.9 and pytorch 1.10. It is recommended to create a new python environment using the package manager miniconda or anaconda and the following steps:
- Open you terminal with
conda
and navigate to the folder containing thisREADME.md
. conda env create -f environment.yml
conda activate env_workshop
Refer to conda manual for more details if needed.
Note that the environment is the same as the one used for the classification tutorial by Pierre-Etienne, so you only need to create it once.
Download the zip file with data separately from Google Drive and unzip its content directly to the folder data
such that there are folders data/samples
, etc.
Use the Issues
tab on github to ask for help.
- Open you terminal with
conda
and navigate to the folder containing thisREADME.md
. cd notebooks
conda activate env_workshop
jupyter notebook
Follow instructions in tutorial_instructions.pptx for further instructions.
In case of issues with local environment a slightly modified version of the tutorial is available on Google Colab.
By default, GPU is disabled which prevents the CNN face detector from working. Runtime
->Change Runtime type
and select Hardware accelerator = GPU
to enable GPU support (CUDA). However, note that this introduces a limit on working time for free accounts. All other parts of the tutorial are functional without the GPU support.