EEG Datasets

In this paper, we present a novel idea where we analyzed EEG signals to classify what type of video a person is watching which we believe is the first step of a BCI based video recommender system. For this, we setup an experiment where 13 subjects were shown three different types of videos. To be able to classify each of these videos from the EEG data of the subjects with a very good classification accuracy, we carried out experiments with several state-of-the-art algorithms for each of the submodules (pre-processing, feature extraction, feature selection and classification) of the Signal Processing module of a BCI system in order to find out what combination of algorithms best predicts what type of video a person is watching. We found, the best results (80.0% with 32.32 ms average total execution time per subject) are obtained when data of channel AF8 are used (i.e. data recorded from the electrode located at the right frontal lobe of the brain). The combination of algorithms that achieved this highest average accuracy of 80.0% are FIR Least Squares, Welch Spectrum, Principal Component Analysis and Adaboost for the submodules pre-processing, feature extraction, feature selection and classification respectively.

Dataset can be downloaded from here.

Citation Request:

Mutasim, A. K., Tipu, R. S., Bashar, M. R., & Amin, M. A. (2017, November). Video Category Classification Using Wireless EEG. In International Conference on Brain Informatics (pp. 39-48). Springer, Cham. Web Link

CLEF Datasets

We are organizing a lab session at the ImageCLEF 2015. ImageCLEF 2015 is an evaluation campaign which is being organized as part of The CLEF Initiative labs. The results of this campaign, including selected works among the participants, will be presented in the Conference and Labs of the Evaluation Forum (CLEF) 2015 which will be held in the city of Toulouse in France, 8-11 September 2015. For the 2015 edition, ImageCLEF will organize three main tasks with a global objective of benchmarking automatic annotation and indexing of images. The tasks tackle different aspects of the annotation problem and are aimed at supporting and promoting cutting-edge research addressing the key challenges in the field. A wide range of source images and annotation objectives are considered, such as general multi-domain images for object or concept detection, as well as domain-specific tasks such as labelling and separation of compound figures from biomedical literature and volumetric medical images for automated structured reporting.

The session computer vision and cybernetics group is organizing called medical clustering. Training Dataset (500 image) for the task can be download from the competition website now. Test dataset (250 image) will be released in March 2015, after the conference in September 2015 we will make all 5000 data public. Which can be downloaded from