A Novel EEG Dataset Utilizing Low-Cost, Sparse Electrode Devices for Emotion Exploration
The dataset is hosted on Zenodo and can be accessed upon request at the following link:
If you find NeuroSense valuable for your research, please cite our work:
@ARTICLE{10737340,
author={Colafiglio, Tommaso and Lombardi, Angela and Sorino, Paolo and Brattico, Elvira and Lofù, Domenico and Danese, Danilo and Sciascio, Eugenio Di and Noia, Tommaso Di and Narducci, Fedelucio},
journal={IEEE Access},
title={NeuroSense: A Novel EEG Dataset Utilizing Low-Cost, Sparse Electrode Devices for Emotion Exploration},
year={2024},
volume={},
number={},
pages={1-1},
keywords={Videos;Electroencephalography;Feature extraction;Electrodes;Color;Physiology;Brain modeling;Emotion recognition;Protocols;Data mining;Emotion Recognition;EEG Dataset;Low-Cost EEG Devices;Machine Learning;Human-Computer Interaction;Russell’s Model},
doi={10.1109/ACCESS.2024.3487932}}
}
The terms and conditions for using this dataset are specified in the LICENCE file included in this repository. Please review these terms carefully before accessing or using the data.
For additional information about the dataset, please contact:
- Name: Angela Lombardi
- Affiliation: Department of Electrical and Information Engineering, Politecnico di Bari
- Email: angela.lombardi@poliba.it
The dataset can be accessed through our dedicated web platform. To request access:
- Visit the main dataset page at: https://sisinflab.poliba.it/neurosense-dataset-request/
- Follow the instructions on the website to submit your access request
- Upon approval, you will receive further instructions for downloading the data
Please ensure you have read and agreed to the terms in the data user agreement before requesting access.
The experiment consists in 40 sessions per user. During each session, users are asked to watch a
music video with the aim to understand their emotions.
Recordings are performed with a Muse EEG headset at a 256 Hz sampling rate.
Channels are recorded as follows:
- Channel 0: AF7
- Channel 1: TP9
- Channel 2: TP10
- Channel 3: AF8
The chosen songs have various Last.fm tags in order to create different feelings. The title of every track can be found in the "TaskName" field of sub-ID***_ses-S***_task-Default_run-001_eeg.json, while the author, the Last.fm tag and additional information in "TaskDescription".
The subject pool is made of 30 college students, aged between 18 and 35. 16 of them are males, 14 females.
The experiment was performed using the same procedures as those to create Deap Dataset, which is a dataset to recognize emotions via a Brain Computer Interface (BCI).
Firstly, music videos were selected. Once 40 songs were picked, the protocol was chosen and the self-assessment questionnaire was created.
In order to evaluate the stimulus, Russell's VAD (Valence-Arousal-Dominance) scale was used.
In this scale, valenza-arousal space can be divided in four quadrants:
- Low Arousal/Low Valence (LALV);
- Low Arousal/High Valence (LAHV);
- High Arousal/Low Valence (HALV);
- High Arousal/High Valence (HAHV).
The experiment was performed in a laboratory located at DEI Department of Politecnico di Bari.
Data recorded during session S019 - Session 2, ID021 - Session 23, user was corrupted, therefore is missing.
Sessions S033 and S038 of ID015 user show a calculated effective sampling rate lower than 256 Hz:
- ID015_ses-S033 has 226.1320 Hz
- ID015_ses-S038 has 216.9549 Hz