Ethics & Privacy Statement: In compliance with the informed consent signed by the elderly participants and IRB regulations, raw data will NOT be publicly released to protect participant privacy. We provide pre-processed and aligned multimodal features for use in all downstream tasks.
Abstract
While affective computing has advanced considerably, multimodal emotion prediction in aging populations remains underexplored, largely due to the scarcity of dedicated datasets. Existing multimodal benchmarks predominantly target young, cognitively healthy subjects, neglecting the influence of cognitive decline on emotional expression and physiological responses. To bridge this gap, we present MECO, a Multimodal dataset for Emotion and Cognitive understanding in Older adults. MECO includes 42 participants and provides approximately 38 hours of multimodal signals, yielding 30,592 synchronized samples. To maximize ecological validity, data collection followed standardized protocols within community-based settings. The modalities cover video, audio, electroencephalography (EEG), and electrocardiography (ECG). In addition, the dataset offers comprehensive annotations of emotional and cognitive states, including self-assessed valence, arousal, six basic emotions, and Mini-Mental State Examination cognitive scores. We further establish baseline benchmarks for both emotion and cognitive prediction. MECO serves as a foundational resource for multimodal modeling of affect and cognition in aging populations, facilitating downstream applications such as personalized emotion recognition and early detection of mild cognitive impairment (MCI) in real-world settings.
42
Elderly Participants
38h
Total Duration
30,592
Sync Samples
2
Label Types
Emotion & Cognition
7
Target Tasks
5 Emotion + 2 Cognition
MODALITIES
Video
Non-intrusiveAudio
High-fidelityEEG
2-Channel WearableECG
1-Channel SensorBASELINE TASKS
MECO supports 7 distinct downstream baseline tasks spanning emotion and cognition.
Emotion Domain (5 Tasks)
Objective → Subjective Coarse-grained → Fine-grainedNote: Task T1 evaluates objective stimulus-induced emotion. Tasks T2–T5 evaluate subjective emotional experiences assessed via the Self-Assessment Manikin (SAM), transitioning from coarse-grained sentiment classes to fine-grained emotion categories and continuous dimensional spaces.
Task T1 (SR): Stimulus-induced Emotion Recognition
- Target: Predict objective stimulus-induced emotions based on intended stimuli.
- Output: Discrete labels
{0: 'angry', 1: 'boredom', 2: 'happy', 3: 'neutral', 4: 'sadness', 5: 'tension'}. - Metric: Unweighted Average Recall (UAR), Weighted Average Recall (WAR).
Task T2 (SA): 3-class Sentiment Analysis
- Target: Predict coarse-grained subjective sentiment.
- Output: Discrete labels
{0: 'Negative', 1: 'Neutral', 2: 'Positive'}. - Metric: UAR, WAR.
Task T3 (ER): 5-class Emotion Recognition
- Target: Predict fine-grained subjective emotion categories.
- Output: Discrete labels
{0: 'Angry', 1: 'Sadness', 2: 'Tension', 3: 'Neutral', 4: 'Happiness'}. - Metric: UAR, WAR.
Task T4 (VR): Valence Regression
- Target: Predict continuous subjective valence dimensional scores.
- Output: Continuous values in
[1, 9](spanning from negative to positive). - Metric: Mean Absolute Error (MAE), Concordance Correlation Coefficient (CCC).
Task T5 (AR): Arousal Regression
- Target: Predict continuous subjective arousal dimensional scores.
- Output: Continuous values in
[1, 9](spanning from calm to excited). - Metric: MAE, CCC.
Cognition Domain (2 Tasks)
Coarse Screening → Fine-grained ScoringNote: Cognition tasks progress from coarse-grained binary screening for cognitive impairment (T6) to fine-grained regression of specific cognitive scores (T7).
Task T6 (CR): Binary Cognitive Impairment Screening
- Target: Predict coarse-grained cognitive state for Mild Cognitive Impairment (MCI) screening.
- Output: Binary labels
{0: 'HC', 1: 'MCI'}. - Metric: Accuracy (ACC), Macro-F1.
Task T7 (MR): MMSE Score Regression
- Target: Predict fine-grained Mini-Mental State Examination (MMSE) scores.
- Output: Values in
[0, 30](evaluated as a continuous regression task, though underlying target values are inherently discrete integers). - Metric: MAE, CCC.
EVALUATION PROTOCOLS
Personalization
Subject-Dependent
EVALUATION SPECTRUM COVERAGE
Generalization
Subject-Independent
Subject-Dependent (SD)
Personalized Modeling • Emotion Tasks Only
Designed for personalized applications (e.g., in-home elderly care systems). To preserve the temporal dynamics of elicited emotional responses, a Chronological Split is applied within each trial.
Chronological Split per Trial:
Subject-Independent (SI)
Generalization • Emotion & Cognition Tasks
Designed to evaluate model generalization across unseen participants. This is strictly required for cognition screening (MCI vs. HC) to ensure valid real-world applicability.
Subject-wise 5-Fold Cross-Validation:
Train (~80%)
Test (~20%)
POTENTIAL APPLICATIONS
Unlocking new frontiers in scientific discovery and clinical deployment.
DATA ACCESS
Get started with the MECO dataset for your research.
Step 1
Download Encrypted Features
The extracted multimodal feature files are hosted on Baidu Netdisk in an encrypted archive. Please download the files first.
Baidu Netdisk (Pan)
Extraction Code:
b3yj
Step 2
Request Decryption Key
To obtain the password to unzip the dataset, please send a formal request from your institutional email to hongbinchen@stu.njmu.edu.cn. Please use the following template:
Subject: Data Access Request for MECO Dataset
Dear MECO Dataset Team,
I am writing to request the decryption password for the MECO dataset. I agree to use the data strictly for academic research purposes and will not distribute it to any third parties without permission.
- Full Name: [Your Name]
- Institution / Affiliation: [Your University / Research Center]
- Position / Title: [e.g., Master Student, PhD Student, Professor]
- Intended Purpose of Use: [Briefly describe your research project, e.g., multimodal emotion recognition]
I confirm that I have read the ethics and privacy statement, and I commit to citing your paper if this dataset is used in my publications.
Best regards,
[Your Name]
* Applications will typically be reviewed and processed within 1-2 business days.
CITATION
If you find the MECO useful for your research, please cite our work:
@article{chen_meco2026,
title = {MECO: A Multimodal Dataset for Emotion and Cognitive Understanding in Older Adults},
author = {Chen, Hongbin and Li, Jie and Wang, Wei and Song, Siyang and Gu, Xiao and Li, Jianqing and Xiang, Wentao},
journal = {arXiv preprint arXiv:2604.03050},
doi = {10.48550/ARXIV.2604.03050},
url = {https://arxiv.org/abs/2604.03050},
year = {2026}
}