Skip to content

Latest commit

 

History

History
45 lines (34 loc) · 1.46 KB

README.md

File metadata and controls

45 lines (34 loc) · 1.46 KB

Overview

This project contains the code for the paper Multimodal Emotion-Cause Pair Extraction with Holistic Interaction and Label Constraint published in ACM Transactions on Multimedia (ToMM).

Dependencies

This project is based on PyTorch and Transformers. You can create the conda environment using the following command:

conda env create -f environment.yml

Then activate the environment:

conda activate hilo 

Configuration

The configurations for the model and training process are stored in src/config.yaml. You can modify this file to adjust the settings.

Data

The dataset is located in data/dataset. Please follow the instructions in data/dataset/README.md to download the audio and video features, and place them in the data/dataset directory.

Usage

You can run the following command to train & evaluate the model:
python main.py

Citation

If you use this code in your research, please cite our paper:

@article{li23mecpe,
    author = {Li, Bobo and Fei, Hao and Li, Fei and Chua, Tat-seng and Ji, Donghong},
    title = {Multimodal Emotion-Cause Pair Extraction with Holistic Interaction and Label Constraint},
    year = {2024},
    publisher = {Association for Computing Machinery},
    address = {New York, NY, USA},
    issn = {1551-6857},
    url = {https://doi.org/10.1145/3689646},
    doi = {10.1145/3689646},
    journal = {ACM Trans. Multimedia Comput. Commun. Appl.},
}