Datasets:

Modalities:
Text
Languages:
English
License:
ffwang commited on
Commit
cfbfd2f
·
verified ·
1 Parent(s): ed35a05

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -12,7 +12,7 @@ For the task named Multimodal Emotion-Cause Pair Extraction in Conversation, we
12
 
13
  For more details, please refer to our GitHub:
14
 
15
- - [Multimodal Emotion-Cause Pair Extraction in Conversations](https://github.com/NUSTM/MECPE)
16
  - [SemEval-2024 Task 3](https://github.com/NUSTM/SemEval-2024_ECAC)
17
 
18
  ## Dataset Statistics
@@ -24,6 +24,23 @@ For more details, please refer to our GitHub:
24
  | Emotion (utterances) | 5,577 | 668 | 1,445 | 7,690 |
25
  | Emotion-cause (utterance) pairs | 7,055 | 866 | 1,873 | 9,794 |
26
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  ## Citation
28
 
29
  If you find ECF useful for your research, please cite our paper using the following BibTeX entries:
 
12
 
13
  For more details, please refer to our GitHub:
14
 
15
+ - [Multimodal Emotion-Cause Pair Extraction in Conversations](https://github.com/NUSTM/MECPE/tree/main/data)
16
  - [SemEval-2024 Task 3](https://github.com/NUSTM/SemEval-2024_ECAC)
17
 
18
  ## Dataset Statistics
 
24
  | Emotion (utterances) | 5,577 | 668 | 1,445 | 7,690 |
25
  | Emotion-cause (utterance) pairs | 7,055 | 866 | 1,873 | 9,794 |
26
 
27
+ ## About Multimodal Data   
28
+
29
+ ⚠️ Due to potential copyright issues with the TV show "Friends", we do not provide pre-segmented video clips.
30
+
31
+ If you need to utilize multimodal data, you may consider the following options:
32
+
33
+ 1. Use the acoustic and visual features we provide:
34
+ - [`audio_embedding_6373.npy`](https://drive.google.com/file/d/1EhU2jFSr_Vi67Wdu1ARJozrTJtgiQrQI/view?usp=share_link): the embedding table composed of the 6373-dimensional acoustic features of each utterances extracted with openSMILE
35
+ - [`video_embedding_4096.npy`](https://drive.google.com/file/d/1NGSsiQYDTqgen_g9qndSuha29JA60x14/view?usp=share_link): the embedding table composed of the 4096-dimensional visual features of each utterances extracted with 3D-CNN
36
+
37
+ 2. Since ECF is constructed based on the MELD dataset, you can download the raw video clips from [MELD](https://github.com/declare-lab/MELD).
38
+ Most utterances in ECF align with MELD. However, **we have made certain modifications to MELD's raw data while constructing ECF, including but not limited to editing utterance text, adjusting timestamps, and adding or removing utterances**. Therefore, some timestamps provided in ECF have been corrected, and there are also new utterances that cannot be found in MELD. Given this, we recommend option (3) if feasible.
39
+
40
+ 3. Download the raw videos of _Friends_ from the website, and use the FFmpeg toolkit to extract audio-visual clips of each utterance based on the timestamps we provide.
41
+
42
+
43
+
44
  ## Citation
45
 
46
  If you find ECF useful for your research, please cite our paper using the following BibTeX entries: