One implementation of the paper "Dynamic Sliding Window for Meeting Summarization" / "Dynamic Sliding Window Modeling for Abstractive Meeting Summarization" (Interspeech 2022).
- pytorch==1.7.1
- transformers==4.8.2
- click==7.1.2
- sentencepiece==0.1.92
- allennlp==2.6.0
- allennlp-models==2.6.0
There are 4 steps to process the original AMI meeting conversations, to get the segmented samples described in the paper.
Step1_build_AMI_data_with_segments.py
Splitting the meeting conversations as well as the summaries to multiple segments/snippets, this is based on the extractive supporting annotation in AMI data;Step2_process_segmented_data.py
Conducting simple coreference resolution on the summary snippets;Step3_build_longABS_target_sentences.py
The AMI data provides long abstract, to improve the informativeness of generation, we further merge them to the summary snippets;Step4_build_AMI_all_utterances.py
Building the dialogue level sample for testing;
The segmented samples are used for training, and the dialogue level samples can be used for evaluation.
./AMI_experiments/tmp_data_with_segments/
There are train/eval/test samples after our pre-processing. For model training, using thetrain.source
(dialogue content) andtrain.target.plus_long
(reference summary);./AMI_experiments/tmp_data_dialogue_level/
For evaluation for meeting level summarization, using the filetrain.all.utter.source
(meeting level dialogue content) andtrain.doc_level_longabs
(reference summary);
@inproceedings{Liu2022DynamicSW,
title={Dynamic Sliding Window Modeling for Abstractive Meeting Summarization},
author={Zhengyuan Liu and Nancy F. Chen},
booktitle={Interspeech},
year={2022},
}
@article{liu2021dynamic,
title={Dynamic Sliding Window for Meeting Summarization},
author={Liu, Zhengyuan and Chen, Nancy F},
journal={arXiv preprint arXiv:2108.13629},
year={2021}
}