Document-level event argument extraction with self-augmentation and a cross-domain joint training mechanism

作者:

Highlights:

• A new approach for resolving the data sparsity issue in document-level event argument extraction from two complementing viewpoints.

• A self-augmentation technique that combines pre-trained language models and a label-conditioned pre-training procedure to retain word-label consistency.

• A cross-domain joint training framework for knowledge transfer from datasets of varied granularity, task specifications, and the event schema description language.

• A novel noise filtering method in teacher–student framework to mitigate the data quality issue.

• State-of-the-art performance on standard datasets.

摘要

•A new approach for resolving the data sparsity issue in document-level event argument extraction from two complementing viewpoints.•A self-augmentation technique that combines pre-trained language models and a label-conditioned pre-training procedure to retain word-label consistency.•A cross-domain joint training framework for knowledge transfer from datasets of varied granularity, task specifications, and the event schema description language.•A novel noise filtering method in teacher–student framework to mitigate the data quality issue.•State-of-the-art performance on standard datasets.

论文关键词:Document-level event argument extraction,Event extraction,Information extraction

论文评审过程:Received 30 June 2022, Revised 13 September 2022, Accepted 13 September 2022, Available online 19 September 2022, Version of Record 28 September 2022.

论文官网地址:https://doi.org/10.1016/j.knosys.2022.109904