Zero-shot domain adaptation for natural language inference by projecting superficial words out

作者:

Highlights:

摘要

In natural language inference, the semantics of some words do not affect the inference. Such information is considered superficial and brings overfitting. In this paper, we project the superficial information out to learn a more general representation. The generality refer to the adaptation to different domains. The projection is over the dropout framework. To project all information of recurring words out, we propose a parameter-free model (HardDrop). By further noticing that the information of some recurring words needs to be reserved, we propose SoftDrop to learn to cautiously project the information out. Our approaches outperform the competitors over the source domain and over zero-shot target domains. For some target domains (e.g. from RTE to MRPC), the accuracy/f1-score increases from 47.6%/0.501 to 68.1%/0.771.

论文关键词:Natural language inference,Zero-shot learning,Domain adaptation

论文评审过程:Received 21 July 2020, Revised 30 May 2021, Accepted 2 June 2021, Available online 14 June 2021, Version of Record 14 June 2021.

论文官网地址:https://doi.org/10.1016/j.knosys.2021.107193