Unsupervised and few-shot parsing from pretrained language models

作者:

摘要

Pretrained language models are generally acknowledged to be able to encode syntax [46], [16], [13]. In this article, we propose UPOA, an Unsupervised constituent Parsing model that calculates an Out Association score solely based on the self-attention weight matrix learned in a pretrained language model as the syntactic distance for span segmentation. We further propose an enhanced version, UPIO, which exploits both inside association and outside association scores for estimating the likelihood of a span. Experiments with UPOA and UPIO disclose that the linear projection matrices for the query and key in the self-attention mechanism play an important role in parsing. We therefore extend the unsupervised models to few-shot parsing models (FPOA, FPIO) that use a few annotated trees to learn better linear projection matrices for parsing. Experiments on the Penn Treebank demonstrate that our unsupervised parsing model UPIO achieves results comparable to the state of the art on short sentences (length <= 10). Our few-shot parsing model FPIO trained with only 20 annotated trees outperforms a previous few-shot parsing method trained with 50 annotated trees. Experiments on cross-lingual parsing show that both unsupervised and few-shot parsing methods are better than previous methods on most languages of SPMRL [39].

论文关键词:Unsupervised constituent parsing,Few-shot parsing,Pretrained language model,Self-attention mechanism

论文评审过程:Received 8 February 2021, Revised 7 July 2021, Accepted 20 January 2022, Available online 26 January 2022, Version of Record 31 January 2022.

论文官网地址:https://doi.org/10.1016/j.artint.2022.103665