Probabilistic logic with minimum perplexity: Application to language modeling

作者:

Highlights:

摘要

Any statistical model based on training encounters sparse configurations. These data are those that have not been encountered (or seen) during the training phase. This inherent problem is a big challenge to many scientific communities. The statistical estimation of rare events is usually performed through the maximum likelihood (ML) criterion. However, it is well-known that the ML estimator is sensitive to extreme values that is therefore non-reliable. To answer this challenge, we propose a novel approach based on probabilistic logic (PL) and the minimal perplexity criterion. In our approach, configurations are considered as probabilistic events such as predicates related through logical connectors. Our method was applied to estimate word trigram probability values from a corpus. Experimental results conducted on several test sets show that the PL method with minimal perplexity has outperformed both the “Absolute Discounting”, and the “Good-Turing Discounting” techniques.

论文关键词:Word trigrams,Probabilistic logic,Statistical language model,Maximum likelihood estimation,Sparseness problem,Minimum perplexity,Entropy maximization

论文评审过程:Received 27 May 2004, Accepted 27 December 2004, Available online 14 March 2005.

论文官网地址:https://doi.org/10.1016/j.patcog.2004.12.009