Survey of distance measures for quantifying concept drift and shift in numeric data

作者:Igor Goldenberg, Geoffrey I. Webb

摘要

Deployed machine learning systems are necessarily learned from historical data and are often applied to current data. When the world changes, the learned models can lose fidelity. Such changes to the statistical properties of data over time are known as concept drift. Similarly, models are often learned in one context, but need to be applied in another. This is called concept shift. Quantifying the magnitude of drift or shift, especially in the context of covariate drift or shift, or unsupervised learning, requires use of measures of distance between distributions. In this paper, we survey such distance measures with respect to their suitability for estimating drift and shift magnitude between samples of numeric data.

论文关键词:Multivariate concept drift, Mahalanobis distance, Hotelling distance, Hellinger distance, Kullback–Leibler divergence

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10115-018-1257-z