A novel domain adaptation theory with Jensen–Shannon divergence

作者:

Highlights:

摘要

Domain adaptation aims to alleviate the shift between training and test distribution, where the DA theory is crucial in understanding the success of domain adaptation algorithms. In this paper, we reveal the incoherence between the empirical domain adversarial training and its generally assumed theoretical counterpart based on H-divergence. Concretely, we find that H-divergence is not equivalent to Jensen–Shannon divergence, the optimization objective in domain adversarial training. To this end, we establish a new theoretical framework by directly proving the upper and lower target risk bounds based on the joint distributional Jensen–Shannon divergence. We further derive bidirectional upper bounds for marginal and conditional shifts. Our framework exhibits inherent flexibility for different transfer learning problems, which is usable for various scenarios. From an algorithmic perspective, our theory enables a generic guideline of the unified principles of semantic conditional matching, feature marginal matching, and label marginal shift correction. We employ algorithms for each principle and empirically validate the benefits of our framework.

论文关键词:Learning theory,Transfer learning,Domain adaptation,Representation learning,Deep learning

论文评审过程:Received 25 March 2022, Revised 19 July 2022, Accepted 27 August 2022, Available online 24 September 2022, Version of Record 7 October 2022.

论文官网地址:https://doi.org/10.1016/j.knosys.2022.109808