The Stochastic Augmented Lagrangian method for domain adaptation

作者:

Highlights:

摘要

Among various topics explored in the transfer learning community, domain adaptation (DA) has been of primary interest and successfully applied in diverse fields. However, theoretical understanding of learning convergence in DA has not been sufficiently explored. To address such an issue, this paper presents the Stochastic Augmented Lagrangian method (SALM) to solve the optimization problem associated with domain adaptation. In contrast to previous works, the SALM is able to find the optimal Lagrangian multipliers, as opposed to manually selecting the multipliers which could result in significantly suboptimal solutions. Additionally, the SALM is the first algorithm which can find a feasible point with arbitrary precision for domain adaptation problems with bounded penalty parameters. We also observe that with unbounded penalty parameters, the proposed algorithm is able to find an approximate stationary point of infeasibility. We validate our theoretical analysis with several experimental results using benchmark data sets including MNIST, SYNTH, SVHN, and USPS.

论文关键词:Domain adaptation,Augmented Lagrangian,Optimization,Convergence

论文评审过程:Received 23 January 2021, Revised 31 July 2021, Accepted 9 October 2021, Available online 19 October 2021, Version of Record 29 October 2021.

论文官网地址:https://doi.org/10.1016/j.knosys.2021.107593