Domain Adaptation with Conditional Distribution Matching and Generalized Label Shift

  • Remi Tachet des Combes ,
  • Han Zhao ,
  • Yu-Xiang Wang ,
  • Geoff Gordon

NeurIPS |

Adversarial learning has demonstrated good performance in the unsupervised domain adaptation setting, by learning domain-invariant representations that perform well on the source domain. However, recent work has underlined limitations of existing methods in the presence of mismatched label distributions between the source and target domains. In this paper, we extend a recent upper-bound on the performance of adversarial domain adaptation to multi-class classification and more general discriminators. We then propose generalized label shift (GLS) as a way to improve robustness against mismatched label distributions. GLS states that, conditioned on the label, there exists a representation of the input that is invariant between the source and target domains. Under GLS, we provide theoretical guarantees on the transfer performance of any classifier. We also devise necessary and sufficient conditions for GLS to hold. The conditions are based on the estimation of the relative class weights between domains and on an appropriate reweighting of samples. Guided by our theoretical insights, we modify three widely used algorithms, JAN, DANN and CDAN and evaluate their performance on standard domain adaptation tasks where our method outperforms the base versions. We also demonstrate significant gains on artificially created tasks with large divergences between their source and target label distributions.

Publication Downloads

Domain Adaptation with Conditional Distribution Matching and Generalized Label Shift (code)

December 3, 2020

Unsupervised Domain Adaptation algorithms robust to mismatched label distributions