Adjustment and Alignment for Unbiased Open Set Domain Adaptation

Wuyang Li, Jie Liu, Bo Han, Yixuan Yuan*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

60 Citations (Scopus)

Abstract

Open Set Domain Adaptation (OSDA) transfers the model from a label-rich domain to a label-free one containing novel-class samples. Existing OSDA works overlook abundant novel-class semantics hidden in the source domain, leading to a biased model learning and transfer. Although the causality has been studied to remove the semantic-level bias, the non-available novel-class samples result in the failure of existing causal solutions in OSDA. To break through this barrier, we propose a novel causality-driven solution with the unexplored front-door adjustment theory, and then implement it with a theoretically grounded framework, coined AdjustmeNt aNd Alignment (ANNA), to achieve an unbiased OSDA. In a nutshell, ANNA consists of Front-Door Adjustment (FDA) to correct the biased learning in the source domain and Decoupled Causal Alignment (DCA) to transfer the model unbiasedly. On the one hand, FDA delves into fine-grained visual blocks to discover novel-class regions hidden in the base-class image. Then, it corrects the biased model optimization by implementing causal debiasing. On the other hand, DCA disentangles the base-class and novel-class regions with orthogonal masks, and then adapts the decoupled distribution for an unbiased model transfer. Extensive experiments show that ANNA achieves state-of-the-art results. The code is available at https://github.com/CityU-AIM-Group/Anna. ©2023 IEEE.
Original languageEnglish
Title of host publicationProceedings - 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition CVPR 2023
PublisherIEEE
Pages24110-24119
ISBN (Electronic)979-8-3503-0129-8
ISBN (Print)979-8-3503-0130-4
DOIs
Publication statusPublished - 2023
Event2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023) - Vancouver Convention Center, Vancouver, Canada
Duration: 18 Jun 202322 Jun 2023
https://cvpr2023.thecvf.com/Conferences/2023
https://openaccess.thecvf.com/menu
https://ieeexplore.ieee.org/xpl/conhome/1000147/all-proceedings

Publication series

NameProceedings - IEEE Computer Society Conference on Computer Vision and Pattern Recognition
ISSN (Print)1063-6919
ISSN (Electronic)2575-7075

Conference

Conference2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023)
Abbreviated titleCVPR2023
PlaceCanada
CityVancouver
Period18/06/2322/06/23
Internet address

Bibliographical note

Research Unit(s) information for this publication is provided by the author(s) concerned.

Funding

This work was supported by Hong Kong Research Grants Council (RGC) General Research Fund 11211221, and Innovation and Technology Commission-Innovation and Technology Fund ITS/100/20.

RGC Funding Information

  • RGC-funded

Fingerprint

Dive into the research topics of 'Adjustment and Alignment for Unbiased Open Set Domain Adaptation'. Together they form a unique fingerprint.

Cite this