OAK

Learning CNN on ViT: A Hybrid Model to Explicitly Class-specific Boundaries for Domain Adaptation

Metadata Downloads
Abstract
Most domain adaptation (DA) methods are based on either a convolutional neural networks (CNNs) or a vision transformers (ViTs). They align the distribution differences between domains as encoders without considering their unique characteristics. For instance, ViT excels in accuracy due to its superior ability to capture global representations, while CNN has an advantage in capturing local representations. This fact has led us to design a hybrid method to fully take advantage of both ViT and CNN, called
xplicitly
lass-specific
oundaries (
). ECB learns CNN on ViT to combine their distinct strengths. In particular, we leverage ViT's properties to explicitly find class-specific decision boundaries by maximizing the discrepancy between the outputs of the two classifiers to detect target samples far from the source support. In contrast, the CNN encoder clusters target features based on the previously defined class-specific boundaries by minimizing the discrepancy between the probabilities of the two classifiers. Finally, ViT and CNN mutually exchange knowledge to improve the quality of pseudo labels and reduce the knowledge discrepancies of these models. Compared to conventional DA methods, our ECB achieves superior performance, which verifies its effectiveness in this hybrid model. The project website can be found https://dotrannhattuong.github.io/ECB/website/.
Author(s)
Ba Hung NgoNhat-Tuong Do-TranTuan-Ngoc NguyenJeon, Hae-GonTae Jong Choi
Issued Date
2024-06-17
Type
Conference Paper
DOI
10.1109/CVPR52733.2024.02697
URI
https://scholar.gist.ac.kr/handle/local/20915
Publisher
IEEE Computer Society
Citation
2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024, pp.28545 - 28554
ISSN
10636919
Conference Place
US
Seattle Convention Center
Appears in Collections:
Department of AI Convergence > 2. Conference Papers
공개 및 라이선스
  • 공개 구분공개
파일 목록

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.