OAK

Generalizable Depth Perception via Foundation Model

Metadata Downloads
Author(s)
Chanhwi Jeong
Type
Thesis
Degree
Master
Department
대학원 AI대학원
Advisor
Jeon, Hae-Gon
Abstract
Monocular depth foundation models have recently advanced significantly, yet their application to depth perception tasks utilizing sensor-derived measurements remains underexplored. This paper introduces two methodologies for integrating depth foun- dation models into generalizable depth perception frameworks. The first employs the foundation model as a teacher to generate high-quality pseudo dense labels for depth enhancement, using a robust scale alignment technique to correct inherent scale dis- crepancies between monocular predictions and sensor measurements. The second ap- plies test-time visual prompt tuning, allowing dynamic adaptation to new sensor data without altering pretrained parameters, reducing computational costs while preserv- ing strong generalization. Extensive experiments across multiple datasets demonstrate superior performance of both methods over existing approaches, highlighting their po- tential to enhance depth perception by leveraging the rich knowledge embedded in monocular depth models.
URI
https://scholar.gist.ac.kr/handle/local/31898
Fulltext
http://gist.dcollection.net/common/orgView/200000896252
Alternative Author(s)
정찬휘
Appears in Collections:
Department of AI Convergence > 3. Theses(Master)
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.