Detail View
A Study on the Generality of Neural Network Structures for Monocular Depth Estimation
WEB OF SCIENCE
SCOPUS
- Title
- A Study on the Generality of Neural Network Structures for Monocular Depth Estimation
- Issued Date
- 2024-04
- Citation
- Bae, Jinwoo. (2024-04). A Study on the Generality of Neural Network Structures for Monocular Depth Estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 46(4), 2224–2238. doi: 10.1109/TPAMI.2023.3332407
- Type
- Article
- Author Keywords
- Monocular depth estimation ; Out-of-Distribution ; Generalization ; Transformer
- ISSN
- 0162-8828
- Abstract
-
Monocular depth estimation has been widely studied, and significant improvements in performance have been recently reported. However, most previous works are evaluated on a few benchmark datasets, such as KITTI datasets, and none of the works provide an in-depth analysis of the generalization performance of monocular depth estimation. In this paper, we deeply investigate the various backbone networks (
더보기e.g. CNN and Transformer models) toward the generalization of monocular depth estimation. First, we evaluate state-of-the-art models on both in-distribution and out-of-distribution datasets, which have never been seen during network training. Then, we investigate the internal properties of the representations from the intermediate layers of CNN-/Transformer-based models using synthetic texture-shifted datasets. Through extensive experiments, we observe that the Transformers exhibit a strong shape-bias rather than CNNs, which have a strong texture-bias. We also discover that texture-biased models exhibit worse generalization performance for monocular depth estimation than shape-biased models. We demonstrate that similar aspects are observed in real-world driving datasets captured under diverse environments. Lastly, we conduct a dense ablation study with various backbone networks which are utilized in modern strategies. The experiments demonstrate that the intrinsic locality of the CNNs and the self-attention of the Transformers induce texture-bias and shape-bias, respectively. IEEE
- Publisher
- Institute of Electrical and Electronics Engineers
File Downloads
- There are no files associated with this item.
공유
Related Researcher
- Im, Sunghoon임성훈
-
Department of Electrical Engineering and Computer Science
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
