Detail View
Vision Transformer 기반 Trimap 분할 기법
WEB OF SCIENCE
SCOPUS
Metadata Downloads
- Title
- Vision Transformer 기반 Trimap 분할 기법
- Alternative Title
- Trimap Segmentation Method Based on Vision Transformer
- Issued Date
- 2025-06-26
- Citation
- 대한전자공학회 2025년도 하계종합학술대회, pp.2219 - 2222
- Type
- Conference Paper
- Abstract
-
In this study, we propose a trimap prediction model aimed at achieving more accurate foreground extraction compared to conventional background removal techniques. The proposed model is based on a click-free structure that does not rely on user clicks or point-based interactions and is constructed by modifying a Vision Transformer (ViT) backbone to accept both an image and a segmentation mask as inputs. The output consists of a three-class trimap comprising background, unknown, and foreground regions. For training, we constructed a large-scale trimap dataset by integrating real-world datasets such as AM-2K, AIM-500, P3M-10K, and Composition-431K [1]. In addition, a composite loss function combining Normalized Focal Loss and Unknown Region Distance Transform Loss was applied to encourage the model to focus more effectively on the unknown regions. Experimental results demonstrate that the proposed model accurately predicts trimap boundaries without any user interaction and effectively enhances segmentation performance.
더보기
- Publisher
- 대한전자공학회
File Downloads
- There are no files associated with this item.
공유
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
