Detail View

Vision Transformer 기반 Trimap 분할 기법

Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

Title
Vision Transformer 기반 Trimap 분할 기법
Alternative Title
Trimap Segmentation Method Based on Vision Transformer
Issued Date
2025-06-26
Citation
대한전자공학회 2025년도 하계종합학술대회, pp.2219 - 2222
Type
Conference Paper
Abstract

In this study, we propose a trimap prediction model aimed at achieving more accurate foreground extraction compared to conventional background removal techniques. The proposed model is based on a click-free structure that does not rely on user clicks or point-based interactions and is constructed by modifying a Vision Transformer (ViT) backbone to accept both an image and a segmentation mask as inputs. The output consists of a three-class trimap comprising background, unknown, and foreground regions. For training, we constructed a large-scale trimap dataset by integrating real-world datasets such as AM-2K, AIM-500, P3M-10K, and Composition-431K [1]. In addition, a composite loss function combining Normalized Focal Loss and Unknown Region Distance Transform Loss was applied to encourage the model to focus more effectively on the unknown regions. Experimental results demonstrate that the proposed model accurately predicts trimap boundaries without any user interaction and effectively enhances segmentation performance.

더보기
URI
https://scholar.dgist.ac.kr/handle/20.500.11750/60073
Publisher
대한전자공학회
Show Full Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

이상헌
Lee, Sang-Heon이상헌

Division of Mobility Technology

read more

Total Views & Downloads

???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???: