Detail View

DC Field Value Language
dc.contributor.author Ullah, Ihsan -
dc.contributor.author An, Sion -
dc.contributor.author Kang, Myeongkyun -
dc.contributor.author Chikontwe, Philip -
dc.contributor.author Lee, HyunKi -
dc.contributor.author Choi, Jinwoo -
dc.contributor.author Park, Sang Hyun -
dc.date.accessioned 2024-12-20T21:40:19Z -
dc.date.available 2024-12-20T21:40:19Z -
dc.date.created 2024-08-05 -
dc.date.issued 2024-11 -
dc.identifier.issn 0893-6080 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/57336 -
dc.description.abstract Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous and related labeled datasets (sources) to a new unlabeled dataset (target). Despite the impressive performance, existing approaches have largely focused on image-based UDA only, and video-based UDA has been relatively understudied and received less attention due to the difficulty of adapting diverse modal video features and modeling temporal associations efficiently. To address this, existing studies use optical flow to capture motion cues between in-domain consecutive frames, but is limited by heavy compute requirements and modeling flow patterns across diverse domains is equally challenging. In this work, we propose an adversarial domain adaptation approach for video semantic segmentation that aims to align temporally associated pixels in successive source and target domain frames without relying on optical flow. Specifically, we introduce a Perceptual Consistency Matching (PCM) strategy that leverages perceptual similarity to identify pixels with high correlation across consecutive frames, and infer that such pixels should correspond to the same class. Therefore, we can enhance prediction accuracy for video-UDA by enforcing consistency not only between in-domain frames, but across domains using PCM objectives during model training. Extensive experiments on public datasets show the benefit of our approach over existing state-of-the-art UDA methods. Our approach not only addresses a crucial task in video domain adaptation but also offers notable improvements in performance with faster inference times. © 2024 Elsevier Ltd -
dc.language English -
dc.publisher Elsevier -
dc.title Video domain adaptation for semantic segmentation using perceptual consistency matching -
dc.type Article -
dc.identifier.doi 10.1016/j.neunet.2024.106505 -
dc.identifier.wosid 001271589600001 -
dc.identifier.scopusid 2-s2.0-85198316613 -
dc.identifier.bibliographicCitation Ullah, Ihsan. (2024-11). Video domain adaptation for semantic segmentation using perceptual consistency matching. Neural Networks, 179. doi: 10.1016/j.neunet.2024.106505 -
dc.description.isOpenAccess FALSE -
dc.subject.keywordAuthor Unsupervised domain adaptation -
dc.subject.keywordAuthor Video domain adaptation -
dc.subject.keywordAuthor Semantic segmentation -
dc.subject.keywordAuthor Consistency matching -
dc.citation.title Neural Networks -
dc.citation.volume 179 -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.relation.journalResearchArea Computer Science; Neurosciences & Neurology -
dc.relation.journalWebOfScienceCategory Computer Science, Artificial Intelligence; Neurosciences -
dc.type.docType Article -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

이현기
Lee, HyunKi이현기

Division of Intelligent Robotics

read more

Total Views & Downloads