Detail View

Improving Adversarial Robustness via Distillation-Based Purification
Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

DC Field Value Language
dc.contributor.author Koo, Inhwa -
dc.contributor.author Chae, Dong-Kyu -
dc.contributor.author Lee, Sang-Chul -
dc.date.accessioned 2023-12-22T21:40:22Z -
dc.date.available 2023-12-22T21:40:22Z -
dc.date.created 2023-10-25 -
dc.date.issued 2023-10 -
dc.identifier.issn 2076-3417 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/46724 -
dc.description.abstract Despite the impressive performance of deep neural networks on many different vision tasks, they have been known to be vulnerable to intentionally added noise to input images. To combat these adversarial examples (AEs), improving the adversarial robustness of models has emerged as an important research topic, and research has been conducted in various directions including adversarial training, image denoising, and adversarial purification. Among them, this paper focuses on adversarial purification, which is a kind of pre-processing that removes noise before AEs enter a classification model. The advantage of adversarial purification is that it can improve robustness without affecting the model’s nature, while another defense techniques like adversarial training suffer from a decrease in model accuracy. Our proposed purification framework utilizes a Convolutional Autoencoder as a base model to capture the features of images and their spatial structure.We further aim to improve the adversarial robustness of our purification model by distilling the knowledge from teacher models. To this end, we train two Convolutional Autoencoders (teachers), one with adversarial training and the other with normal training. Then, through ensemble knowledge distillation, we transfer the ability of denoising and restoring of original images to the student model (purification model). Our extensive experiments confirm that our student model achieves high purification performance(i.e., how accurately a pre-trained classification model classifies purified images). The ablation study confirms the positive effect of our idea of ensemble knowledge distillation from two teachers on performance. © 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). -
dc.language English -
dc.publisher MDPI -
dc.title Improving Adversarial Robustness via Distillation-Based Purification -
dc.type Article -
dc.identifier.doi 10.3390/app132011313 -
dc.identifier.wosid 001119484700001 -
dc.identifier.scopusid 2-s2.0-85192464467 -
dc.identifier.bibliographicCitation Koo, Inhwa. (2023-10). Improving Adversarial Robustness via Distillation-Based Purification. Applied Sciences, 13(20). doi: 10.3390/app132011313 -
dc.description.isOpenAccess TRUE -
dc.subject.keywordAuthor adversarial robustness -
dc.subject.keywordAuthor adversarial attacks -
dc.subject.keywordAuthor adversarial purification -
dc.subject.keywordAuthor knowledge distillation -
dc.subject.keywordAuthor image classification -
dc.subject.keywordAuthor convolutional autoencoders -
dc.citation.number 20 -
dc.citation.title Applied Sciences -
dc.citation.volume 13 -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.relation.journalResearchArea Chemistry; Engineering; Materials Science; Physics -
dc.relation.journalWebOfScienceCategory Chemistry, Multidisciplinary; Engineering, Multidisciplinary; Materials Science, Multidisciplinary; Physics, Applied -
dc.type.docType Article -
Show Simple Item Record

File Downloads

공유

qrcode
공유하기

Related Researcher

이상철
Lee, Sang-Chul이상철

Division of Nanotechnology

read more

Total Views & Downloads