Sliced Wasserstein adversarial training for improving adversarial robustness

Woojin Lee, Sungyoon Lee, Hoki Kim, Jaewook Lee

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Recently, deep-learning-based models have achieved impressive performance on tasks that were previously considered to be extremely challenging. However, recent works have shown that various deep learning models are susceptible to adversarial data samples. In this paper, we propose the sliced Wasserstein adversarial training method to encourage the logit distributions of clean and adversarial data to be similar to each other. We capture the dissimilarity between two distributions using the Wasserstein metric and then align distributions using an end-to-end training process. We present the theoretical background of the motivation for our study by providing generalization error bounds for adversarial data samples. We performed experiments on three standard datasets and the results demonstrate that our method is more robust against white box attacks compared to previous methods.

Original languageEnglish
Pages (from-to)3229-3242
Number of pages14
JournalJournal of Ambient Intelligence and Humanized Computing
Volume15
Issue number8
DOIs
StatePublished - Aug 2024

Keywords

  • Adversarial attack
  • Adversarial defense
  • Adversarial training
  • Sliced Wasserstein Distance

Fingerprint

Dive into the research topics of 'Sliced Wasserstein adversarial training for improving adversarial robustness'. Together they form a unique fingerprint.

Cite this