Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
37th (2023)
Session ID : 2K5-GS-2-01
Conference information

An Attempt to Rectify Classification Results Using Vulnerability of Adversarial Examples
*Fumiya MORIMOTOKeigo AKAGAKISatoshi ONO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Deep neural networks (DNNs) have shown high performance in various fields, such as image classification and speech recognition, and are being applied in real-world applications. On the other hand, recent studies have revealed that DNN-based classifiers have the vulnerability of misrecognizing Adversarial Examples (AEs), which are small and specially perturbed input data to the extent that they are difficult for humans to perceive. For this reason, research on defense methods against AEs has been widely conducted. For example, detection methods that discriminate AEs based on features of input samples have been proposed, but they only detect AEs and do not consider AEs’ correct categories. While many tasks can simply reject detected AEs, some tasks, such as sign recognition for autonomous driving, require correct categories of AEs. This is because, when an attack is made on a stop sign, DNNs with the defense method cannot recognize it as a stop sign though they can detect the attacked sign. Such tasks require some post-processing in addition to detect AEs. For this reason, we propose a label rectification method for AEs detected by the defense method, that is, a method to estimate the correct labels in the original images of the AEs. This method based on counter-attacking can correct the misclassification results to those of the original images.

Content from these authors
© 2023 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top