L1-Norm Based Adversarial Example against CNN

碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particul...

Full description

Bibliographic Details
Main Authors: Pei-Hsuan Lu, 呂姵萱
Other Authors: Chia-Mu Yu
Format: Others
Language:en_US
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/ua49z8
id ndltd-TW-106NCHU5394030
record_format oai_dc
spelling ndltd-TW-106NCHU53940302019-05-16T01:24:30Z http://ndltd.ncl.edu.tw/handle/ua49z8 L1-Norm Based Adversarial Example against CNN 植基於 L1 範數針對卷積神經網路的對抗樣本 Pei-Hsuan Lu 呂姵萱 碩士 國立中興大學 資訊科學與工程學系 106 In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particular, MagNet consisting of an adversary detector and a data reformer is by far one of the strongest defenses in the black-box setting, where the attacker aims to craft transferable adversarial examples from an undefended DNN model to bypass a defense module without knowing its existence. MagNet can successfully defend a variety of attacks in DNNs, including the Carlini and Wagner''s transfer attack based on the L2 distortion metric. However, in this thesis, under the black-box transfer attack setting we show that adversarial examples crafted based on the L1 distortion metric can easily bypass MagNet and fool the target DNN image classifiers on MNIST and CIFAR-10. We also provide theoretical justification on why the considered approach can yield adversarial examples with superior attack transferability and conduct extensive experiments on variants of MagNet to verify its lack of robustness to L1 distortion based transfer attacks. Notably, our results substantially weaken the existing transfer attack assumption of knowing the deployed defense technique when attacking defended DNNs (i.e., the gray-box setting). Chia-Mu Yu 游家牧 2018 學位論文 ; thesis 27 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particular, MagNet consisting of an adversary detector and a data reformer is by far one of the strongest defenses in the black-box setting, where the attacker aims to craft transferable adversarial examples from an undefended DNN model to bypass a defense module without knowing its existence. MagNet can successfully defend a variety of attacks in DNNs, including the Carlini and Wagner''s transfer attack based on the L2 distortion metric. However, in this thesis, under the black-box transfer attack setting we show that adversarial examples crafted based on the L1 distortion metric can easily bypass MagNet and fool the target DNN image classifiers on MNIST and CIFAR-10. We also provide theoretical justification on why the considered approach can yield adversarial examples with superior attack transferability and conduct extensive experiments on variants of MagNet to verify its lack of robustness to L1 distortion based transfer attacks. Notably, our results substantially weaken the existing transfer attack assumption of knowing the deployed defense technique when attacking defended DNNs (i.e., the gray-box setting).
author2 Chia-Mu Yu
author_facet Chia-Mu Yu
Pei-Hsuan Lu
呂姵萱
author Pei-Hsuan Lu
呂姵萱
spellingShingle Pei-Hsuan Lu
呂姵萱
L1-Norm Based Adversarial Example against CNN
author_sort Pei-Hsuan Lu
title L1-Norm Based Adversarial Example against CNN
title_short L1-Norm Based Adversarial Example against CNN
title_full L1-Norm Based Adversarial Example against CNN
title_fullStr L1-Norm Based Adversarial Example against CNN
title_full_unstemmed L1-Norm Based Adversarial Example against CNN
title_sort l1-norm based adversarial example against cnn
publishDate 2018
url http://ndltd.ncl.edu.tw/handle/ua49z8
work_keys_str_mv AT peihsuanlu l1normbasedadversarialexampleagainstcnn
AT lǚpèixuān l1normbasedadversarialexampleagainstcnn
AT peihsuanlu zhíjīyúl1fànshùzhēnduìjuǎnjīshénjīngwǎnglùdeduìkàngyàngběn
AT lǚpèixuān zhíjīyúl1fànshùzhēnduìjuǎnjīshénjīngwǎnglùdeduìkàngyàngběn
_version_ 1719175205421580288