L1-Norm Based Adversarial Example against CNN
碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particul...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/ua49z8 |
id |
ndltd-TW-106NCHU5394030 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-106NCHU53940302019-05-16T01:24:30Z http://ndltd.ncl.edu.tw/handle/ua49z8 L1-Norm Based Adversarial Example against CNN 植基於 L1 範數針對卷積神經網路的對抗樣本 Pei-Hsuan Lu 呂姵萱 碩士 國立中興大學 資訊科學與工程學系 106 In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particular, MagNet consisting of an adversary detector and a data reformer is by far one of the strongest defenses in the black-box setting, where the attacker aims to craft transferable adversarial examples from an undefended DNN model to bypass a defense module without knowing its existence. MagNet can successfully defend a variety of attacks in DNNs, including the Carlini and Wagner''s transfer attack based on the L2 distortion metric. However, in this thesis, under the black-box transfer attack setting we show that adversarial examples crafted based on the L1 distortion metric can easily bypass MagNet and fool the target DNN image classifiers on MNIST and CIFAR-10. We also provide theoretical justification on why the considered approach can yield adversarial examples with superior attack transferability and conduct extensive experiments on variants of MagNet to verify its lack of robustness to L1 distortion based transfer attacks. Notably, our results substantially weaken the existing transfer attack assumption of knowing the deployed defense technique when attacking defended DNNs (i.e., the gray-box setting). Chia-Mu Yu 游家牧 2018 學位論文 ; thesis 27 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立中興大學 === 資訊科學與工程學系 === 106 === In recent years, defending adversarial perturbations to natural examples in order to build robust machine learning models trained by deep neural networks (DNNs) has become an emerging research field in the conjunction of deep learning and security. In particular, MagNet consisting of an adversary detector and a data reformer is by far one of the strongest defenses in the black-box setting, where the attacker aims to craft transferable adversarial examples from an undefended DNN model to bypass a defense module without knowing its existence. MagNet can successfully defend a variety of attacks in DNNs, including the Carlini and Wagner''s transfer attack based on the L2 distortion metric. However, in this thesis, under the black-box transfer attack setting we show that adversarial examples crafted based on the L1 distortion metric can easily bypass MagNet and fool the target DNN image classifiers on MNIST and CIFAR-10. We also provide theoretical justification on why the considered approach can yield adversarial examples with superior attack transferability and conduct extensive experiments on variants of MagNet to verify its lack of robustness to L1 distortion based transfer attacks. Notably, our results substantially weaken the existing transfer attack assumption of knowing the deployed defense technique when attacking defended DNNs (i.e., the gray-box setting).
|
author2 |
Chia-Mu Yu |
author_facet |
Chia-Mu Yu Pei-Hsuan Lu 呂姵萱 |
author |
Pei-Hsuan Lu 呂姵萱 |
spellingShingle |
Pei-Hsuan Lu 呂姵萱 L1-Norm Based Adversarial Example against CNN |
author_sort |
Pei-Hsuan Lu |
title |
L1-Norm Based Adversarial Example against CNN |
title_short |
L1-Norm Based Adversarial Example against CNN |
title_full |
L1-Norm Based Adversarial Example against CNN |
title_fullStr |
L1-Norm Based Adversarial Example against CNN |
title_full_unstemmed |
L1-Norm Based Adversarial Example against CNN |
title_sort |
l1-norm based adversarial example against cnn |
publishDate |
2018 |
url |
http://ndltd.ncl.edu.tw/handle/ua49z8 |
work_keys_str_mv |
AT peihsuanlu l1normbasedadversarialexampleagainstcnn AT lǚpèixuān l1normbasedadversarialexampleagainstcnn AT peihsuanlu zhíjīyúl1fànshùzhēnduìjuǎnjīshénjīngwǎnglùdeduìkàngyàngběn AT lǚpèixuān zhíjīyúl1fànshùzhēnduìjuǎnjīshénjīngwǎnglùdeduìkàngyàngběn |
_version_ |
1719175205421580288 |