SwePub
Sök i LIBRIS databas

  Extended search

onr:"swepub:oai:DiVA.org:liu-167712"
 

Search: onr:"swepub:oai:DiVA.org:liu-167712" > Cross-Domain Transf...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Cross-Domain Transferability of Adversarial Perturbations

Naseer, Muzammal (author)
Australian Natl Univ, Australia; Incept Inst Artificial Intelligence, U Arab Emirates
Khan, Salman (author)
Australian Natl Univ, Australia; Incept Inst Artificial Intelligence, U Arab Emirates
Khan, Muhammad Haris (author)
Incept Inst Artificial Intelligence, U Arab Emirates
show more...
Khan, Fahad (author)
Linköpings universitet,Datorseende,Tekniska fakulteten,Incept Inst Artificial Intelligence, U Arab Emirates
Porikli, Fatih (author)
Australian Natl Univ, Australia
show less...
 (creator_code:org_t)
NEURAL INFORMATION PROCESSING SYSTEMS (NIPS), 2019
2019
English.
In: ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019). - : NEURAL INFORMATION PROCESSING SYSTEMS (NIPS).
  • Conference paper (peer-reviewed)
Abstract Subject headings
Close  
  • Adversarial examples reveal the blind spots of deep neural networks (DNNs) and represent a major concern for security-critical applications. The transferability of adversarial examples makes real-world attacks possible in black-box settings, where the attacker is forbidden to access the internal parameters of the model. The underlying assumption in most adversary generation methods, whether learning an instance-specific or an instance-agnostic perturbation, is the direct or indirect reliance on the original domain-specific data distribution. In this work, for the first time, we demonstrate the existence of domain-invariant adversaries, thereby showing common adversarial space among different datasets and models. To this end, we propose a framework capable of launching highly transferable attacks that crafts adversarial patterns to mislead networks trained on entirely different domains. For instance, an adversarial function learned on Paintings, Cartoons or Medical images can successfully perturb ImageNet samples to fool the classifier, with success rates as high as similar to 99% (l(infinity) <= 10). The core of our proposed adversarial function is a generative network that is trained using a relativistic supervisory signal that enables domain-invariant perturbations. Our approach sets the new state-of-the-art for fooling rates, both under the white-box and black-box scenarios. Furthermore, despite being an instance-agnostic perturbation function, our attack outperforms the conventionally much stronger instance-specific attack methods.

Subject headings

NATURVETENSKAP  -- Matematik -- Sannolikhetsteori och statistik (hsv//swe)
NATURAL SCIENCES  -- Mathematics -- Probability Theory and Statistics (hsv//eng)

Publication and Content Type

ref (subject category)
kon (subject category)

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Find more in SwePub

By the author/editor
Naseer, Muzammal
Khan, Salman
Khan, Muhammad H ...
Khan, Fahad
Porikli, Fatih
About the subject
NATURAL SCIENCES
NATURAL SCIENCES
and Mathematics
and Probability Theo ...
Articles in the publication
By the university
Linköping University

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view