Aktionen

Adversarial Attacks: Unterschied zwischen den Versionen

Aus exmediawiki

Zeile 1: Zeile 1:
KNN's sind extrem anfällig für...
 
 
 
* Praxis-Beispiele: https://boingboing.net/tag/adversarial-examples
 
* Praxis-Beispiele: https://boingboing.net/tag/adversarial-examples
 
* https://bdtechtalks.com/2018/12/27/deep-learning-adversarial-attacks-ai-malware/
 
* https://bdtechtalks.com/2018/12/27/deep-learning-adversarial-attacks-ai-malware/
Zeile 8: Zeile 6:
  
 
* https://en.wikipedia.org/wiki/Deep_learning#Cyberthreat
 
* https://en.wikipedia.org/wiki/Deep_learning#Cyberthreat
 
+
[[File:Adversarial-pig.png|800]]
----
 
----  
 
 
=WHITE BOX ATTACKS=
 
=WHITE BOX ATTACKS=
 
* https://cv-tricks.com/how-to/breaking-deep-learning-with-adversarial-examples-using-tensorflow/
 
* https://cv-tricks.com/how-to/breaking-deep-learning-with-adversarial-examples-using-tensorflow/

Version vom 16. April 2019, 16:41 Uhr

800

WHITE BOX ATTACKS


Untargeted Adversarial Attacks

Adversarial attacks that just want your model to be confused and predict a wrong class are called Untargeted Adversarial Attacks.

  • nicht zielgerichtet

Fast Gradient Sign Method(FGSM)

FGSM is a single step attack, ie.. the perturbation is added in a single step instead of adding it over a loop (Iterative attack).

Basic Iterative Method

Störung, anstatt in einem einzelnen Schritt in mehrere kleinen Schrittgrößen anwenden

Iterative Least-Likely Class Method

ein Bild erstellen, welches in der Vorhersage den niedrigsten Score trägt


Targeted Adversarial Attacks

Attacks which compel the model to predict a (wrong) desired output are called Targeted Adversarial attacks

  • zielgerichtet

(Un-)Targeted Adversarial Attacks

kann beides...

Projected Gradient Descent (PGD)

Eine Störung finden die den Verlust eines Modells bei einer bestimmten Eingabe maximiert:



BLACK BOX ATTACKS


on computer vision

propose zeroth order optimization (ZOO)

Black-Box Attacks using Adversarial Samples

  • a technique that uses the victim model as an oracle to label a synthetic training set for the substitute, so the attacker need not even collect a training set to mount the attack

new Tesla Hack


on voice (ASR)

hidden voice commands

Psychoacoustic Hiding (Attacking Speech Recognition)


on written text (NLP)

paraphrasing attacks


Anti Surveillance

http://dismagazine.com/dystopia/evolved-lifestyles/8115/anti-surveillance-how-to-hide-from-machines/


libraries