{"id":10442,"date":"2020-08-19T16:00:00","date_gmt":"2020-08-19T14:00:00","guid":{"rendered":"https:\/\/blog.mi.hdm-stuttgart.de\/?p=10442"},"modified":"2023-08-06T21:45:42","modified_gmt":"2023-08-06T19:45:42","slug":"adversarial-attacks","status":"publish","type":"post","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2020\/08\/19\/adversarial-attacks\/","title":{"rendered":"The Dark Side of AI &#8211; Part 2: Adversarial Attacks"},"content":{"rendered":"\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/08\/photo-1555255707-c07966088b7b.jpg\" alt=\"\"\/><\/figure>\n\n\n\n<p><a href=\"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2020\/08\/19\/ai-cyberattacks-deepfakes\/\"><strong>Part 1<\/strong><\/a> of this series discussed how AI technology can be used for good or evil. But what if the AI itself becomes an attack vector? Could an attacker use my models against me? Also, what\u2019s the worst that could happen? Welcome to the domain of adversarial AI!<\/p>\n\n\n\n<!--more-->\n\n\n\n<h1 class=\"wp-block-heading\">Introduction<\/h1>\n\n\n\n<p>If you\u2019re reading technology-focused magazines or blogs, you might have stumbled upon the term <em>adversarial examples<\/em> before. In recent years (mostly since 2013), there has been a large number of publications on them and their possible impact on AI technology. According to the <a href=\"https:\/\/openai.com\/blog\/adversarial-example-research\/\">OpenAI blog<\/a>:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>Adversarial examples are inputs to machine learning models that an attacker has intentionally designed to cause the model to make a mistake; they\u2019re like optical illusions for machines.<\/p>\n<\/blockquote>\n\n\n\n<h2 class=\"wp-block-heading\">Some Fun Examples<\/h2>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"336\" data-attachment-id=\"10498\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/fgsm_panda_image_preds_1\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1.png\" data-orig-size=\"1590,522\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"fgsm_panda_image_preds_1\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-1024x336.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-1024x336.png\" alt=\"\" class=\"wp-image-10498\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-1024x336.png 1024w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-300x98.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-768x252.png 768w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1-1536x504.png 1536w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds_1.png 1590w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">An untargeted attack using the <em>Fast Gradient Sign Method<\/em> (FGSM)<br>Source: <a href=\"https:\/\/arxiv.org\/pdf\/1412.6572.pdf\">Explaining and Harnessing Adversarial Examples, Goodfellow et al., 2015<\/a><\/figcaption><\/figure>\n\n\n\n<p>The example above shows one of the earlier attacks. In short, an attacker generates some very specific noise, which turns a regular image into one that is classified incorrectly. This noise is so small that it is invisible to the human eye.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"642\" data-attachment-id=\"10495\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/adversarial_patches\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches.png\" data-orig-size=\"1231,772\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"adversarial_patches\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches-1024x642.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches-1024x642.png\" alt=\"\" class=\"wp-image-10495\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches-1024x642.png 1024w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches-300x188.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches-768x482.png 768w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adversarial_patches.png 1231w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">Source: <a href=\"https:\/\/arxiv.org\/pdf\/1712.09665.pdf\">Adversarial Patch, Brown et al., 2018<\/a><\/figcaption><\/figure>\n\n\n\n<p>Adversarial attacks can also be applied in the physical world. In this example, a so-called <em>adversarial patch<\/em> is created. If the patch is present, the attacked classifier detects the image as a toaster, instead of a banana. Such a patch is visible to the human eye, but there are other ways to hide or disguise it. For example, it might be shaped like a pair of glasses. Patches on traffic signs may likewise be designed to look similar to graffiti.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Some Not so Fun Examples<\/h2>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"883\" height=\"418\" data-attachment-id=\"10494\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/robust_adversarial_samples\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples.png\" data-orig-size=\"883,418\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"robust_adversarial_samples\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples.png\" alt=\"\" class=\"wp-image-10494\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples.png 883w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples-300x142.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/robust_adversarial_samples-768x364.png 768w\" sizes=\"auto, (max-width: 883px) 100vw, 883px\" \/><figcaption class=\"wp-element-caption\">Source: <a href=\"https:\/\/arxiv.org\/pdf\/1707.08945.pdf\">Robust Physical-World Attacks on Deep Learning Visual Classification, Eykholt et al., 2018<\/a><\/figcaption><\/figure>\n\n\n\n<p>Adversarial attacks are most prevalent in the domain of <em>computer vision<\/em>, the majority being aimed at misclassification. There are many attack scenarios, from spoofing facial recognition systems to fooling autonomous vehicles. In this example, adversarial examples cause a classifier to mistake a \u201cStop\u201d sign for a \u201cSpeed limit 45\u201d sign.&nbsp;Notably, this attack is robust against changes in scale, rotation and shift of the adversarial patch, which means that it can fool a driving car.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"810\" height=\"575\" data-attachment-id=\"10491\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/fake_lane_attack_mix\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix.png\" data-orig-size=\"810,575\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"fake_lane_attack_mix\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix.png\" alt=\"\" class=\"wp-image-10491\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix.png 810w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix-300x213.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fake_lane_attack_mix-768x545.png 768w\" sizes=\"auto, (max-width: 810px) 100vw, 810px\" \/><figcaption class=\"wp-element-caption\">Source: <a href=\"https:\/\/keenlab.tencent.com\/en\/whitepapers\/Experimental_Security_Research_of_Tesla_Autopilot.pdf\">Experimental Security Research of Tesla Autopilot, Tencent Keen Security Lab, 2019<\/a><\/figcaption><\/figure>\n\n\n\n<p>A security report on \u201cTesla Autopilot\u201d (HW 2.5) from March 2019 found that the platform can be attacked with adversarial examples. The researchers placed adversarial patches on the road to cause the vehicle to detect and follow lanes which weren\u2019t there. The fact that this flaw is (or was) present in a commercial product, firmly places this example in the \u201cnot so fun\u201d section.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"716\" height=\"300\" data-attachment-id=\"10488\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/bert_attack_2_acc\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/bert_attack_2_acc.png\" data-orig-size=\"716,300\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"bert_attack_2_acc\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/bert_attack_2_acc.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/bert_attack_2_acc.png\" alt=\"\" class=\"wp-image-10488\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/bert_attack_2_acc.png 716w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/bert_attack_2_acc-300x126.png 300w\" sizes=\"auto, (max-width: 716px) 100vw, 716px\" \/><figcaption class=\"wp-element-caption\">Performance of the BERT language model on various text classification tasks, before and after the adversarial attack.&nbsp;<br>Source: <a href=\"http:\/\/BERT-ATTACK: Adversarial Attack Against BERT Using BERT, Li et al., 2020\">BERT-ATTACK: Adversarial Attack Against BERT Using BERT, Li et al., 2020<\/a><\/figcaption><\/figure>\n\n\n\n<p>Adversarial attacks are by no means restricted to the domain of computer vision. Here is an example from the domain of <em>Natural Language Processing<\/em>. It shows a successful adversarial attack on the powerful BERT language model, using \u201cBERT-Attack\u201d. In this example, one original task of the model was to classify Yelp reviews as \u201cpositive\u201d or \u201cnegative\u201d &#8211; which BERT did with an accuracy of 95.6 percent. After the attack, BERT\u2019s accuracy was only 5.1 percent &#8211; much worse than random. To achieve this, BERT-Attack was able to find the exact words that needed to be replaced with synonyms, to fool the language model.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Not Fun at All<\/h2>\n\n\n\n<p>Among adversarial attacks, physical attacks against computer vision are arguably one of the more immediate threats. This is especially true in areas where AI is used in systems with safety- or security relevance, such as autonomous driving or others:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>What you don\u2019t want is your enemy putting an adversarial image on top of a hospital so that you strike that hospital.<\/p>\n<cite>Jeff Clune, University of Wyoming, on <a href=\"https:\/\/www.theverge.com\/2017\/4\/12\/15271874\/ai-adversarial-images-fooling-attacks-artificial-intelligence\">The Verge<\/a><\/cite><\/blockquote>\n\n\n\n<p>Purely digital attacks might not have the same immediate impact, but they also have the potential to cause significant trouble as well. There are many more domains that I didn\u2019t touch on, such as cybersecurity, sentiment analysis, movement prediction etc., and there are some quite uncomfortable scenarios imaginable if one is only creative enough. Adversarial attacks are clearly more than just a theoretical threat. While the exploration of adversarial attack mechanisms has been a rather academic exercise in the past, there is now a growing awareness of them and their potential impact.&nbsp;<\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity\"\/>\n\n\n\n<h1 class=\"wp-block-heading\">Threat Model<\/h1>\n\n\n\n<p>Technically, adversarial attacks have been around for almost two decades now, since long before the recent success of deep learning. Early adversarial attacks often involved manual input. For example, a spammer might try to design their spam emails to appear innocuous to spam filters. To that end they would guess \u201cbad words\u201d that are likely to trigger a filter rule and replace them with others. Modern attacks are largely automated and employ machine learning methods themselves. <br>Most adversarial attacks are designed against some classification mechanism, for example machine learning models. Such classifiers receive some input and react in a certain way, e.g. by returning an output class. An attacker tries to change this behavior in some way the mechanism was not designed.<br>While adversarial attacks can target any artificial intelligence and machine learning model, today most research into adversarial attacks concerns (deep) neural networks. <br>There are several factors to consider when modeling an adversarial attack threat:&nbsp; The attacker\u2019s assumed control over the target mechanism, their knowledge about it and finally their goal.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Attack Strategy<\/h2>\n\n\n\n<p>Adversarial attacks come in two categories: Poisoning and Evasion attacks. The key difference lies in the attacker\u2019s influence on the target mechanism:<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Poisoning Attacks<\/h3>\n\n\n\n<p>In a poisoning attack, an attacker has access to the creation process itself. In machine learning models this may happen if an attacker can somehow manipulate the training data. Take our spam example from earlier: An \u201cintelligent\u201d spam filter learns which emails are spam and which are not over time. If an attacker can insert spam emails labeled as \u201cnot spam\u201d during this training process, the spam filter will learn to ignore them and won\u2019t filter such spam emails in the future.<br>Of course, this scenario only works if the attacker has at some point access to the training process. Due to this restriction, poisoning attacks are somewhat less prevalent than evasion attacks.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Evasion Attacks<\/h3>\n\n\n\n<p>Evasion attacks on the other hand assume that the target mechanism is already fully developed and immutable. It only assumes that an attacker can send some input to it.<br>The basic idea is as follows: An attacker causes some minimal change in the input data that they send to the target. This change is either imperceptible to humans, or seems so innocuous that it is simply ignored. The machine learning model on the other hand is affected by this change and produces a different result. Such manipulated inputs are called <em>adversarial examples<\/em>. An Attacker now has two options: Either the adversarial examples simply throw off the AI model enough as to make it unsuitable for its task (untargeted attack), or they force the model to behave in a specific way designed by the attacker (targeted attack).<\/p>\n\n\n\n<p>In the field of deep learning and especially computer vision, evasion attacks are much more prevalent than poisoning attacks, due not in small part to their applicability in the real world. For that reason, this article will focus on evasion attacks.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Adversarial Example Crafting<\/h2>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"290\" data-attachment-id=\"10492\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/adv_sample_crafting\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting.png\" data-orig-size=\"1137,322\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"adv_sample_crafting\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting-1024x290.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting-1024x290.png\" alt=\"\" class=\"wp-image-10492\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting-1024x290.png 1024w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting-300x85.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting-768x217.png 768w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/adv_sample_crafting.png 1137w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\">The basic crafting process for adversarial examples<\/figcaption><\/figure>\n\n\n\n<p> The process of crafting adversarial examples usually requires knowledge about the target\u2019s internal <em>decision boundaries<\/em>. This information is necessary, since an attacker will try to change the inputs in such a way that the target comes to a different decision, e.g. to make it predict another class than it normally would.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Decision Boundaries<\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"544\" height=\"413\" data-attachment-id=\"10490\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/decision_boundaries\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/decision_boundaries.png\" data-orig-size=\"544,413\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"decision_boundaries\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/decision_boundaries.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/decision_boundaries.png\" alt=\"\" class=\"wp-image-10490\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/decision_boundaries.png 544w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/decision_boundaries-300x228.png 300w\" sizes=\"auto, (max-width: 544px) 100vw, 544px\" \/><figcaption class=\"wp-element-caption\">A sample <strong><em>x<\/em><\/strong> with <strong><em>y = 4<\/em><\/strong> is turned into an adversarial sample <strong>x\u2019<\/strong> that is classified as a <strong><em>3<\/em><\/strong>.<br>Source: <a href=\"https:\/\/arxiv.org\/pdf\/1909.08072.pdf\">Adversarial AI Survey Paper, Xu et al., 2019<\/a><\/figcaption><\/figure>\n\n\n\n<p>A neural network classifier learns decision boundaries, that allow it to classify an input sample <strong><em>x<\/em><\/strong> as its corresponding class <strong><em>y<\/em><\/strong>. These decision boundaries are represented by the network\u2019s internal parameters. If an attacker knows where the decision boundaries are, they can change the input example <strong><em>x<\/em><\/strong> in such a way that \u201cpushes\u201d it over the boundary.&nbsp;<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Perturbation and Robustness<\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"719\" height=\"230\" data-attachment-id=\"10489\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/fgsm_panda_image_preds\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds.png\" data-orig-size=\"719,230\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"fgsm_panda_image_preds\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds.png\" alt=\"\" class=\"wp-image-10489\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds.png 719w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/fgsm_panda_image_preds-300x96.png 300w\" sizes=\"auto, (max-width: 719px) 100vw, 719px\" \/><figcaption class=\"wp-element-caption\">Source: <a href=\"https:\/\/arxiv.org\/pdf\/1412.6572.pdf\">Explaining and Harnessing Adversarial Examples, Goodfellow et al., 2015<\/a><\/figcaption><\/figure>\n\n\n\n<p>In some attack methods like FGSM, an adversarial example <strong><em>x\u2019<\/em><\/strong> is created from a regular input example by adding some perturbation <strong><em>\u03b7<\/em><\/strong>. This perturbation is also often called <em>adversarial noise<\/em>. The noise is then added on top of the original input <strong><em>x<\/em><\/strong>. Usually, an attacker tries to make this noise as imperceptible as possible, while still having an adverse effect on the target\u2019s performance.<br>Conversely, a model is more robust against adversarial attacks, the more perturbation is necessary to fool it. If the minimum necessary perturbation is large enough, the attack may become obvious to the naked eye and the attack is easily detected.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Adversarial Patch<\/h3>\n\n\n\n<p>Other attacks, such as most physical attacks, rely on <em>adversarial patches<\/em>. In such cases, the perturbation isn\u2019t hidden but constrained to some patch within the input image. Here the attacker\u2019s goal is to make the patch as small and unassuming as possible, while retaining its adverse effect.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Transferability<\/h3>\n\n\n\n<p>Transferability refers to a property of many adversarial attacks, that allows them to be successful on more than one specific target. This concept can even work across different architectures. Only the domain of the models must be the same.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"880\" height=\"343\" data-attachment-id=\"10493\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/blackbox_concept_01\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01.png\" data-orig-size=\"880,343\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"blackbox_concept_01\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01.png\" alt=\"\" class=\"wp-image-10493\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01.png 880w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01-300x117.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_01-768x299.png 768w\" sizes=\"auto, (max-width: 880px) 100vw, 880px\" \/><figcaption class=\"wp-element-caption\">Transferability example: If an attacker can create a substitute model <strong><em>F<\/em><\/strong> that is close enough to the black-box <strong><em>O<\/em><\/strong>, they can then use <strong><em>F<\/em><\/strong> to create adversarial noise that is effective on <strong><em>O<\/em><\/strong>.<\/figcaption><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">Attacker\u2019s Knowledge<\/h2>\n\n\n\n<p>How successful an attack can be largely depends on the information available to the attacker. One common approach to model such information is to consider black-box, gray-box and white-box attacks:<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">White-box<\/h3>\n\n\n\n<p>The attacker knows everything about the target model. They have access to the training data, the neural network graph and the trained parameters. The attacker can essentially use the model against itself.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Gray-box<\/h3>\n\n\n\n<p>The attacker only has access to parts of the model, such as the architecture or the classifier\u2019s confidence values. This information is then used to recreate the target model as close to the original as possible.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Black-box<\/h3>\n\n\n\n<p>The attacker has no knowledge about the target model at all. However, they <strong>can<\/strong> make educated guesses. There are a number of attack strategies, that are each able to extract valuable information from a black-box model. Most of these strategies assume that the attacker correctly guesses the target model&#8217;s domain and that they can communicate with the black-box, e.g. through an API. They can then send queries to the black-box model, to learn how it classifies inputs. The goal is again to acquire enough information to recreate a close approximation of the target model.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"988\" height=\"343\" data-attachment-id=\"10500\" data-permalink=\"https:\/\/blog.mi.hdm-stuttgart.de\/blackbox_concept_02\/\" data-orig-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02.png\" data-orig-size=\"988,343\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"blackbox_concept_02\" data-image-description=\"\" data-image-caption=\"\" data-large-file=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02.png\" src=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02.png\" alt=\"\" class=\"wp-image-10500\" srcset=\"https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02.png 988w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02-300x104.png 300w, https:\/\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/blackbox_concept_02-768x267.png 768w\" sizes=\"auto, (max-width: 988px) 100vw, 988px\" \/><figcaption class=\"wp-element-caption\">Substitute model example:  <br>1. The attacker finds input data that fits the domain of the black-box (<em>oracle<\/em>).<br>2. The attacker sends the input data to the oracle, which returns predictions.<br>3. The attacker uses the input data and the labels predicted by the oracle to train a substitute model.<\/figcaption><\/figure>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity\"\/>\n\n\n\n<h1 class=\"wp-block-heading\">Defending Against Adversarial AI<\/h1>\n\n\n\n<p>So what can you do about adversarial attacks? Can they be averted or mitigated? Well, yes and no. Here are some security practices I found &#8211; you\u2019ll likely recognize some of the more general ones. Let\u2019s explore what works, what doesn\u2019t, and why.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Gradient Masking<\/h2>\n\n\n\n<p>Since many white-box attacks require access to the gradients of a neural network, one way to prevent them is to simply hide their parameters. In practice this boils down to: Don\u2019t put your trained deep neural network up on GitHub for everyone to see, if you want to use it in production. This is what many autonomous car manufacturers do, they keep their network graphs and parameters closed source, which in theory makes it harder to attack them. If that sounds like \u201csecurity through obscurity\u201d to you: That\u2019s because it is.&nbsp;<br>The problem with this \u201cdefense\u201d is that many attacks are transferable. An attack on one machine learning model may not work perfectly on another, but just well enough. In addition, there are black-box attacks out there that don\u2019t require full access to the original network parameters at all.<\/p>\n\n\n\n<p>Verdict: Gradient Masking is better than nothing, but it doesn\u2019t really work as a defense strategy on its own, other than giving you a false sense of security.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Rate Limiting<\/h2>\n\n\n\n<p>Say you employ gradient masking and put your neural network behind an API. A user sends some input (e.g. an image), the API returns some output (e.g. a classification). An attacker can only observe the black-box from outside. Without rate limiting, you\u2019re now inviting a substitute model black-box attack. An attacker can send as many queries against the API as they like, until they have labeled a complete dataset. They then train their own substitute model that has the same internal decision boundaries as the original, and attack from there. <br>Rate limiting can prevent some substitute model attacks, because sending millions of queries would simply take forever. Depending on your use case, you can even use CAPTCHAs or charge money for every API use.<br>On the flip side, this still doesn\u2019t protect against some of the more advanced black-box attacks. Some of these are especially designed to minimize the number of queries necessary to create a good substitute.<\/p>\n\n\n\n<p>Verdict: It\u2019s a good idea to use rate limiting, but it doesn\u2019t offer consistent protection either and may significantly impact user experience.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Input Validation<\/h2>\n\n\n\n<p>Given that most recent adversarial attack methods use machine learning techniques, it only makes sense to use machine learning techniques to defend against them. To neutralize adversarial examples before they reach their target, one can use a secondary <em>detector model<\/em>. Such a model is either trained to flag and filter adversarial examples, or to turn them into regular, harmless examples. The latter can be done by training a <em>denoising autoencoder<\/em> to remove any adversarial noise from its input examples and pass only the \u201ccleaned\u201d inputs to the primary model. The advantage of input validation is that the actual model (the one with the business logic) can stay as-is. The downside here is that you\u2019ll have to train and maintain yet another model. The second disadvantage is the increase in latency, because every input goes through two neural networks now. Finally, there\u2019s no guarantee that a detector itself cannot be fooled.<\/p>\n\n\n\n<p>Verdict: Input validation with a detector model can actually be a reasonable protection against known adversarial attacks, but it comes with high recurring costs.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Adversarial Training<\/h2>\n\n\n\n<p>Finally, one can try and make the model itself more robust to attacks. Remember that the main problem with adversarial examples is that the target model has never seen that specific type of noise during training. Adversarial training means that adversarial examples are added to the training data, so that some specific attacks become useless against the model. Unfortunately, this can make the model larger and more complex, as it must learn more information. This drives up training cost for no immediate performance gain. Finally, adversarial training might theoretically open the door for poisoning attacks. If detected adversarial examples are used in training, an attacker might thus manipulate the training process itself.<\/p>\n\n\n\n<p>Verdict: This is a valid defense method, but it comes with high recurring costs, because it must be done again once new attacks are discovered. The space of possible adversarial attacks is very large, so it is intractable to defend against all of them. This means that adversarial training is only a crutch, but cannot fix the deeper problem.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">The Deeper Problem<\/h2>\n\n\n\n<p>The issue with the \u201csolutions\u201d above is that they are all only symptomatic fixes. The deeper problem lies within a general limitation of state-of-the-art machine learning, that has been documented as early as 2014:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>These results suggest that classifiers based on modern machine learning techniques, even those that obtain excellent performance on the test set, are not learning the true underlying concepts that determine the correct output label. Instead, these algorithms have built a Potemkin village that works well on naturally occuring data, but is exposed as a fake when one visits points in space that do not have high probability in the data distribution. [\u2026]<br>We regard the knowledge of this flaw as an opportunity to fix it.<\/p>\n<cite>Goodfellow et al., <a href=\"http:\/\/arxiv.org\/pdf\/1412.6572.pdf\">Explaining and Harnessing Adversarial Examples<\/a><\/cite><\/blockquote>\n\n\n\n<p>In other words: Current machine learning models do not really understand the concept of what they see. All they learn is some shallow set of rules that happens to result in the expected behavior. As soon as some pattern appears which they haven\u2019t seen during training, their flaw will be revealed.<br>This means in practice that there is never a guarantee against adversarial attacks, because there are simply too many ways to create such patterns. Machine learning models cannot be made proactively robust against every possible attack, because that would require them to understand the true concept behind the data. That leaves us in a situation where:<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>The current state-of-the-art attacks will likely be neutralized by new defenses, and these defenses will subsequently be circumvented.<\/p>\n<cite>Xu et al., Michigan State University, <a href=\"http:\/\/arxiv.org\/pdf\/1909.08072.pdf\">2019 Survey Paper<\/a><\/cite><\/blockquote>\n\n\n\n<p>Some even call it an <a href=\"https:\/\/www.mcafee.com\/blogs\/mcafee-labs\/2018-threats-predictions#machinelearning\">adversarial AI arms race<\/a>. A grim outlook? Perhaps, but it is at least something we can work with.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">General Strategy<\/h2>\n\n\n\n<p><a href=\"https:\/\/www.accenture.com\/_acnmedia\/Accenture\/Redesign-Assets\/DotCom\/Documents\/Global\/1\/Accenture-Trustworthy-AI-POV-Updated.pdf\">A recent report published by Accenture Labs<\/a> features a three-step plan on how to prepare for adversarial threats. I\u2019ve compiled the items that I found most actionable, as well as a few of my own thoughts:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Take stock of everything in your organization that uses AI models.<\/li>\n\n\n\n<li>Think about which AI components are exposed to third parties (and who they are).<\/li>\n\n\n\n<li>Have some process to assess the risks of your active models regularly.<\/li>\n\n\n\n<li>Keep your models current, define a model lifecycle and replacement strategy.<\/li>\n\n\n\n<li>Rank and prioritize each model by its associated risk.<\/li>\n<\/ul>\n\n\n\n<p>There is not yet a dedicated security process (that I know of) for adversarial attacks. That said, many existing security principles and management processes should readily apply to adversarial risks as well.&nbsp;<\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity\"\/>\n\n\n\n<h1 class=\"wp-block-heading\">Conclusion<\/h1>\n\n\n\n<p>AI and deep learning models expose a new attack surface that has appeared alongside the deep learning boom of recent years. New papers on attacks and defenses keep appearing constantly, some even call it an arms race.&nbsp;<br>So far, we know adversarial attacks mostly from controlled academic settings. That said, the possibility of actual large scale attacks is certainly a reality, since many productive systems have been shown to be prone to such attacks.&nbsp;<br>There is not (yet) a panacea against adversarial attacks, because they exploit the limitations of machine learning as it exists today. However, there are ways to protect against adversarial threats, if one knows which systems are vulnerable and how to strengthen them.&nbsp;<\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity\"\/>\n\n\n\n<h1 class=\"wp-block-heading\">Further Reading<\/h1>\n\n\n\n<h2 class=\"wp-block-heading\">Sources<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a href=\"https:\/\/www.accenture.com\/us-en\/insights\/artificial-intelligence\/adversarial-ai\">Accenture Labs AI Security Report Overview&nbsp;<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/www.accenture.com\/_acnmedia\/Accenture\/Redesign-Assets\/DotCom\/Documents\/Global\/1\/Accenture-Trustworthy-AI-POV-Updated.pdf\">Accenture Labs AI Security Full Report<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/arxiv.org\/pdf\/1909.08072.pdf\">Adversarial AI Survey Paper, Xu et al., 2019<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/arxiv.org\/pdf\/1712.09665.pdf\">Adversarial Patch, Brown et al., 2018<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/openai.com\/blog\/adversarial-example-research\/\">Attacking Machine Learning with Adversarial Examples (OpenAI blog)<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/arxiv.org\/pdf\/2004.09984v1.pdf\">BERT-ATTACK: Adversarial Attack Against BERT Using BERT, Li et al., 2020<\/a><\/li>\n\n\n\n<li><a href=\"http:\/\/www.cleverhans.io\">Cleverhans AI Security Blog (OpenAI)<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/keenlab.tencent.com\/en\/whitepapers\/Experimental_Security_Research_of_Tesla_Autopilot.pdf\">Experimental Security Research of Tesla Autopilot, Tencent Keen Security Lab, 2019<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/arxiv.org\/pdf\/1412.6572.pdf\">Explaining and Harnessing Adversarial Examples, Goodfellow et al., 2015<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/www.theverge.com\/2017\/4\/12\/15271874\/ai-adversarial-images-fooling-attacks-artificial-intelligence\">Magic AI: these are the optical illusions that trick, fool, and flummox computers (The Verge)<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/www.mcafee.com\/blogs\/mcafee-labs\/2018-threats-predictions#machinelearning\">McAfee Labs 2018 Threats Predictions<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/arxiv.org\/pdf\/1707.08945.pdf\">Robust Physical-World Attacks on Deep Learning Visual Classification, Eykholt et al., 2018<\/a><\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">Code Repositories on Adversarial AI and Defenses<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a href=\"https:\/\/github.com\/tensorflow\/cleverhans\">Cleverhans Repository<\/a><\/li>\n\n\n\n<li><a href=\"https:\/\/github.com\/IBM\/adversarial-robustness-toolbox\">IBM Adversarial Robustness Toolbox<\/a><\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">More Interesting Reads<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li>A great German article on physical adversarial attacks against autonomous vehicles: <a href=\"https:\/\/www.heise.de\/hintergrund\/Pixelmuster-irritieren-die-KI-autonomer-Fahrzeuge-4852995.html\">Pixelmuster irritieren die KI autonomer Fahrzeuge<\/a><\/li>\n<\/ul>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Find out how AI may become an attack vector! Could an attacker use your models against your? Also, what\u2019s the worst that could happen? Welcome to the domain of adversarial AI!<\/p>\n","protected":false},"author":998,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[1,652,664,26],"tags":[357,356,358,355,106,119,359,57,58],"ppma_author":[814],"class_list":["post-10442","post","type-post","status-publish","format-standard","hentry","category-allgemein","category-artificial-intelligence","category-ethics","category-secure-systems","tag-adversarial-ai","tag-adversarial-attacks","tag-adversarial-examples","tag-ai","tag-artificial-intelligence","tag-deep-learning","tag-evasion-attacks","tag-machine-learning","tag-secure-systems"],"aioseo_notices":[],"jetpack_featured_media_url":"","jetpack-related-posts":[{"id":7327,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2019\/08\/30\/about-the-robustness-of-machine-learning\/","url_meta":{"origin":10442,"position":0},"title":"About the Robustness of Machine Learning","author":"Marcel Heisler","date":"30. August 2019","format":false,"excerpt":"In the past couple of years research in the field of machine learning (ML) has made huge progress which resulted in applications like automated translation, practical speech recognition for smart assistants, useful robots, self-driving cars and lots of others. But so far we only have reached the point where ML\u2026","rel":"","context":"In &quot;Secure Systems&quot;","block_context":{"text":"Secure Systems","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/system-designs\/secure-systems\/"},"img":{"alt_text":"Glitch","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=700%2C400&ssl=1 2x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=1050%2C600&ssl=1 3x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2019\/08\/glitch-2463363_1920.jpg?resize=1400%2C800&ssl=1 4x"},"classes":[]},{"id":2615,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2017\/08\/27\/fooling-the-intelligence\/","url_meta":{"origin":10442,"position":1},"title":"FOOLING THE INTELLIGENCE","author":"Jonas Miederer","date":"27. August 2017","format":false,"excerpt":"Adversarial machine learning and its dangers The world is led by machines, humans are subjected to the robot\u2019s rule. Omniscient computer systems hold the control of the world. The newest technology has outpaced human knowledge, while the mankind is powerless in the face of the stronger, faster, better and almighty\u2026","rel":"","context":"In &quot;Allgemein&quot;","block_context":{"text":"Allgemein","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/allgemein\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2017\/08\/AAEAAQAAAAAAAAxmAAAAJDcyNzkyZjkzLTUzZTEtNGU1ZS04OWYxLWU4NDU5Y2QxOTRjYQ.png?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2017\/08\/AAEAAQAAAAAAAAxmAAAAJDcyNzkyZjkzLTUzZTEtNGU1ZS04OWYxLWU4NDU5Y2QxOTRjYQ.png?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2017\/08\/AAEAAQAAAAAAAAxmAAAAJDcyNzkyZjkzLTUzZTEtNGU1ZS04OWYxLWU4NDU5Y2QxOTRjYQ.png?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2017\/08\/AAEAAQAAAAAAAAxmAAAAJDcyNzkyZjkzLTUzZTEtNGU1ZS04OWYxLWU4NDU5Y2QxOTRjYQ.png?resize=700%2C400&ssl=1 2x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2017\/08\/AAEAAQAAAAAAAAxmAAAAJDcyNzkyZjkzLTUzZTEtNGU1ZS04OWYxLWU4NDU5Y2QxOTRjYQ.png?resize=1050%2C600&ssl=1 3x"},"classes":[]},{"id":10415,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2020\/08\/19\/ai-cyberattacks-deepfakes\/","url_meta":{"origin":10442,"position":2},"title":"The Dark Side of AI &#8211; Part 1: Cyberattacks and Deepfakes","author":"Micha Christ","date":"19. August 2020","format":false,"excerpt":"Introduction Who hasn't seen a cinema production in which an AI-based robot threatens individual people or the entire human race? It is in the stars when or if such a technology can really be developed. With this series of blog entries we want to point out that AI does not\u2026","rel":"","context":"In &quot;Allgemein&quot;","block_context":{"text":"Allgemein","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/allgemein\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/screen-shot-2018-08-03-at-10-34-32-1.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/screen-shot-2018-08-03-at-10-34-32-1.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2020\/08\/screen-shot-2018-08-03-at-10-34-32-1.jpg?resize=525%2C300&ssl=1 1.5x"},"classes":[]},{"id":25006,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2023\/07\/24\/sicherheitscheck-wie-sicher-sind-deep-learning-systeme\/","url_meta":{"origin":10442,"position":3},"title":"Sicherheitscheck \u2013 Wie sicher sind Deep Learning Systeme?","author":"Silas Blumenstock","date":"24. July 2023","format":false,"excerpt":"Was ist Deep Learning?Angriffe auf Deep Learning Systeme\u00dcbersicht von AngriffenVor\/W\u00e4hrend des TrainingsPoisoning AttacksModel Extraction AttacksModel Inversion AttacksNach dem Training (Inferenzphase)Adversarial AttacksGegenma\u00dfnahmenAusblickLiteratur In einer immer st\u00e4rker digitalisierten Welt haben Neuronale Netze und Deep Learning eine immer wichtigere Rolle eingenommen und viele Bereiche unseres Alltags in vielerlei Hinsicht bereichert. Von Sprachmodellen \u00fcber\u2026","rel":"","context":"In &quot;Allgemein&quot;","block_context":{"text":"Allgemein","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/allgemein\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/ga_banner_EA15O715KD.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/ga_banner_EA15O715KD.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/ga_banner_EA15O715KD.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/ga_banner_EA15O715KD.jpg?resize=700%2C400&ssl=1 2x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/ga_banner_EA15O715KD.jpg?resize=1050%2C600&ssl=1 3x"},"classes":[]},{"id":24243,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2023\/03\/03\/modern-application-of-voice-ai-technology\/","url_meta":{"origin":10442,"position":4},"title":"Modern application of Voice AI technology","author":"Ngoc Ton","date":"3. March 2023","format":false,"excerpt":"With the advancement of technology and the gradually increasing use of artificial intelligence, new markets are developed. One of such is the market of Voice AI which became a commercial success with voice bots such as Alexa or Siri. They were mainly used as digital assistants who could answer questions,\u2026","rel":"","context":"In &quot;Allgemein&quot;","block_context":{"text":"Allgemein","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/allgemein\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/03\/01.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/03\/01.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/03\/01.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/03\/01.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]},{"id":25062,"url":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/2023\/07\/27\/machine-learning-fluch-oder-segen-fur-die-it-security\/","url_meta":{"origin":10442,"position":5},"title":"Machine Learning: Fluch oder Segen f\u00fcr die IT Security?","author":"Sabrina Berg","date":"27. July 2023","format":false,"excerpt":"Im heutigen digitalen Zeitalter ist die Sicherheit von IT-Systemen ein allgegenw\u00e4rtiges Thema von enormer Wichtigkeit. Rund um die Uhr m\u00fcssen riesige Mengen an sensible Daten sicher gespeichert und \u00fcbertragen werden k\u00f6nnen und die Funktionalit\u00e4t von unz\u00e4hligen Systemen muss zuverl\u00e4ssig aufrecht gehalten werden. Industrie 4.0, unz\u00e4hlige Onlinediensten und das Internet of\u2026","rel":"","context":"In &quot;Allgemein&quot;","block_context":{"text":"Allgemein","link":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/category\/allgemein\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/Machine-learning-1024x587-1-e1690452250291.png?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/Machine-learning-1024x587-1-e1690452250291.png?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/Machine-learning-1024x587-1-e1690452250291.png?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/blog.mi.hdm-stuttgart.de\/wp-content\/uploads\/2023\/07\/Machine-learning-1024x587-1-e1690452250291.png?resize=700%2C400&ssl=1 2x"},"classes":[]}],"jetpack_sharing_enabled":true,"authors":[{"term_id":814,"user_id":998,"is_guest":0,"slug":"fw055","display_name":"Florian Wintel","avatar_url":"https:\/\/secure.gravatar.com\/avatar\/f8d34dc1797988220eb614c256d1afeb61b7883695f224b57ce3992705443c6d?s=96&d=mm&r=g","0":null,"1":"","2":"","3":"","4":"","5":"","6":"","7":"","8":""}],"_links":{"self":[{"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/posts\/10442","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/users\/998"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/comments?post=10442"}],"version-history":[{"count":90,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/posts\/10442\/revisions"}],"predecessor-version":[{"id":25435,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/posts\/10442\/revisions\/25435"}],"wp:attachment":[{"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/media?parent=10442"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/categories?post=10442"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/tags?post=10442"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/blog.mi.hdm-stuttgart.de\/index.php\/wp-json\/wp\/v2\/ppma_author?post=10442"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}