Please use this identifier to cite or link to this item: doi:10.22028/D291-34355
Title: Why is Machine Learning Security so hard?
Author(s): Grosse, Kathrin
Language: English
Year of Publication: 2020
DDC notations: 621.3 Electrical engineering, electronics
004 Computer science, internet
Publikation type: Dissertation
Abstract: The increase of available data and computing power has fueled a wide application of machine learning (ML). At the same time, security concerns are raised: ML models were shown to be easily fooled by slight perturbations on their inputs. Furthermore, by querying a model and analyzing output and input pairs, an attacker can infer the training data or replicate the model, thereby harming the owner’s intellectual property. Also, altering the training data can lure the model into producing specific or generally wrong outputs at test time. So far, none of the attacks studied in the field has been satisfactorily defended. In this work, we shed light on these difficulties. We first consider classifier evasion or adversarial examples. The computation of such examples is an inherent problem, as opposed to a bug that can be fixed. We also show that adversarial examples often transfer from one model to another, different model. Afterwards, we point out that the detection of backdoors (a training-time attack) is hindered as natural backdoor-like patterns occur even in benign neural networks. The question whether a pattern is benign or malicious then turns into a question of intention, which is hard to tackle. A different kind of complexity is added with the large libraries nowadays in use to implement machine learning. We introduce an attack that alters the library, thereby decreasing the accuracy a user can achieve. In case the user is aware of the attack, however, it is straightforward to defeat. This is not the case for most classical attacks described above. Additional difficulty is added if several attacks are studied at once: we show that even if the model is configured for one attack to be less effective, another attack might perform even better. We conclude by pointing out the necessity of understanding the ML model under attack. On the one hand, as we have seen throughout the examples given here, understanding precedes defenses and attacks. On the other hand, an attack, even a failed one, often yields new insights and knowledge about the algorithm studied.
Link to this record: urn:nbn:de:bsz:291--ds-343554
hdl:20.500.11880/31543
http://dx.doi.org/10.22028/D291-34355
Advisor: Backes, Michael
Date of oral examination: 20-May-2021
Date of registration: 19-Jul-2021
Third-party funds sponsorship: This work was supported by the German Federal Ministry of Education and Research (BMBF) through funding for the Center for IT-Security,Privacy and Accountability (CISPA) (FKZ: 16KIS0753)
Description of the related object: PhD thesis of Kathrin Grosse
Faculty: MI - Fakultät für Mathematik und Informatik
Department: MI - Informatik
Professorship: MI - Prof. Dr. Michael Backes
Collections:SciDok - Der Wissenschaftsserver der Universität des Saarlandes

Files for this record:
File Description SizeFormat 
grosse_thesis.pdf4,39 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons