LDR | | 03024nmm uu200421 4500 |
001 | | 000000334306 |
005 | | 20240805180036 |
008 | | 181129s2018 |||||||||||||||||c||eng d |
020 | |
▼a 9780438135536 |
035 | |
▼a (MiAaPQ)AAI10903728 |
040 | |
▼a MiAaPQ
▼c MiAaPQ
▼d 248032 |
082 | 0 |
▼a 004 |
100 | 1 |
▼a Papernot, Nicolas. |
245 | 10 |
▼a Characterizing the Limits and Defenses of Machine Learning in Adversarial Settings. |
260 | |
▼a [S.l.] :
▼b The Pennsylvania State University.,
▼c 2018 |
260 | 1 |
▼a Ann Arbor :
▼b ProQuest Dissertations & Theses,
▼c 2018 |
300 | |
▼a 178 p. |
500 | |
▼a Source: Dissertation Abstracts International, Volume: 79-12(E), Section: B. |
502 | 1 |
▼a Thesis (Ph.D.)--The Pennsylvania State University, 2018. |
520 | |
▼a Advances in machine learning (ML) in recent years have enabled a dizzying array of applications such as object recognition, autonomous systems, security diagnostics, and playing the game of Go. Machine learning is not only a new paradigm for bui |
520 | |
▼a In this thesis, I focus my study on the integrity of ML models. Integrity refers here to the faithfulness of model predictions with respect to an expected outcome. This property is at the core of traditional machine learning evaluation, as demon |
520 | |
▼a A large fraction of ML techniques were designed for benign execution environments. Yet, the presence of adversaries may invalidate some of these underlying assumptions by forcing a mismatch between the distributions on which the model is trained |
520 | |
▼a I explore the space of attacks against ML integrity at test time. Given full or limited access to a trained model, I devise strategies that modify the test data to create a worst-case drift between the training and test distributions. The implic |
520 | |
▼a Hence, my efforts to increase the robustness of models to these adversarial manipulations strive to decrease the confidence of predictions made far from the training distribution. Informed by my progress on attacks operating in the black-box thr |
520 | |
▼a I then describe recent defensive efforts addressing these shortcomings. To this end, I introduce the Deep k-Nearest Neighbors classifier, which augments deep neural networks with an integrity check at test time. The approach compares internal re |
520 | |
▼a This research calls for future efforts to investigate the robustness of individual layers of deep neural networks rather than treating the model as a black-box. This aligns well with the modular nature of deep neural networks, which orchestrate |
590 | |
▼a School code: 0176. |
650 | 4 |
▼a Computer science. |
690 | |
▼a 0984 |
710 | 20 |
▼a The Pennsylvania State University.
▼b Computer Science and Engineering. |
773 | 0 |
▼t Dissertation Abstracts International
▼g 79-12B(E). |
773 | |
▼t Dissertation Abstract International |
790 | |
▼a 0176 |
791 | |
▼a Ph.D. |
792 | |
▼a 2018 |
793 | |
▼a English |
856 | 40 |
▼u http://www.riss.kr/pdu/ddodLink.do?id=T15000703
▼n KERIS |
980 | |
▼a 201812
▼f 2019 |
990 | |
▼a 관리자 |