Selective Poisoning Attack on Deep Neural Networks †
AbstractStudies related to pattern recognition and visualization using computer technology have been introduced. In particular, deep neural networks (DNNs) provide good performance for image, speech, and pattern recognition. However, a poisoning attack is a serious threat to a DNN’s security. A poisoning attack reduces the accuracy of a DNN by adding malicious training data during the training process. In some situations, it may be necessary to drop a specifically chosen class of accuracy from the model. For example, if an attacker specifically disallows nuclear facilities to be selectively recognized, it may be necessary to intentionally prevent unmanned aerial vehicles from correctly recognizing nuclear-related facilities. In this paper, we propose a selective poisoning attack that reduces the accuracy of only the chosen class in the model. The proposed method achieves this by training malicious data corresponding to only the chosen class while maintaining the accuracy of the remaining classes. For the experiment, we used tensorflow as the machine-learning library as well as MNIST, Fashion-MNIST, and CIFAR10 as the datasets. Experimental results show that the proposed method can reduce the accuracy of the chosen class by 43.2%, 41.7%, and 55.3% in MNIST, Fashion-MNIST, and CIFAR10, respectively, while maintaining the accuracy of the remaining classes. View Full-Text
Share & Cite This Article
Kwon, H.; Yoon, H.; Park, K.-W. Selective Poisoning Attack on Deep Neural Networks †. Symmetry 2019, 11, 892.
Kwon H, Yoon H, Park K-W. Selective Poisoning Attack on Deep Neural Networks †. Symmetry. 2019; 11(7):892.Chicago/Turabian Style
Kwon, Hyun; Yoon, Hyunsoo; Park, Ki-Woong. 2019. "Selective Poisoning Attack on Deep Neural Networks †." Symmetry 11, no. 7: 892.
Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.