skip to main content
10.1145/3097983.3098158acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Adversary Resistant Deep Neural Networks with an Application to Malware Detection

Authors Info & Claims
Published:13 August 2017Publication History

ABSTRACT

Outside the highly publicized victories in the game of Go, there have been numerous successful applications of deep learning in the fields of information retrieval, computer vision, and speech recognition. In cybersecurity, an increasing number of companies have begun exploring the use of deep learning (DL) in a variety of security tasks with malware detection among the more popular. These companies claim that deep neural networks (DNNs) could help turn the tide in the war against malware infection. However, DNNs are vulnerable to adversarial samples, a shortcoming that plagues most, if not all, statistical and machine learning models. Recent research has demonstrated that those with malicious intent can easily circumvent deep learning-powered malware detection by exploiting this weakness.

To address this problem, previous work developed defense mechanisms that are based on augmenting training data or enhancing model complexity. However, after analyzing DNN susceptibility to adversarial samples, we discover that the current defense mechanisms are limited and, more importantly, cannot provide theoretical guarantees of robustness against adversarial sampled-based attacks. As such, we propose a new adversary resistant technique that obstructs attackers from constructing impactful adversarial samples by randomly nullifying features within data vectors. Our proposed technique is evaluated on a real world dataset with 14,679 malware variants and 17,399 benign programs. We theoretically validate the robustness of our technique, and empirically show that our technique significantly boosts DNN robustness to adversarial samples while maintaining high accuracy in classification. To demonstrate the general applicability of our proposed method, we also conduct experiments using the MNIST and CIFAR-10 datasets, widely used in image recognition research.

References

  1. Hyrum Anderson, Jonathan Woodbridge, and Bobby Filar. 2016. DeepDGA: Adversarially-Tuned Domain Generation and Detection. arXiv:1610.01969 [cs.CR] (2016).Google ScholarGoogle Scholar
  2. Matt Wolff Andrew Davis. 2015. Deep Learning on Dis- assembly. https://www.blackhat.com/docs/us-15/materials/ us-15-Davis-Deep-Learning-On-Disassembly.pdf.Google ScholarGoogle Scholar
  3. Marco Barreno, Blaine Nelson, Anthony D. Joseph, and J. D. Tygar. 2010. The Security of Machine Learning. Mach. Learn. 81, 2 (Nov. 2010), 121--148. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Konstantin Berlin, David Slater, and Joshua Saxe. 2015. Malicious behavior detection using windows audit logs. In Proceedings of the 8th ACM Workshop on Artificial Intelligence and Security. ACM, 35--44. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Ran Bi. 2015. Deep Learning can be easily fooled. http://www.kdnuggets.com/ 2015/01/deep-learning-can-be-easily-fooled.html.Google ScholarGoogle Scholar
  6. Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim Srndic, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. 2013. Evasion Attacks against Machine Learning at Test Time.. In ECML/PKDD (3). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. BIZETY 2016. Deep Learning Neural Nets Are Effective Against AI Malware. BIZETY. https://www.bizety.com/2016/02/05/ deep-learning-neural-nets-are-effective-against-ai-malware/.Google ScholarGoogle Scholar
  8. George Dahl, Jack W. Stokes, Li Deng, and Dong Yu. 2013. Large-Scale Malware Classification Using Random Projections and Neural Networks. In Proceedings IEEE Conference on Acoustics, Speech, and Signal Processing. Google ScholarGoogle ScholarCross RefCross Ref
  9. Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572 (2014).Google ScholarGoogle Scholar
  10. Kathrin Grosse, Nicolas Papernot, Praveen Manoharan, Michael Backes, and Patrick McDaniel. 2016. Adversarial Perturbations Against Deep Neural Networks for Malware Classification. arXiv preprint arXiv:1606.04435 (2016).Google ScholarGoogle Scholar
  11. Shixiang Gu and Luca Rigazio. 2014. Towards deep neural network architectures robust to adversarial examples. arXiv:1412.5068 [cs] (2014).Google ScholarGoogle Scholar
  12. Mike James. 2014. The Flaw Lurking In Every Deep Neural Net . http://www.i-programmer.info/news/105-artificial-intelligence/ 7352-the-flaw-lurking-in-every-deep-neural-net.html.Google ScholarGoogle Scholar
  13. D.K. Kang, J. Zhang, A. Silvescu, and V. Honavar. 2005. Multinomial event model based abstraction for sequence and text classification. Abstraction, Reformulation and Approximation (2005), 901--901.Google ScholarGoogle Scholar
  14. Will Knight. 2015. Antivirus that Mimics the Brain Could Catch More Malware. https://www.technologyreview.com/s/542971/ antivirus-that-mimics-the-brain-could-catch-more-malware/.Google ScholarGoogle Scholar
  15. Alex Krizhevsky and Geoffrey Hinton. 2009. Learning multiple layers of features from tiny images. (2009).Google ScholarGoogle Scholar
  16. Yann LeCun, Corinna Cortes, and Christopher JC Burges. 1998. The MNIST database of handwritten digits. (1998).Google ScholarGoogle Scholar
  17. Cade Metz. 2015. Baidu, the Chinese Google, Is Teaching AI to Spot Malware. https://www.wired.com/2015/11/ baidu-the-chinese-google-is-teaching-ai-to-spot-malware/.Google ScholarGoogle Scholar
  18. MIT Technology Review 2016. Machine-Learning Algorithm Combs the Darknet for Zero Day Exploits, and Finds them. MIT Technology Review.Google ScholarGoogle Scholar
  19. Linda Musthaler. 2016. How to use deep learning AI to detect and prevent malware and APTs in real-time.Google ScholarGoogle Scholar
  20. Alexander G. Ororbia II, C. Lee Giles, and Daniel Kifer. 2016. Unifying Adversarial Training Algorithms with Flexible Deep Data Gradient Regularization. arXiv:1601.07213 [cs] (2016).Google ScholarGoogle Scholar
  21. Nicolas Papernot, Patrick McDaniel, Somesh Jha, Matt Fredrikson, Z Berkay Celik, and Ananthram Swami. 2016. The limitations of deep learning in adversarial settings. In 2016 IEEE European Symposium on Security and Privacy (EuroS&P). IEEE, 372--387.Google ScholarGoogle ScholarCross RefCross Ref
  22. Nicolas Papernot, Patrick McDaniel, Xi Wu, Somesh Jha, and Ananthram Swami. 2015. Distillation as a defense to adversarial perturbations against deep neural networks. arXiv preprint arXiv:1511.04508 (2015).Google ScholarGoogle Scholar
  23. Joshua Saxe and Konstantin Berlin. 2015. Deep Neural Network Based Malware Detection Using Two Dimensional Binary Program Features. CoRR (2015).Google ScholarGoogle Scholar
  24. Nitish Srivastava, Geoffrey E Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: a simple way to prevent neural networks from overfitting. Journal of Machine Learning Research 15, 1 (2014), 1929--1958.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Nedim Srndic and Pavel Laskov. 2014. Practical Evasion of a Learning-Based Classifier: A Case Study. In Proceedings of the 2014 IEEE Symposium on Security and Privacy . Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Symantec 2016. Internet Security Threat Report. Symantec. https://www.symantec. com/content/dam/symantec/docs/reports/istr-21--2016-en.pdf.Google ScholarGoogle Scholar
  27. Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2014. Intriguing properties of neural networks. In International Conference on Learning Representations.Google ScholarGoogle Scholar
  28. Zhenlong Yuan, Yongqiang Lu, Zhaoguo Wang, and Yibo Xue. 2014. Droid-Sec: Deep Learning in Android Malware Detection. In Proceedings of the 2014 ACM Conference on SIGCOMM (SIGCOMM '14) Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Adversary Resistant Deep Neural Networks with an Application to Malware Detection

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            KDD '17: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
            August 2017
            2240 pages
            ISBN:9781450348874
            DOI:10.1145/3097983

            Copyright © 2017 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 13 August 2017

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            KDD '17 Paper Acceptance Rate64of748submissions,9%Overall Acceptance Rate1,133of8,635submissions,13%

            Upcoming Conference

            KDD '24

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader