AGNES: Abstraction-Guided Framework for Deep Neural Networks Security

Akshay Dhonthi Ramesh Babu*, Marcello Eiermann, Ernst Moritz Hahn, Vahid Hashemi

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

16 Downloads (Pure)

Abstract

Deep Neural Networks (DNNs) are becoming widespread, particularly in safety-critical areas. One prominent application is image recognition in autonomous driving, where the correct classification of objects, such as traffic signs, is essential for safe driving. Unfortunately, DNNs are prone to backdoors, meaning that they concentrate on attributes of the image that should be irrelevant for their correct classification. Backdoors are integrated into a DNN during training, either with malicious intent (such as a manipulated training process, because of which a yellow sticker always leads to a traffic sign being recognised as a stop sign) or unintentional (such as a rural background leading to any traffic sign being recognised as “animal crossing”, because of biased training data). In this paper, we introduce AGNES, a tool to detect backdoors in DNNs for image recognition. We discuss the principle approach on which AGNES is based. Afterwards, we show that our tool performs better than many state-of-the-art methods for multiple relevant case studies.

Original languageEnglish
Title of host publicationVerification, Model Checking, and Abstract Interpretation - 25th International Conference, VMCAI 2024, Proceedings
EditorsRayna Dimitrova, Ori Lahav, Sebastian Wolff
PublisherSpringer
Pages124-138
Number of pages15
ISBN (Print)9783031505201
DOIs
Publication statusPublished - 2024
Event25th International Conference on Verification, Model Checking, and Abstract Interpretation, VMCAI 2024 - London, United Kingdom
Duration: 15 Jan 202416 Jan 2024
Conference number: 25

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14500 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference25th International Conference on Verification, Model Checking, and Abstract Interpretation, VMCAI 2024
Abbreviated titleVMCAI 2024
Country/TerritoryUnited Kingdom
CityLondon
Period15/01/2416/01/24
Otherco-located with 51st ACM SIGPLAN Symposium on Principles of Programming Languages, POPL 2024

Keywords

  • 2024 OA procedure
  • Neural network analysis
  • Security testing
  • Backdoor detection

Fingerprint

Dive into the research topics of 'AGNES: Abstraction-Guided Framework for Deep Neural Networks Security'. Together they form a unique fingerprint.

Cite this