Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
Ergebnis 22 von 54426
IEEE transactions on information forensics and security, 2021, Vol.16, p.4924-4938
2021

Details

Autor(en) / Beteiligte
Titel
Adversarial XAI Methods in Cybersecurity
Ist Teil von
  • IEEE transactions on information forensics and security, 2021, Vol.16, p.4924-4938
Ort / Verlag
IEEE
Erscheinungsjahr
2021
Link zum Volltext
Quelle
IEEEXplore
Beschreibungen/Notizen
  • Machine Learning methods are playing a vital role in combating ever-evolving threats in the cybersecurity domain. Explanation methods that shed light on the decision process of black-box classifiers are one of the biggest drivers in the successful adoption of these models. Explaining predictions that address 'Why?/Why Not?' questions help users/stakeholders/analysts understand and accept the predicted outputs with confidence and build trust. Counterfactual explanations are gaining popularity as an alternative method to help users to not only understand the decisions of black-box models (why?) but also to provide a mechanism to highlight mutually exclusive data instances that would change the outcomes (why not?). Recent Explainable Artificial Intelligence literature has focused on three main areas: (a) creating and improving explainability methods that help users better understand how the internal of ML models work as well as their outputs; (b) attacks on interpreters with a white-box setting; (c) defining the relevant properties, metrics of explanations generated by models. Nevertheless, there is no thorough study of how the model explanations can introduce new attack surfaces to the underlying systems. A motivated adversary can leverage the information provided by explanations to launch membership inference, and model extraction attacks to compromise the overall privacy of the system. Similarly, explanations can also facilitate powerful evasion attacks such as poisoning and back door attacks. In this paper, we cover this gap by tackling various cybersecurity properties and threat models related to counterfactual explanations. We propose a new black-box attack that leverages Explainable Artificial Intelligence (XAI) methods to compromise the confidentiality and privacy properties of underlying classifiers. We validate our approach with datasets and models used in the cyber security domain to demonstrate that our method achieves the attacker's goal under threat models which reflect the real-world settings.
Sprache
Englisch
Identifikatoren
ISSN: 1556-6013
eISSN: 1556-6021
DOI: 10.1109/TIFS.2021.3117075
Titel-ID: cdi_crossref_primary_10_1109_TIFS_2021_3117075

Weiterführende Literatur

Empfehlungen zum selben Thema automatisch vorgeschlagen von bX