Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluating Concept Explanations for CNNs Under Adversarial Image Transformations
0
Zitationen
6
Autoren
2024
Jahr
Abstract
Concept-based explainers for convolutional neural networks (CNNs) provide human-understandable explanations by revealing what the CNN sees, rather than merely indicating where it looked. However, their performance is limited by the reducer at its core and adversarial attacks. Although CNN classification performance may be enhanced by some image transformations in small amounts whereas intense image transformations can cause noticeable variations to CNN predictions, it is uncertain how explainers perform in such cases. This paper investigates the performance of state-of-the-art concept-based explainers at different levels of adversarial attacks for the first time. We achieve this by exploring different image transformations as adversarial attacks, including Gaussian noise, elastic transform, rotation, and contrast on the ILSVRC2012 dataset. Our study shows that image transformation techniques altering only image coordinates have little impact on classifier and explainer performance, whereas methods modifying image pixels, such as elastic transform and contrast, significantly affect performance, akin to introducing Gaussian noise. Our work underscores the significance of scrutinizing explainers during their development and adoption for CNNs.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.597 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.822 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.710 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.455 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.633 Zit.