Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Invertible Concept-based Explanations for CNN Models with Non-negative Concept Activation Vectors (2006.15417v4)

Published 27 Jun 2020 in cs.CV, cs.AI, and cs.LG

Abstract: Convolutional neural network (CNN) models for computer vision are powerful but lack explainability in their most basic form. This deficiency remains a key challenge when applying CNNs in important domains. Recent work on explanations through feature importance of approximate linear models has moved from input-level features (pixels or segments) to features from mid-layer feature maps in the form of concept activation vectors (CAVs). CAVs contain concept-level information and could be learned via clustering. In this work, we rethink the ACE algorithm of Ghorbani et~al., proposing an alternative invertible concept-based explanation (ICE) framework to overcome its shortcomings. Based on the requirements of fidelity (approximate models to target models) and interpretability (being meaningful to people), we design measurements and evaluate a range of matrix factorization methods with our framework. We find that non-negative concept activation vectors (NCAVs) from non-negative matrix factorization provide superior performance in interpretability and fidelity based on computational and human subject experiments. Our framework provides both local and global concept-level explanations for pre-trained CNN models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ruihan Zhang (26 papers)
  2. Prashan Madumal (7 papers)
  3. Tim Miller (53 papers)
  4. Benjamin I. P. Rubinstein (69 papers)
  5. Krista A. Ehinger (14 papers)
Citations (77)

Summary

We haven't generated a summary for this paper yet.