Modern deep neural networks have now reached human-level performance across a variety of tasks. However, unlike humans they lack the ability to explain their decisions by showing where and telling what concepts guided them. In this work, we present a unified framework for transforming any vision neural network into a spatially and conceptually interpretable model. We introduce a spatially-aware concept bottleneck layer that projects “black-box” features of pre-trained backbone models into interpretable concept maps, without requiring human labels. By training a classification layer over this bottleneck, we obtain a self-explaining model that articulates which concepts most influenced its prediction, along with heatmaps that ground them in the input image. Accordingly, we name this method “Spatially-Aware and Label-Free Concept Bottleneck Model” (SALF-CBM). Our results show that the proposed SALF-CBM: (1) Outperforms non-spatial CBM methods, as well as the original backbone, on a variety of classification tasks; (2) Produces high-quality spatial explanations, outperforming widely used heatmap-based methods on a zero-shot segmentation task; (3) Facilitates model exploration and debugging, enabling users to query specific image regions and refine the model's decisions by locally editing its concept maps.
Given a pre-trained backbone model, we transform it into an explainable SALF-CBM as follows:
SALF-CBM explains its predictions as an integral part of its inference process by specifying which concepts contributed to its output the most, and grounding them in the input image. Below, we show qualitative examples with ResNet-50 backbone pre-trained on ImageNet.
When prompted with a specific image region, SALF-CBM reveals how the model perceives it by identifying the most strongly activated concepts within that region. Below, we show qualitative examples with ResNet-50 backbone pre-trained on ImageNet.
By applying our method to video sequences in a frame-by-frame manner, we can visualize how the model recognizes concepts over time. Below, we show qualitative results on DAVIS 2017 videos with a ResNet-50 backbone pre-trained on ImageNet.
If you find this project useful for your research, please cite the following:
@article{benou2025show,
title={Show and Tell: Visually Explainable Deep Neural Nets via Spatially-Aware Concept Bottleneck Models},
author={Benou, Itay and Riklin-Raviv, Tammy},
journal={arXiv preprint arXiv:2502.20134},
year={2025}
}