Emergent Mind

Dimensions underlying the representational alignment of deep neural networks with humans

(2406.19087)
Published Jun 27, 2024 in cs.CV , cs.AI , cs.LG , and q-bio.QM

Abstract

Determining the similarities and differences between humans and artificial intelligence is an important goal both in machine learning and cognitive neuroscience. However, similarities in representations only inform us about the degree of alignment, not the factors that determine it. Drawing upon recent developments in cognitive science, we propose a generic framework for yielding comparable representations in humans and deep neural networks (DNN). Applying this framework to humans and a DNN model of natural images revealed a low-dimensional DNN embedding of both visual and semantic dimensions. In contrast to humans, DNNs exhibited a clear dominance of visual over semantic features, indicating divergent strategies for representing images. While in-silico experiments showed seemingly-consistent interpretability of DNN dimensions, a direct comparison between human and DNN representations revealed substantial differences in how they process images. By making representations directly comparable, our results reveal important challenges for representational alignment, offering a means for improving their comparability.

Comparison of human- and DNN-derived representational embeddings with semantic, visual-semantic, and visual dimensions.

Overview

  • The paper presents a framework for comparing representational embeddings of humans and Deep Neural Networks (DNNs) using the same experimental paradigms, focusing on visual and semantic dimensions.

  • Through a series of experiments including similarity judgments for object images and various interpretability techniques, it was found that human embeddings are primarily semantic, while DNN embeddings show a strong visual bias.

  • The findings highlight the need for AI models that better balance semantic and visual information to achieve more human-like cognitive capabilities, suggesting future research directions to bridge this gap.

Dimensions Underlying the Representational Alignment of Deep Neural Networks with Humans

The paper "Dimensions underlying the representational alignment of deep neural networks with humans" by Florian P. Mahner et al. provides a systematic framework for comparing the representational embeddings of humans and Deep Neural Networks (DNNs) using the same experimental and modeling paradigms. This study explore the often-discussed topic of representational alignment and contributes a detailed examination of the visual and semantic dimensions that underlie image representations in both domains.

Overview of Methodology

To accomplish their objective, the authors collected similarity judgments for object images from both humans and a pretrained VGG-16 model. Human data was derived from the THINGS database, while the DNN data was synthetically generated using 24,102 images for the same objects. The paper employs a triplet odd-one-out task, where subjects (both human and DNN) select the most dissimilar object from a set of three, to create low-dimensional embeddings that can be directly compared.

Key Findings

  1. Representational Embeddings:

    • Both human and DNN embeddings were optimized to predict the odd-one-out choices, yielding 70 stable dimensions for DNNs and 68 for humans.
    • Human embeddings captured 90.85% of the explainable variance, whereas VGG-16 embeddings captured 84.03%.
  2. Visual versus Semantic Bias:

    • The DNN showed a striking visual bias with many dimensions reflecting visual-perceptual properties.
    • Human embeddings were primarily dominated by semantic dimensions.
    • This contrast was quantitatively verified by expert ratings, which showed a higher proportion of semantic dimensions in human embeddings compared to DNNs.
  3. Interpretability and Causality:

    • Various interpretability techniques such as Grad-CAM, activation maximization with StyleGAN, and causal image manipulations were used to identify and verify the contributory image features.
    • Grad-CAM heatmaps and dimension-specific synthetic images revealed high interpretability in the DNN dimensions.
  4. Comparative Analysis:

    • Pairwise RSA analyses indicated a moderate global alignment between human and DNN RSMs with a Pearson correlation of 0.55.
    • High-correlating dimensions between the two modalities still revealed significant differences in interpretation, with seeming alignment driven by different feature sets.
  5. Behavioral Implications:

    • Analyzing the relevance of individual dimensions for odd-one-out tasks showed that humans prioritize semantic properties, while DNN prioritizes visual features.
    • Jackknife resampling demonstrated this behavioral divergence, emphasizing differing categorical strategies between humans and DNNs.

Implications and Future Directions

Practical Implications: The detailed comparison of representational embeddings suggests that while DNNs can approximate human-like semantic dimensions, they predominantly rely on visual features. This has direct implications for developing AI systems aimed at human-like cognitive tasks, such as those in medical imaging, autonomous driving, and interactive AI systems, recommending a focus on models that better balance semantic and visual information.

Theoretical Implications: From a theoretical perspective, these findings challenge the notion that DNNs naturally align with human cognition simply by increasing the complexity of visual tasks. The observed visual bias suggests that additional explicit training on semantic associations may be necessary to achieve closer representational fidelity to human cognition.

Future Developments: Future work could explore various DNN architectures and training paradigms to see if and how they impact the representational alignment with humans. The integration of higher-level cognitive tasks and richer semantic datasets might help bridge the visual-semantic gap observed. Furthermore, expanding the framework to different sensory modalities or complex stimuli can provide a holistic understanding of human and AI representational capabilities.

Conclusion

The presented framework and findings by Mahner et al. substantially advance our understanding of the dimensions underlying representational alignment between humans and DNNs. While showing a reasonable degree of global alignment, the deep dive into visual and semantic dimensions uncovers divergent strategies, emphasizing the need for new approaches in AI development to achieve more human-like cognitive models. The study sets the stage for further exploration into optimization and interpretability techniques that can better align DNN representations with human cognition.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.

YouTube