Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 126 tok/s
Gemini 2.5 Pro 45 tok/s Pro
GPT-5 Medium 29 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 127 tok/s Pro
Kimi K2 183 tok/s Pro
GPT OSS 120B 425 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Coupling Artificial Neurons in BERT and Biological Neurons in the Human Brain (2303.14871v1)

Published 27 Mar 2023 in cs.CL

Abstract: Linking computational NLP models and neural responses to language in the human brain on the one hand facilitates the effort towards disentangling the neural representations underpinning language perception, on the other hand provides neurolinguistics evidence to evaluate and improve NLP models. Mappings of an NLP model's representations of and the brain activities evoked by linguistic input are typically deployed to reveal this symbiosis. However, two critical problems limit its advancement: 1) The model's representations (artificial neurons, ANs) rely on layer-level embeddings and thus lack fine-granularity; 2) The brain activities (biological neurons, BNs) are limited to neural recordings of isolated cortical unit (i.e., voxel/region) and thus lack integrations and interactions among brain functions. To address those problems, in this study, we 1) define ANs with fine-granularity in transformer-based NLP models (BERT in this study) and measure their temporal activations to input text sequences; 2) define BNs as functional brain networks (FBNs) extracted from functional magnetic resonance imaging (fMRI) data to capture functional interactions in the brain; 3) couple ANs and BNs by maximizing the synchronization of their temporal activations. Our experimental results demonstrate 1) The activations of ANs and BNs are significantly synchronized; 2) the ANs carry meaningful linguistic/semantic information and anchor to their BN signatures; 3) the anchored BNs are interpretable in a neurolinguistic context. Overall, our study introduces a novel, general, and effective framework to link transformer-based NLP models and neural activities in response to language and may provide novel insights for future studies such as brain-inspired evaluation and development of NLP models.

Citations (10)

Summary

  • The paper introduces a framework that couples BERT's 9216 fine-grained artificial neurons with fMRI-derived functional brain networks to reflect linguistic processing.
  • It employs the element-wise product of query and key vectors and synchronizes activations using Pearson correlation to map artificial to biological neurons.
  • Experimental results on the 'Narratives' fMRI dataset reveal significant coupling between deep-layer ANs and BNs, highlighting neurolinguistic integration.

Overview of the Framework

The paper "Coupling Artificial Neurons in BERT and Biological Neurons in the Human Brain" introduces a framework to link artificial neurons (ANs) within transformer-based NLP models, such as BERT, and biological neurons (BNs) within the human brain. This framework seeks to resolve existing limitations by increasing the granularity of ANs and capturing the functional interactions between BNs. By embedding input text sequences with a pre-trained BERT and analyzing functional magnetic resonance imaging (fMRI) data, the paper enhances our understanding of linguistic processing at a neural level.

Defining Artificial and Biological Neurons

The paper addresses the lack of fine-granularity in existing models by considering each dimension of the query/key vectors within BERT’s multi-head self-attention (MSA) as an individual AN. This approach results in a large number of ANs (9216 in BERT). Temporal activations of these ANs are quantified by considering the element-wise product of queries and keys, offering a deeper insight into linguistic perception.

Conversely, biological neurons (BNs) are defined as functional brain networks (FBNs), extracted via the volumetric sparse deep belief network (VS-DBN) applied to fMRI data. This model captures the integrational and segregational brain activities, providing a fine-grained analysis of brain function interaction in response to linguistic stimuli. Figure 1

Figure 1: The framework for coupling artificial neurons in BERT and biological neurons in the brain.

Synchronization and Experiments

Synchronization between ANs and BNs is achieved by aligning their temporal activations. The Pearson correlation coefficient (PCC) is employed to measure synchronization, identifying the coupling between ANs and BNs by maximizing the correlations of their temporal activations.

The paper utilizes the "Narratives" fMRI dataset, including sessions with spoken stories, to demonstrate the framework’s effectiveness. The experimental results reveal that a significant portion of ANs in BERT’s deep layers are synchronized with BNs, uncovering the association between linguistic perception and neural activation. Figure 2

Figure 2: The definitions of AN and its activation in BERT. (a) The simplified attention matrix by removing the softmax operations. (b) The contribution of each dimension in query/key to the attention matrix. (c) The element-wise product of queries and keys. (d) The AN's activation is measured by each dimension of the element-wise product.

Interpretation of Results

Significantly synchronized ANs and BNs underpin the framework's capacity to reflect meaningful linguistic and semantic information. The distinct spatial distributions of BNs indicate that they are not merely passive responders but play a crucial role aligned with linguistic integration and perception. The paper observed different patterns of neural activation for different story stimuli, offering insights into the dynamic nature of linguistic processing within the brain. Figure 3

Figure 3: The definitions of BN and its activation in fMRI. (a) The model input is fMRI volumes. (b) The VS-DBN consists of three RBM layers. (c) The global latent variable W. (d) The spatial distribution of FBNs. (e) The temporal activations of FBNs.

Conclusion and Future Directions

The framework presented reveals significant potential in linking linguistic processing models with neural activities. It shows that transformer-based NLP models can serve as representative tools to explore the neural basis of language processing. Despite its limitations, such as the reliance on unidirectional attention dynamics in BERT and simplistic measures of AN activation, the framework sets a foundation for future research. Future work could explore other linguistic tasks or models, such as GPT-3, to assess further consistency in AN-BN coupling patterns, enriching the understanding of neurolinguistic interactions.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.