Emergent Mind

Large Language Models are Parallel Multilingual Learners

(2403.09073)
Published Mar 14, 2024 in cs.CL

Abstract

In this study, we reveal an in-context learning (ICL) capability of multilingual LLMs: by translating the input to several languages, we provide Parallel Input in Multiple Languages (PiM) to LLMs, which significantly enhances their comprehension abilities. To test this capability, we design extensive experiments encompassing 8 typical datasets, 7 languages and 8 state-of-the-art multilingual LLMs. Experimental results show that (1) incorporating more languages help PiM surpass the conventional ICL further; (2) even combining with the translations that are inferior to baseline performance can also help. Moreover, by examining the activated neurons in LLMs, we discover a counterintuitive but interesting phenomenon. Contrary to the common thought that PiM would activate more neurons than monolingual input to leverage knowledge learned from diverse languages, PiM actually inhibits neurons and promotes more precise neuron activation especially when more languages are added. This phenomenon aligns with the neuroscience insight about synaptic pruning, which removes less used neural connections, strengthens remainders, and then enhances brain intelligence.

Overview

  • The paper presents a study on the in-context learning capabilities of LLMs for multilingual processing, introducing a novel prompting approach called Parallel Input in Multiple Languages (PiM).

  • PiM enhances multilingual LLM performance by augmenting input with translations in several languages, showing significant improvement across various tasks and languages.

  • Findings reveal that PiM may optimize neuron activation in LLMs, suggesting a more efficient use of neural networks for multilingual processing.

  • The study underscores PiM's broad applicability and potential for improving LLM performance in real-world scenarios, opening avenues for further interdisciplinary research.

LLMs are Parallel Multilingual Learners

Introduction

In this study, we explore the in-context learning (ICL) capabilities of LLMs for processing and understanding information provided in multiple languages simultaneously. This work introduces a novel prompting approach, Parallel Input in Multiple Languages (PiM), which significantly enhances the comprehension abilities of multilingual LLMs by augmenting the standard input with translations of the task's prompt into several languages. Through extensive experiments across a diverse set of datasets, languages, and state-of-the-art multilingual LLMs, this paper demonstrates the efficacy of PiM in improving model performance across a variety of tasks, including machine translation, language inference, reading comprehension, text simplification, and abstractive summarization.

Parallel Input in Multiple Languages

The study presents PiM as a method to leverage the inherent capability of multilingual LLMs to process inputs in multiple languages. PiM involves translating the original input into several languages and presenting these translations alongside the original input to the LLM. This approach is theorized to enrich the context available to the model, thereby improving its performance. The hypothesis is substantiated by significant improvements observed across eight datasets, seven languages, and eight leading multilingual LLMs. Notably, PiM demonstrates effectiveness even when translations do not outperform direct translations in baseline scenarios, suggesting a robust method to enhance multilingual model performance.

Insights and Theoretical Implications

A counterintuitive discovery made through neuron activation analysis in LLMs suggests that, contrary to expectations, PiM does not necessarily increase the number of activated neurons. Instead, it inhibits neurons while promoting more precise neuron activation, especially with the addition of more languages to the input. This observation indicates a potential optimization in how LLMs access and utilize multilingual knowledge, aligning with processes of synaptic pruning observed in neurological studies. These findings suggest that PiM's effectiveness may stem from inducing a more efficient use of the model's neural network, emphasizing quality over quantity in neuron activation.

Practical Applications and Future Directions

The paper evidences the broad applicability of PiM across various NLP tasks and its compatibility with multiple LLM architectures, from 7B to 176B parameters. The success of PiM in improving translation tasks, even with machine-translated inputs, opens new pathways for its use in enhancing LLM performance in real-world scenarios. Furthermore, the study highlights an intriguing direction for future research on understanding neuron activation patterns in LLMs and their relation to learning processes in human brains. Given the effectiveness of PiM, further exploration into tailored prompting strategies for different types of tasks and languages could yield additional gains in model performance and efficiency.

Conclusions

This research contributes significantly to the field by demonstrating a simple yet effective strategy to improve the performance of multilingual LLMs across a range of tasks. By adopting PiM, the study not only provides a practical method for leveraging the multilingual capabilities of LLMs but also offers new insights into the optimization of neural networks for multilingual understanding. The revelations regarding neuron activation patterns offer a fascinating glimpse into the potential analogs between artificial and biological learning processes, presenting an exciting avenue for interdisciplinary research bridging AI and neurosciences.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.