Papers
Topics
Authors
Recent
Search
2000 character limit reached

SpeechUT: Bridging Speech and Text with Hidden-Unit for Encoder-Decoder Based Speech-Text Pre-training

Published 7 Oct 2022 in cs.CL and eess.AS | (2210.03730v1)

Abstract: The rapid development of single-modal pre-training has prompted researchers to pay more attention to cross-modal pre-training methods. In this paper, we propose a unified-modal speech-unit-text pre-training model, SpeechUT, to connect the representations of a speech encoder and a text decoder with a shared unit encoder. Leveraging hidden-unit as an interface to align speech and text, we can decompose the speech-to-text model into a speech-to-unit model and a unit-to-text model, which can be jointly pre-trained with unpaired speech and text data respectively. Our proposed SpeechUT is fine-tuned and evaluated on automatic speech recognition (ASR) and speech translation (ST) tasks. Experimental results show that SpeechUT gets substantial improvements over strong baselines, and achieves state-of-the-art performance on both the LibriSpeech ASR and MuST-C ST tasks. To better understand the proposed SpeechUT, detailed analyses are conducted. The code and pre-trained models are available at https://aka.ms/SpeechUT.

Citations (54)

Summary

  • The paper introduces a unified speech-to-text pre-training model that bridges modalities via hidden-unit representations.
  • The methodology decouples tasks into speech-to-unit and unit-to-text stages, using multi-task learning with masked unit modeling.
  • The model achieves state-of-the-art results on LibriSpeech ASR and MuST-C ST, outperforming existing encoder and encoder-decoder approaches.

Overview of SpeechUT: Bridging Speech and Text with Hidden-Unit for Encoder-Decoder Based Speech-Text Pre-training

The paper presents SpeechUT, a pre-training model designed to effectively bridge the modalities of speech and text using hidden-unit representations. This approach leverages both unpaired speech and text data, offering a novel solution for tasks such as automatic speech recognition (ASR) and speech translation (ST). SpeechUT's innovative design capitalizes on a shared unit encoder to align the outputs of a speech encoder and a text decoder, thus facilitating improved performance in cross-modal pre-training.

Key Contributions

  1. Unified Speech-Unit-Text Pre-training Model: SpeechUT integrates a speech encoder and a text decoder via a unit encoder, facilitating seamless conversion between speech and text modalities. This structure allows the decomposition of tasks into speech-to-unit (S2U) and unit-to-text (U2T) models, which are pre-trained independently.
  2. Integration of Hidden-Units: The approach uses hidden-unit representations derived from models like HuBERT to align speech and text, providing a semantic interface between the two modalities.
  3. State-of-the-Art Performance: The model demonstrates significant improvements over existing baselines, achieving state-of-the-art results on the LibriSpeech ASR and MuST-C ST tasks.

Technical Insights

  • Pre-training Tasks: SpeechUT performs multi-task learning with three principal tasks:
    • Speech-to-Unit (S2U): Similar to HuBERT's masked prediction, this task predicts unit categories from masked speech features.
    • Unit-to-Text (U2T): A sequence-to-sequence task that reconstructs the text from the intermediate unit representation.
    • Masked Unit Modeling (MUM): Inspired by BERT, this task involves predicting masked unit tokens to enhance unit representation learning.
  • Embedding Mixing Mechanism: This technique strategically replaces parts of the speech features with corresponding unit embeddings, enhancing alignment between speech and unit representations.
  • Pre-training and Fine-tuning: The model is pre-trained with a combination of speech, unit, and text data. For ASR and ST tasks, all modules, including the text decoder, are fine-tuned without introducing new parameters.

Experimental Results

SpeechUT achieves remarkable results across various benchmarks. On the LibriSpeech ASR task, the model surpasses encoder-based and encoder-decoder models, such as wav2vec 2.0 and SpeechT5. Furthermore, SpeechUT attains superior BLEU scores in ST evaluations compared to recent works like STPT, even with reduced pre-training data.

Implications and Future Directions

The successful integration of hidden units as a bridge between speech and text opens avenues for more efficient and scalable pre-training methods. The decoupled pre-training strategy could inspire future research in unified multi-modal learning. Potential areas of exploration include multilingual extensions and refining the T2U generator to eliminate dependency on paired ASR data.

This paper positions SpeechUT as a powerful tool for advancing speech-related AI applications, underscoring the potential benefits of intermediary representations in cross-modal alignments.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.