We present a scalable method to build a high quality instruction following language model by automatically labelling human-written text with corresponding instructions. Our approach, named instruction backtranslation, starts with a language model finetuned on a small amount of seed data, and a given web corpus. The seed model is used to construct training examples by generating instruction prompts for web documents (self-augmentation), and then selecting high quality examples from among these candidates (self-curation). This data is then used to finetune a stronger model. Finetuning LLaMa on two iterations of our approach yields a model that outperforms all other LLaMa-based models on the Alpaca leaderboard not relying on distillation data, demonstrating highly effective self-alignment.
The paper presents a method called instruction backtranslation for improving the instruction-following capabilities of LLMs by using an iterative self-training algorithm on unlabelled data.
The methodology includes two primary phases: self-augmentation and self-curation, where a seed model generates and evaluates high-quality (instruction, output) pairs from unlabelled web documents.
The resultant model, named Humpback, demonstrates superior performance on benchmarks without relying on extensive human annotations or distillation from more powerful models, notably excelling in zero-shot settings.
The paper titled "Self-Alignment with Instruction Backtranslation" proposes an innovative method for finetuning LLMs to improve their instruction-following capabilities. The core of the methodology is an iterative self-training algorithm termed instruction backtranslation, which leverages unlabelled data to create high-quality training datasets through a two-step process of self-augmentation and self-curation. This approach is inspired by backtranslation in machine translation.
The approach hinges on two primary phases:
The process is iterated to improve the model's performance incrementally. This paper introduces Humpback, a model built through two iterations of instruction backtranslation using the LLaMa model as the base. The resultant model demonstrates superior performance compared to other non-distilled models on established benchmarks such as the Alpaca leaderboard.
The transformed LLM, Humpback
, is empirically validated against various baselines:
The model's instruction-following capability and general quality are assessed through both automated and human evaluations:
Moreover, commonsense reasoning and massive multitask language understanding (MMLU) benchmarks show notable improvements, particularly in zero-shot settings, suggesting that Humpback has enhanced generalization capabilities.
The implications of this research span both practical and theoretical domains. Practically, the method enables the creation of high-quality instruction-following models without reliance on extensive human annotations or distillation from more powerful models, markedly reducing resource requirements. Theoretically, it underscores the efficacy of self-alignment in LLMs, potentially setting a new paradigm in model training.
This paper presents a compelling case for instruction backtranslation as a scalable method for finetuning LLMs, demonstrating substantial improvements in instruction-following performance. Future developments could explore scaling this method further by harnessing larger unlabeled corpora and integrating advanced curation strategies to meet diverse application requirements. This path may well drive the next leap in autonomous AI systems proficiency in understanding and executing complex instructions.