Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 37 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 10 tok/s Pro
GPT-5 High 15 tok/s Pro
GPT-4o 84 tok/s Pro
Kimi K2 198 tok/s Pro
GPT OSS 120B 448 tok/s Pro
Claude Sonnet 4 31 tok/s Pro
2000 character limit reached

NEFTune: Noisy Embeddings Improve Instruction Finetuning (2310.05914v2)

Published 9 Oct 2023 in cs.CL and cs.LG

Abstract: We show that LLM finetuning can be improved, sometimes dramatically, with a simple augmentation. NEFTune adds noise to the embedding vectors during training. Standard finetuning of LLaMA-2-7B using Alpaca achieves 29.79% on AlpacaEval, which rises to 64.69% using noisy embeddings. NEFTune also improves over strong baselines on modern instruction datasets. Models trained with Evol-Instruct see a 10% improvement, with ShareGPT an 8% improvement, and with OpenPlatypus an 8% improvement. Even powerful models further refined with RLHF such as LLaMA-2-Chat benefit from additional training with NEFTune.

Citations (60)

Summary

  • The paper introduces NEFTune, a method that adds noise to embedding vectors during fine-tuning to mitigate overfitting and enhance generative performance.
  • The approach improved the AlpacaEval score of LLaMA-2-7B from 29.8% to 64.7%, with additional gains of 8-10% on other instruction datasets.
  • The study highlights NEFTune's advantage of enhancing model generalization without extra computational resources, opening avenues for further research.

Evaluating NEFTune: Noisy Embeddings for Enhanced LLM Fine-Tuning

The paper "NEFTune: Noisy Embeddings Improve Instruction Finetuning" presents a novel yet straightforward enhancement to the fine-tuning process of LMs, specifically demonstrating significant improvements in generative conversational quality across various state-of-the-art datasets. The central premise of this work is the addition of noise to embedding vectors during the fine-tuning stage, a process termed as Noisy Embedding Instruction Fine-Tuning (NEFTune). This method is shown to enhance model performance without incurring any additional computational burden or data requirements.

Summary of Results

Strong numerical results underpin the efficacy of NEFTune. For instance, finetuning the LLaMA-2-7B model on the Alpaca dataset with added noise improves the AlpacaEval score from 29.8% to 64.7%, marking a substantial 35 percentage point gain. Similarly, improvements were reported for models finetuned on diverse instruction datasets such as Evol-Instruct, ShareGPT, and OpenPlatypus, with enhancements ranging from 8% to 10%. These results provide compelling evidence that NEFTune systematically boosts conversational performance, even when applied to models further refined with Reinforcement Learning from Human Feedback (RLHF) methodologies, as evidenced in the case of LLaMA-2-Chat.

Implications and Observations

The authors attribute the success of NEFTune to its ability to mitigate overfitting on small instruction datasets, thereby enhancing the model's generalization capabilities. This hypothesis is supported by an analysis of model training loss and the diversity of token outputs, which indicates that noise introduction reduces exact replication of training data while maintaining or boosting output quality. This decreased overfitting is likely due to the noise perturbing the acquired idiosyncrasies of the fine-tuning set, promoting learning representations that are robust to input variability and possibly extending pre-trained capabilities.

The practical implications of this work are noteworthy. NEFTune presents an efficient mechanism to enhance the conversational and generative capabilities of LLMs without demanding additional data or computational resources, a "free lunch" for LLM fine-tuning. This enhancement is pertinent given the increasing need for models capable of seamless, context-aware human interaction in various applications ranging from automated customer support to complex problem-solving advisories.

Prospective Developments

This paper opens several avenues for future research. Investigating the balance between noise-induced diversity and model coherence will provide deeper insights into optimizing NEFTune settings for different model architectures and datasets. Extending this framework to other modalities, such as multi-modal models handling both text and vision inputs, may leverage noise for enhanced cross-modal understanding and integration. Additionally, examining the interplay between NEFTune and other regularization strategies could yield further performance gains and inform adaptive finetuning techniques.

Finally, beyond empirical improvements, there is merit in exploring theoretical models to better understand and predict the impact of noise on learning dynamics within embedding spaces. Such theoretical advancements could lead to adaptive noise mechanisms that dynamically adjust in response to model and data characteristics, optimizing performance in a principled manner.

In conclusion, this paper contributes a valuable augmentation technique to the instruction fine-tuning process of LLMs, showcasing substantial improvements with minimal overhead. By promoting reduced overfitting and enhanced generalization, NEFTune offers a promising pathway towards more versatile and contextually adept LLMs.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Youtube Logo Streamline Icon: https://streamlinehq.com