Contrastive Instruction Tuning (2402.11138v2)
Abstract: Instruction tuning has been used as a promising approach to improve the performance of LLMs on unseen tasks. However, current LLMs exhibit limited robustness to unseen instructions, generating inconsistent outputs when the same instruction is phrased with slightly varied forms or language styles. This behavior indicates LLMs' lack of robustness to textual variations and generalizability to unseen instructions, potentially leading to trustworthiness issues. Accordingly, we propose Contrastive Instruction Tuning, which maximizes the similarity between the hidden representations of semantically equivalent instruction-instance pairs while minimizing the similarity between semantically different ones. To facilitate this approach, we augment the existing FLAN collection by paraphrasing task instructions. Experiments on the PromptBench benchmark show that CoIN consistently improves LLMs' robustness to unseen instructions with variations across character, word, sentence, and semantic levels by an average of +2.5% in accuracy. Code is available at https://github.com/luka-group/CoIN.
- Enhancing Logical Reasoning of Large Language Models through Logic-Driven Data Augmentation. ArXiv:2305.12599 [cs].
- The second PASCAL recognising textual entailment challenge.
- The Fifth PASCAL Recognizing Textual Entailment Challenge.
- SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pages 1–14, Vancouver, Canada. Association for Computational Linguistics.
- BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 2924–2936, Minneapolis, Minnesota. Association for Computational Linguistics.
- The PASCAL Recognising Textual Entailment Challenge. In Machine Learning Challenges. Evaluating Predictive Uncertainty, Visual Object Classification, and Recognising Tectual Entailment, Lecture Notes in Computer Science, pages 177–190, Berlin, Heidelberg. Springer.
- William B. Dolan and Chris Brockett. 2005. Automatically Constructing a Corpus of Sentential Paraphrases. In Proceedings of the Third International Workshop on Paraphrasing (IWP2005).
- Exploring the Limits of Out-of-Distribution Detection. In Advances in Neural Information Processing Systems, volume 34, pages 7068–7081. Curran Associates, Inc.
- Black-box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers. ArXiv:1801.04354 [cs].
- SimCSE: Simple Contrastive Learning of Sentence Embeddings. ArXiv:2104.08821 [cs].
- The Third PASCAL Recognizing Textual Entailment Challenge. In Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, pages 1–9, Prague. Association for Computational Linguistics.
- Twitter Sentiment Classification using Distant Supervision.
- Robustness of Learning from Task Instructions. ArXiv:2212.03813 [cs].
- Toward Semantics-Based Answer Pinpointing. In Proceedings of the First International Conference on Human Language Technology Research.
- Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment. ArXiv:1907.11932 [cs].
- Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 252–262, New Orleans, Louisiana. Association for Computational Linguistics.
- Active Instruction Tuning: Improving Cross-Task Generalization by Training on Prompt Sensitive Tasks. ArXiv:2311.00288 [cs].
- ContrastNER: Contrastive-based Prompt Tuning for Few-shot NER. In 2023 IEEE 47th Annual Computers, Software, and Applications Conference (COMPSAC), pages 241–249. ArXiv:2305.17951 [cs].
- The Winograd schema challenge. In Proceedings of the Thirteenth International Conference on Principles of Knowledge Representation and Reasoning, KR’12, pages 552–561, Rome, Italy. AAAI Press.
- Xin Li and Dan Roth. 2002. Learning Question Classifiers. In COLING 2002: The 19th International Conference on Computational Linguistics.
- Exploring Format Consistency for Instruction Tuning. ArXiv:2307.15504 [cs].
- How Good Are Large Language Models at Out-of-Distribution Detection? ArXiv:2308.10261 [cs].
- BRIO: Bringing Order to Abstractive Summarization. ArXiv:2203.16804 [cs].
- Robustness Over Time: Understanding Adversarial Examples’ Effectiveness on Longitudinal Versions of Large Language Models. ArXiv:2308.07847 [cs].
- The Flan Collection: Designing Data and Methods for Effective Instruction Tuning. ArXiv:2301.13688 [cs].
- Learning Word Vectors for Sentiment Analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pages 142–150, Portland, Oregon, USA. Association for Computational Linguistics.
- The CommitmentBank: Investigating projection in naturally occurring discourse. Proceedings of Sinn und Bedeutung, 23(2):107–124. Number: 2.
- UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. ArXiv:1802.03426 [cs, stat].
- Cross-Task Generalization via Natural Language Crowdsourcing Instructions. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3470–3487, Dublin, Ireland. Association for Computational Linguistics.
- Adversarial NLI: A New Benchmark for Natural Language Understanding. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4885–4901, Online. Association for Computational Linguistics.
- Learning to Generalize for Cross-domain QA. ArXiv:2305.08208 [cs].
- Training language models to follow instructions with human feedback. ArXiv:2203.02155 [cs].
- Mohammad Taher Pilehvar and Jose Camacho-Collados. 2019. WiC: the Word-in-Context Dataset for Evaluating Context-Sensitive Meaning Representations. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 1267–1273, Minneapolis, Minnesota. Association for Computational Linguistics.
- Controllable Natural Language Generation with Contrastive Prefixes. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2912–2924, Dublin, Ireland. Association for Computational Linguistics.
- Know What You Don’t Know: Unanswerable Questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 784–789, Melbourne, Australia. Association for Computational Linguistics.
- Beyond Accuracy: Behavioral Testing of NLP Models with CheckList. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4902–4912, Online. Association for Computational Linguistics.
- Multitask Prompted Training Enables Zero-Shot Task Generalization. ArXiv:2110.08207 [cs].
- Proximal Policy Optimization Algorithms. ArXiv:1707.06347 [cs].
- Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1631–1642, Seattle, Washington, USA. Association for Computational Linguistics.
- Evaluating the Zero-shot Robustness of Instruction-tuned Language Models. ArXiv:2306.11270 [cs].
- Stanford Alpaca: An Instruction-following LLaMA Model. Original-date: 2023-03-10T23:33:09Z.
- LLaMA: Open and Efficient Foundation Language Models. ArXiv:2302.13971 [cs].
- GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 353–355, Brussels, Belgium. Association for Computational Linguistics.
- Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models.
- On the Robustness of ChatGPT: An Adversarial and Out-of-distribution Perspective.
- How far can camels go? exploring the state of instruction tuning on open resources. In Advances in Neural Information Processing Systems.
- Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks. ArXiv:2204.07705 [cs].
- Neural Network Acceptability Judgments. Transactions of the Association for Computational Linguistics, 7:625–641. Place: Cambridge, MA Publisher: MIT Press.
- Finetuned Language Models Are Zero-Shot Learners. ArXiv:2109.01652 [cs].
- A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1112–1122, New Orleans, Louisiana. Association for Computational Linguistics.
- Contrastive Training for Improved Out-of-Distribution Detection. ArXiv:2007.05566 [cs, stat].
- ZeroPrompt: Scaling Prompt-Based Pretraining to 1,000 Tasks Improves Zero-Shot Generalization. ArXiv:2201.06910 [cs].
- Instruction Tuning for Large Language Models: A Survey. ArXiv:2308.10792 [cs].
- Character-level Convolutional Networks for Text Classification. In Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc.
- PAWS: Paraphrase Adversaries from Word Scrambling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 1298–1308, Minneapolis, Minnesota. Association for Computational Linguistics.
- PromptBench: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts. ArXiv:2306.04528 [cs].