Fine-Tuning on Noisy Instructions: Effects on Generalization and Performance
Abstract
Introducing perturbations in instruction-tuning data can enhance large language models' resistance to noisy instructions and improve performance on benchmarks.
Instruction-tuning plays a vital role in enhancing the task-solving abilities of large language models (LLMs), improving their usability in generating helpful responses on various tasks. However, previous work has demonstrated that they are sensitive to minor variations in instruction phrasing. In this paper, we explore whether introducing perturbations in instruction-tuning data can enhance LLMs' resistance against noisy instructions. We focus on how instruction-tuning with perturbations, such as removing stop words or shuffling words, affects LLMs' performance on the original and perturbed versions of widely-used benchmarks (MMLU, BBH, GSM8K). We further assess learning dynamics and potential shifts in model behavior. Surprisingly, our results suggest that instruction-tuning on perturbed instructions can, in some cases, improve downstream performance. These findings highlight the importance of including perturbed instructions in instruction-tuning, which can make LLMs more resilient to noisy user inputs.
Community
Instruction-tuning is crucial for enhancing large language models’ (LLMs) ability to follow tasks and generate useful responses. Yet, prior work shows that LLMs remain sensitive to small variations in instruction phrasing. This paper investigates whether introducing perturbations during instruction-tuning can improve robustness to noisy inputs. We apply perturbations such as stop-word removal and word shuffling and evaluate performance on original and perturbed versions of MMLU, BBH, and GSM8K. Our results show that tuning on perturbed instructions can, in some cases, enhance downstream performance and stability. These findings suggest that incorporating controlled noise in instruction-tuning may yield more resilient and adaptable LLMs.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Analyzing the Effects of Supervised Fine-Tuning on Model Knowledge from Token and Parameter Levels (2025)
- When Punctuation Matters: A Large-Scale Comparison of Prompt Robustness Methods for LLMs (2025)
- Rethinking Safety in LLM Fine-tuning: An Optimization Perspective (2025)
- When Long Helps Short: How Context Length in Supervised Fine-tuning Affects Behavior of Large Language Models (2025)
- ReSURE: Regularizing Supervision Unreliability for Multi-turn Dialogue Fine-tuning (2025)
- Continuously Steering LLMs Sensitivity to Contextual Knowledge with Proxy Models (2025)
- On Robustness and Reliability of Benchmark-Based Evaluation of LLMs (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper