Instruction Following without Instruction Tuning
Hewitt, John, Liu, Nelson F., Liang, Percy, Manning, Christopher D.
–arXiv.org Artificial Intelligence
Instruction tuning commonly means finetuning a language model on instructionresponse pairs. We discover two forms of adaptation (tuning) that are deficient compared to instruction tuning, yet still yield instruction following; we call this implicit instruction tuning. We first find that instruction-response pairs are not necessary: training solely on responses, without any corresponding instructions, yields instruction following. This suggests pretrained models have an instruction-response mapping which is revealed by teaching the model the desired distribution of responses. However, we then find it's not necessary to teach the desired distribution of responses: instruction-response training on narrow-domain data like poetry still leads to broad instruction-following behavior like recipe generation. In particular, when instructions are very different from those in the narrow finetuning domain, models' responses do not adhere to the style of the finetuning domain. To begin to explain implicit instruction tuning, we hypothesize that very simple changes to a language model's distribution yield instruction following. We support this by hand-writing a rule-based language model which yields instruction following in a product-of-experts with a pretrained model. The rules are to slowly increase the probability of ending the sequence, penalize repetition, and uniformly change 15 words' probabilities. In summary, adaptations made without being designed to yield instruction following can do so implicitly. Instruction tuning, finetuning on a broad distribution of responses (e.g., Tiramisu is made by...) conditioned on instructions (e.g., Give me a recipe for tiramisu), yields instruction following from language models for a wide range of instructions (Ouyang et al., 2022). Prior work has shown that instruction tuning is sample-efficient, requiring as few as 1000 broad-domain instruction-response pairs (Zhou et al., 2023) or a carefully crafted prompt and few-shot instruction-response examples (Lin et al., 2024). We take this a step further, exploring the idea that instruction following can be yielded from language models even implicitly, i.e., through methods not explictly designed to do so.
arXiv.org Artificial Intelligence
Sep-21-2024
- Country:
- Asia
- Europe > Russia (0.04)
- North America
- Canada (0.04)
- United States
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Government (0.46)
- Leisure & Entertainment > Games (0.30)
- Technology: