A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (2024)

Hao Zhao  Maksym Andriushchenko  Francesco Croce  Nicolas Flammarion

Abstract

There is a consensus that instruction fine-tuning of LLMs requires high-quality data, but what are they?LIMA (NeurIPS 2023) and AlpaGasus (ICLR 2024) are state-of-the-art methods for selecting such high-quality examples, either via manual curation or using GPT-3.5-Turbo as a quality scorer.We show that the extremely simple baseline of selecting the 1,000 instructions with longest responses—that intuitively contain more learnable information and are harder to overfit—from standard datasets can consistently outperform these sophisticated methods according to GPT-4 and PaLM-2 as judges, while remaining competitive on the Open LLM benchmarks that test factual knowledge.We demonstrate this for several LLMs (Llama-2-7B, Llama-2-13B, Mistral-7B-v0.1) and datasets (Alpaca-52k, Evol-Instruct-70k).In addition, a lightweight refinement of such long instructions can further improve the abilities of the fine-tuned LLMs, and allows us to obtain competitive results on MT-Bench and the 2nd highest-ranked Llama-2-7B-based model on AlpacaEval 2.0, while training on only 1,000 examples and no extra preference data.We also conduct a thorough analysis of our models to ensure that their enhanced performance is not simply due to GPT-4’s preference for longer responses. Overall, our findings suggest that fine-tuning on the longest responses should be the default baseline for any work on instruction fine-tuning. We provide our code in this GitHub repository.

Machine Learning, ICML

1 Introduction

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (1)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (2)

Pre-trained large language models (LLMs) need to undergo an alignment phase(Askell etal., 2021; Bai etal., 2022a; Ouyang etal., 2022; Wang etal., 2022; Taori etal., 2023) to make them suitable for downstream tasks like user interaction or question answering.While the details may vary, alignment often relies on supervised fine-tuning (SFT) on a dataset of instruction-response pairs to improve conversational ability, followed by reinforcement learning from either human (RLHF)(Ouyang etal., 2022) or automated (RLAIF)(Bai etal., 2022b; Lee etal., 2023) feedback to promote the preferred style and content of replies.It is an active research direction to study whether it is possible to achieve satisfactory results while relying only on SFT, which would avoid the (potentially expensive) process of collecting preference data.Taori etal. (2023) created Alpaca, an open source dataset of 52k instruction-response pairs, and fine-tuned on it a Llama-2-7B model to match the performance of the closed-source text-davinci-003 model.Then, Chen etal. (2023) introduced AlpaGasus, consisting of the 9k examples of Alpaca which are judged of highest quality by GPT-3.5-Turbo, to further improve the instruction-following abilities of the fine-tuned models.The intuition that instruction fine-tuning (IFT) might benefit from fewer demonstrations but of higher quality has been further pursued by Zhou etal. (2023) which manually curated LIMA, a dataset of 1k examples, which outperforms AlpaGasus.While the quality of the instructions seems to play a major role for IFT, it remains unclear which are the distinguishing features of high quality demonstrations.

In this work, we revisit the significant efforts in constructing instruction-tuning datasets from prior work.Inspired by the fact LIMA contains much longer examples than Alpaca and the observation of recent works (Singhal etal., 2023; Yuan etal., 2024) that RLHF and direct preference optimization (DPO) (Rafailov etal., 2023) seem to mostly make the outputs longer, we test selecting longest responses as a simple and inexpensive heuristic to curate a small (only 1k examples) and high-quality IFT dataset from a larger one.Surprisingly, fine-tuning a Llama-2-7B (Touvron etal., 2023) base model on the 1k longest elements of Alpaca outperforms both AlpaGasus and LIMA in one-to-one comparison with different LLMs as judges and on the AlpacaEval 2.0 benchmark (see Fig.1).Moreover, simply improving the quality and the style of the response in Alpaca-1k-longest with GPT-3.5-Turbo, in combination with NEFTune noise augmentation (Jain etal., 2023), allows us to obtain the 2nd highest-ranked Llama-2-7B-based model on AlpacaEval 2.0.In this case, our simple method yields models which surpass LLMs with the same base model but fine-tuned with orders of magnitude more instructions as well as millions of preference data points.

Next we analyze several aspects of our models to understand the unexpected effectiveness of our approach. First, via several ablation studies, we show that our models do not just exploit the bias to favor longer responses of GPT-4 (OpenAI, 2023) or PaLM-2 (Anil etal., 2023), but provide higher quality replies.Then, since Jha etal. (2023); Gudibande etal. (2023) suggest that optimizing performance of instruction-following tasks mightbe disconnected from factual knowledge, we additionally test our models on then Open LLM benchmarks.On these datasets assessing reasoning and factuality, our models perform similarly or better than the baselines fine-tuned on AlpaGasus and LIMA from the same base model, i.e. with the same factual knowledge coming from pre-training.Finally, we confirm our findings with extensive experiments using multiple IFT datasets (Alpaca, Evol-Instruct) and architectures (Llama-2-7B, Llama-2-13B, Mistral-7B-v0.1 (Jiang etal., 2023)), and including head-to-head evaluation and on established benchmarks (AlpacaEval 2.0, Open LLM), to show the generality of our approach.

In summary, we uncover the surprising effectiveness of fine-tuning only on the longest 1,000 instructions of large datasets to obtain aligned models.Moreover, we show that such small datasets, potentially refined via an inexpensive automatic process, constitute a strong and tough-to-beat baseline for any method for instruction fine-tuning.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (3)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (4)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (5)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (6)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (7)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (8)

2 Related work

Instruction fine-tuning of LLMs.Since pre-trained LLMs usually do not accurately understand user intents and provide coherent and beneficial responses, an instruction fine-tuning stage is necessary (Ouyang etal., 2022; Bai etal., 2022a). Diversity of demonstrations and tasks(Chung etal., 2022; Xu etal., 2022) plays a pivotal role in enhancing the instruction-following performance of LMs. InstructGPT(Ouyang etal., 2022) first demonstrated how to achieve impressive performance in handling open-ended queries by fine-tuning GPT-3 models (Brown etal., 2020) with RLHF, which led to the release of ChatGPT. Subsequently, the community attempted to replicate the exceptional performance of proprietary models (Wang etal., 2023; Xu etal., 2023; Chiang etal., 2023), but Gudibande etal. (2023) show that it might be easy to mimic the style but not the factuality of closed-source LLMs. Singhal etal. (2023) identify a strong correlation between response length and reward when doing RLHF, implying that optimizing response length might be an implicit goal of RLHF.Also, Yuan etal. (2024) show that their self-improved reward model based on DPO encourages more verbose responses.

Data selection for IFT.The community has focused on creating IFT datasets of high quality(Peng etal., 2023). As one of the pioneering works, Alpaca(Taori etal., 2023) collects 52k interactions with the text-davinci-003 model using techniques from Self-Instruct(Wang etal., 2022).However, direct distillation from language models without careful screening inevitably introduces demonstrations with incorrect or ill-favored answers.To filter these cases out, AlpaGasus(Chen etal., 2023) measures the quality of each demonstration using a powerful LLM (GPT-3.5-Turbo) as a scorer. Touvron etal. (2023) note that fewer (in the order of tens of thousands) but higher-quality examples annotated by their own vendors significantly improve their Llama-2-Chat models. The definition of data quality also pertains to other factors, such as the complexity of queries(Xu etal., 2023), the difficulty of tasks presented(Mukherjee etal., 2023) and the diversity of semantics(Lu etal., 2023). Zhao etal. (2023) propose to control these factors through an instruction refinement approach, which maintains an instruction semantic tree and yields new instructions by modifying the structure of the semantic tree.To better reflect human intentions, LIMA(Zhou etal., 2023) relies on community forums and human labor to curate 1,000 demonstrations with an emphasis on quality and diversity, achieving strong instruction-following ability, surpassing some proprietary LLMs. They also formulate the Superficial Alignment Hypothesis: the general-purpose capabilities of an LLM mostly come from pre-training, and instruction tuning only guides the LLM to mimic the style, persona, and instruction adherence of desired outputs.Finally, while Liu etal. (2023); Cao etal. (2023) consider response length as an indicator of example quality, our comprehensive exploration is the first to reveal its reliability and effectiveness in data selection for IFT.

3 Fine-tuning on long instructions is a very strong baseline

We first study the importance of length of the training examples for IFT, and its applicability as a simple and inexpensive heuristic to obtain small but effective IFT datasets.Surprisingly, we observe that this simple criterion can often outperform much more sophisticated existing methods.

3.1 Subsampling high-quality IFT datasets

Existing methods.Recent works have shown that IFT on a small curated dataset of instructions is sufficient to enhance the ability of LLMs to follow instructions and complete tasks.In particular, Chen etal. (2023) adopt GPT-3.5-Turbo as the oracle to judge the quality of (instruction, input, output) tuples with grades on a 1-5 scale. Only the highest scoring examples (grade \geq 4.5) from Alpaca-52k (but the same approach can be generalized to other datasets) are used to form the AlpaGasus dataset on 9k instructions.Later, Zhou etal. (2023) collect 750 top instruction-response pairs from community forums with some heuristic rules, such as comments and upvotes, and manually write 250 examples to enhance task diversity and quality. These 1,000 examples are optimized for a uniform response style to turn the LLM into a useful conversational agent, and constitute the LIMA-1k dataset.

Our simple baseline: 1k instructions with the longest responses.Though both AlpaGasus and LIMA present promising performance improvements, they require either access to proprietary LLMs or very expensive human labor.Then, since previous works suggest that longer responses naturally arise during alignment (Singhal etal., 2023; Yuan etal., 2024),we explore response length as the selection criterion to prune IFT datasets.We select the 1,000 longest examples from the popular Alpaca-52k and Evol-Instruct-70k datasets to form our IFT datasets that we refer to as Alpaca-1k-longest and Evol-Instruct-1k-longest.Note that we use the term 1k-longest examples or responses interchangeably for simplicity in the remaining text, but always refer to the length of the responses.We restrict ourselves to using 1,000 examples for consistency with LIMA and since we are interested in testing how far the instruction following ability of LLMs can be pushed with a minimal SFT dataset.Using longer examples can be seen as a natural choice since these are usually more informative and thus contain more features relevant to human intentions.Longer responses are also intuitively harder for LLMs to fit, which forces the model to actually learn the response style rather than just memorize the answer.In addition, fitting longer responses encourages the model to capture long-distance semantic connections, and stay on-topic when answering complicated instructions. We provide empirical evidence to support our intuition in App.B.5.Interestingly, we observe that the instructions with longest responses minimally overlap with those receiving high score by LLMs: for example, most of the 1k longest examples from Alpaca receive a score of 3.5 from GPT-3.5-Turbo, i.e. signficantly lower than those in AlpaGasus (see details in Fig.13 in App.B.1).

3.2 Effectiveness of our approach for open-ended generation

Setting.To test the effectiveness of our approach,we compare our 1k-longest datasets to the full original Alpaca and Evol-Instruct datasets (52k and 70k examples), the 1k examples with highest scores according to GPT-3.5-Turbo as done by Chen etal. (2023) (hence we refer to these as AlpaGasus-1k and Evol-Instruct-AlpaGasus-1k), and LIMA-1k.For each instruction dataset, we fine-tune Llama-2-7B base models (complete training configurations in App.A.2).Then, we test their abilities on five evaluation datasets (LIMA, Vicuna, Koala, WizardLM, Self-Instruct, see the description of the datasets in App.A.1). We provide head-to-head comparisons in terms of win rate, where GPT-4 judges the preferable response (ties are allowed, details in App.A.3).

Results.Fig.2 shows that the responses of our models fine-tuned on the 1k-longest examples of either Alpaca or Evol-Instruct consistently outperform the existing methods across evaluation datasets.In particular, Alpaca-1k-longest is largely preferred over all competitors, and has an average win rate of 46.3% vs. LIMA-1k, with only 28.3% of losses (see Fig.1).This performance is significant when considering that LIMA has been carefully curated manually while our instructions come from a simpler dataset and selected only according to their length.Similarly, Evol-Instruct-1k-longest clearly outperforms LIMA-1k and the full Evol-Instruct-72k, while it has a smaller but consistent advantage over Evol-Instruct-AlpaGasus-1k.We hypothesize that the advantage is smaller on Evol-Instruct because Evol-Instruct contains higher-quality data than Alpaca, thus even selecting examples using GPT-3.5-Turbo scores can find relatively effective training examples.Finally, to exclude the possibility of overfitting to GPT-4 preferences, we repeat this evaluation with PaLM-2 as judge and even in this case our models are largely preferred (see Fig.14 in App.B.2).

Role of response length.As frontier LLMs like GPT-4 might be biased to favor longer responses (Zheng etal., 2023), Fig.1 additionally illustrates the average length (as number of tokens) of the responses in several datasets described above, as well as the average length of the responses generated by the LLMs fine-tuned on them during evaluation (on 1030 new instructions from the 5 evaluation datasets). As expected, both training and generated answers of Alpaca-1k-longest are longer than those of Alpaca and AlpaGasus.Interestingly, the training examples of LIMA-1k are more than two times longer than those of Alpaca-1k-longest, while the generated responses of the two models are similar.We conclude that the length of the responses is not the main factor for our model being consistently preferred to LIMA-1k.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (9)

4 How far can we go with 1,000 instructions?

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (10)

In Sec.3 we have shown that length is a strong heuristic to select which instructions to use for IFT. However, the resulting LLMs still fall short compared to those fine-tuned with either more sophisticated (proprietary) pools of instructions or especially preference data e.g. via RLHF.Then, in the following, we want to explore the limit of the ability that can be achieved from SFT on 1k examples.For this, we first refine the style of the longest-1k instructions to be more amenable for IFT.Second, we show that our dataset and NEFTune (Jain etal., 2023), a recent algorithm to improve IFT via noise augmentation, can be successfully combined.Finally, we test that the ability of our models in instruction-following evaluations (1) is stable even when forcibly changing the response length, and (2) does not negatively impact their performance on factual knowledge benchmarks.

4.1 Refining the instructions via introspection

Models# IFT Data# Pref. DataWin RateAvg. Length
Notable baselines
GPT-4-Turbo*??50.050.050.050.02049204920492049
Alpaca-7B*52k02.592.592.592.59396396396396
Vicuna-7B*70k04.164.164.164.161044104410441044
Base model: Llama-2-7B
Llama-2-Chat-7B*27k3M4.964.964.964.961479147914791479
  + Evol70k-NEFTune*97k3M7.607.607.607.601612161216121612
Tulu-2-DPO-7B*326k64k8.208.208.208.201663166316631663
\hdashline AlpaGasus-1k1k02.692.692.692.69745745745745
LIMA-1k1k02.742.742.742.741360136013601360
Alpaca-52k52k02.742.742.742.74586586586586
Alpaca-1k-longest1k03.163.163.163.161810181018101810
  + max gen. 2048 \rightarrow 40961k03.113.113.113.112290229022902290
Evol-Instruct-70k70k03.443.443.443.44850850850850
Evol-Instruct-1k-longest1k04.094.094.094.091866186618661866
  + max gen. 2048 \rightarrow 40961k04.164.164.164.162486248624862486
Evol-Instruct-AlpaGasus-1k1k04.324.324.324.321156115611561156
Refined-Evol-Instruct-1k-longest1k05.125.125.125.121289128912891289
Refined-Alpaca-1k-longest1k06.006.006.006.001732173217321732
  + max gen. 2048 \rightarrow 40961k06.036.036.036.032326232623262326
  + NEFTune1k07.887.887.887.881801180118011801
  + NEFTune + 2048 \rightarrow 40961k07.837.837.837.832478247824782478
Base model: Mistral-7B-v0.1
Alpaca-52k52k03.423.423.423.42450450450450
AlpaGasus-1k1k04.914.914.914.91502502502502
LIMA-1k1k06.766.766.766.761197119711971197
Alpaca-1k-longest1k07.137.137.137.13937937937937
Refined-Alpaca-1k-longest1k011.7411.7411.7411.741170117011701170
  + max gen. 2048 \rightarrow 40961k011.7611.7611.7611.761330133013301330
  + NEFTune1k011.9411.9411.9411.941199119911991199
Base model: Llama-2-13B
Alpaca-52k52k03.903.903.903.90556556556556
Alpaca-1k-longest1k04.804.804.804.801104110411041104
AlpaGasus-1k1k04.874.874.874.87540540540540
LIMA-1k1k05.645.645.645.641097109710971097
Refined-Alpaca-1k-longest1k08.448.448.448.441646164616461646
  + max gen. 2048 \rightarrow 40961k08.308.308.308.302244224422442244
  + NEFTune1k08.768.768.768.761582158215821582

As suggested by Zhou etal. (2023), the goal of IFT is to teach LLMs the format to employ when interacting with the users rather than instilling new knowledge.We argue that fine-tuning on rich and detailed instructions may improve the ability of the models to capture deeper semantic structure and logic. Then, we want to refine our 1k-longest examples to improve the quality of responses of training examples in terms of style, structure and the level of detail.In fact, there is no guarantee that the instructions selected by length also have high quality in terms of structure, glossary and logic.

Given that LLMs are surprisingly good at self-improving(Huang etal., 2022; Pan etal., 2023) and judging(Zheng etal., 2023; Li etal., 2023b), we propose using an Oracle LLM for this task, via encouraging it to introspect.In particular, inspired by Chain-of-Thought prompting(Wei etal., 2022), we prompt the GPT-3.5-Turbo model to produce a brief review of the original response given the instruction, followed by a new response generation process that has access to the original instruction-response pair and the introspection output. The details of the prompt are presented in Fig.3.Applying this procedure to the 1k-longest examples of Alpaca and Evol-Instruct we obtain new IFT datasets: Refined-Alpaca-1k-longest and Refined-Evol-Instruct-1k-longest.

4.2 Instruction-following evaluation

Setup.First, we provide a pairwise comparison between fine-tuning different LLMs on our Refined-1k-longest and baseline datasets, in particular LIMA-1k.Next, to facilitate a unified comparison of all models and position them among existing baselines, we compute their performance on AlpacaEval 2.0 (Li etal., 2023b) and MT-Bench(Zheng etal., 2023).This allows us to compare many LLMs, including those reported on the existing leaderboards by previous works, more efficiently than with pairwise analyses.

Head-to-head comparisons.We compare fine-tuning on our Refined-Alpaca-1k-longest to the Alpaca-1k-longest against Alpaca-52k , AlpaGasus-1k and LIMA-1k in a head-to-head fashion: Fig.4 reports the average (over the 5 evaluation datasets introduced in Sec.3.2) preference of GPT-4, repeated for three base models, i.e. Llama-2-7B, Mistral-7B-v0.1, Llama-2-13B (the corresponding results with PaLM-2 as judge are shown in Fig.15 in App.B.2).In all cases the models fine-tuned on the plain Alpaca-1k-longest already outperform the baselines with the exception of LIMA-1k for Llama-2-13B.In particular, LIMA-1k makes the strongest existing method: however, when we compare it with our Refined-Alpaca-1k-longest, this last one has a significant advantage over LIMA-1k, with an average win rate of 59.9% across architectures vs the 20.2% of LIMA. This shows the effectiveness of the refinement via introspection on the longest examples from Alpaca, even when used by different base models.Moreover, we complement the head-to-head comparisons by conducting human-based evaluations, which reflect the preferences of real users. Concretely, we design a user study (see more details in App.A.3) to compare the responses generated by our Alpaca-1k-longest to those of Alpaca-52k, with Llama-2-7B as the base model. Note that we instruct the evaluators not to consider the length of the responses in their judgment. Finally, we collect 425 human preferences, over which Alpaca-1k-longest obtains 71.0% win rate, which agrees with the conclusion of the LLMs as judges.

AlpacaEval 2.0 evaluation.In Table1 we report the results on the AlpacaEval 2.0 benchmark of our models and some baselines copied from the public leaderboard.111https://tatsu-lab.github.io/alpaca_eval/Moreover, we show the architecture, size of IFT and preference datasets, and average response length for each entry.Among Llama-2-7B models, both LIMA-1k and Alpaca-52k fine-tuned models achieve win rate below 3%, which is outperformed by Alpaca-1k-longest (3.11%).Switching to the instructions refined by introspection (Refined-Alpaca-1k-longest) almost doubles the win rate, achieving 6.00%, which even surpasses the original Llama-2-Chat-7B, fine-tuned with 27k instructions and 3M preference pairs. Since Jain etal. (2023) showed that NEFTune, which injects noise on the embedded inputs as augmentation, can improve the performance of IFT, we test it in combination with our dataset: this yields 7.88% win rate, i.e. the second best Llama-2-7B model appearing on the leaderboard, ahead of Llama-2-7B-Evol-Instruct-NEFTune (Jain etal., 2023) and not far from the 8.20% win rate of Tulu-2-DPO-7B(Ivison etal., 2023). Interestingly, fine-tuned models with similar average response lengths, exhibit wildly distinct win rates. For example, when we refine Alpaca-1k-longest via introspection and enable NEFTune in fine-tuning, the win rate rises from 3.16% to 7.88%, while the average response lengths are almost the same.Overall, these results illustrate how using a simple dataset of 1,000 instructions which did not necessitate any manual curation can compete with more expensive and sophisticated alignment schemes relying on SFT with hundreds of thousands of examples and involving RLHF on up to 3M preference pairs. Moreover, we observe similar behavior with other architectures: for Mistral-7B-v0.1 Alpaca-1k-longest already outperforms the baseline methods, but the refined instructions give the most notable increase (7.13% to 11.74%) in win rate.Similarly, Refined-Alpaca-1k-longest attains the best results for Llama-2-13B.Interestingly, unlike for Llama-2-7B, in these cases the improvements given by NEFTune are marginal (\leq 0.32%), which highlights the importance of the fine-tuning dataset.Furthermore, we surprisingly find that Refined-Evol-Instruct-1k-longest (5.12%) underperforms compared to Refined-Alpaca-1k-longest (6.00%), which may be attributed to the limited capability of GPT-3.5-Turbo in understanding long-form text (see details of average response lengths in Fig.12), such as demonstrations consisting of thousands of tokens.

DatasetsLlama-2-7BLlama-2-13BMistral-7B-v0.1
Alpaca-52k3.743.743.743.745.405.405.405.405.355.355.355.35
AlpaGasus-1k3.633.633.633.634.704.704.704.706.066.066.066.06
LIMA-1k3.953.953.953.955.185.185.185.186.18
Alpaca-1k-longest3.963.963.963.965.325.325.325.325.805.805.805.80
Refined-Alpaca-1k-longest4.18¯¯4.18\underline{4.18}under¯ start_ARG 4.18 end_ARG6.096.006.006.006.00
 + NEFTune4.285.98¯¯5.98\underline{5.98}under¯ start_ARG 5.98 end_ARG6.18

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (11)

Changing response length does not affect quality.As shown in Table1, the LLMs fine-tuned on (Refined-)1k-longest lead to longer generation than most competitors.To test if longer replies are sufficient for higher scores on AlpacaEval 2.0, we increase the maximum number of generated tokens from the default 2048 (used for all baselines as well) to 4096.This makes the average response length of our best Llama-2-7B model (refined dataset with NEFTune) to increase from 1801 to 2478. However, this slightly degrades win rate (-0.05%).Similar small variations can be also observed for other models and architectures (see Table1).Then, length alone does not significantly influence the results on the benchmark.

MT-Bench evaluation.We show the score-based results on MT-Bench of LLMs fine-tuned on different instruction datasets in Table2.The baselines (Alpaca-52k, AlpaGasus-1k, and LIMA-1k) achieve scores below 4 when employing Llama-2-7B as the base model. Alpaca-1k-longest without refinement (3.96) already matches the best of them, and refining the 1,000 raw instructions yields a 4.18 MT-Bench score.Switching to Llama-2-13B and Mistral-7B-v0.1, we show consistent improvements of using Alpaca-1k-longest and its refined variants compared to baselines.For all base models, unlike what we observe in AlpacaEval 2.0, applying NEFTune does not consistently lead to stronger instruction-following capability, as indicated by MT-Bench scores.

4.3 Evaluation on factual knowledge benchmarks

In the following, we study how the models trained on small instruction datasets behave in tasks other than instruction following with an LLM as a judge, and the shortcomings it entails.For this, we evaluate them on a subset of the Open LLM benchmark: it includes six datasets, from which we exclude HellaSwag because it contains examples also present in the training set of LIMA-1k (see discussion in App.E) and GSM80K since all models fail to achieve non trivial performance, which assess several abilities of an LLM including commonsense reasoning, multitask knowledge and truthfulness, at various difficulty levels.

Fig.5 reports the results of the models fine-tuned from Llama-2-7B on the dataset derived from Alpaca and LIMA-1k (the corresponding evaluations for other architectures and Evol-Instruct-based datasets can be found in App.B.4).We observe that, on average over the datasets, IFT on Alpaca-52k yields marginal improvement over the base model, while both AlpaGasus-1k and 1k-longest give around a 1% increase. Significantly better results are achieved by LIMA-1k, with 55.9% vs 53.1% of the base model.However, the two models relying on Alpaca-Refined-1k-longest, without and with NEFTune, are the best performing ones with 56.4% and 56.5% (without and with NEFTune respectively).This suggests that the IFT dataset might have an effect beyond quality of user interactions. In fact, all LLMs are fine-tuned from the same base model, thus we can assume that they have the same factual knowledge, and the different performance is due to how well the alignment phase teaches the model how to follow the right steps to accomplish a given task.We hypothesize that using longer and more detailed instructions, which force the LLM to better capture the semantics of the task at hand, might positively influence the performance on quantitative (e.g. multiple choice questions answering) tasks as those in Open LLM.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (12)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (13)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (14)

5 Additional analyses of our models

While we uncover the effectiveness of fine-tuning on instructions with long responses, the reason for this success remains elusive. In the following we provide some insights about this phenomenon.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (15)

Comparison on generations of similar length.To further support the idea that the length of responses does not explain our models’ performance, we artificially increase the response length of the replies from Llama-2-7B models fine-tuned on Alpaca-52k and AlpaGasus-1k.This extension is achieved by ensuring that the end-of-sentence token does not appear until after the 150th token has been generated.Fig.6(c) shows that this adjustment makes both baselines output responses of similar length as our Alpaca-1k-longest.However, even in this case, both GPT-4 and PaLM-2 judges still significantly prefer our Alpaca-longest-1k model (Fig.6(a)), indicating that artificially increasing the number of generated tokens does not effectively enhance response quality.Meanwhile, we further test using prompting strategies to control the response length of all models, which might be a more natural choice than directly postponing the appearance of the end-of-sentence token. After extensive exploration with Mistral-7B-v0.1 models (note that none of the prompting strategies we tried are effective with Llama-2-7B models), we could increase the response length of Alpaca-52k and AlpaGasus-1k models by asking in the prompt to use N𝑁Nitalic_N paragraphs, and reduce the response length of our Alpaca-1k-longest by asking to “answer in as few words as possible”. When conducting head-to-head comparisons, we ensure that there is minimal variation in average length between the two candidates and have Alpaca-52k-prompt (197 tokens) vs Alpaca-1k-longest (184 tokens) and AlpaGasus-1k-prompt (143 tokens) vs Alpaca-1k-longest-prompt (158 tokens). The results in Fig.6(b) indicate that our Alpaca-1k-longest outperforms both length-controlled counterparts across the two judges, which aligns with the conclusion we obtain in Fig.6(a).

Length and win rate are anticorrelated during fine-tuning.We track the average length of replies over epochs when fine-tuning for one of our models. As shown in Fig.7, except for the early stage of fine-tuning, the response length progressively decreases while the win rate keeps improving.This indicates that the model does not simply learn to output long generations from long training examples but also to produce more refined structures.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (16)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (17)

Example generations.In Fig.8 we provide two examples of completions generated by our Llama-2-7B model fine-tuned on the Alpaca-1k-longest dataset. We see that the LLM provides organic and detailed responses.We provide an extended qualitative comparison to other models in App.D, where one can see that, for example, LIMA can sometimes lead to repetitive outputs while 1k-longest models tend to have a more engaging tone.

Additional comparisons.In App.B.6, we show how our approach to data selection works in concert with other sampling techniques that promote diversity. Moreover, we verify that our 1k-longest models remain effective on text summarization tasks (see results in Table4), in which concise answers are preferred. For space reasons, we defer to the appendix the comparison of our Alpaca-1k-longest to two additional baselines, AlpaGasus-9k and the dataset obtained improving Alpaca-52k with reflection-tuning in Li etal. (2023a). As shown in App.C.1 and App.C.2 respectively, our approach consistently outperforms both baselines.

6 Discussion

Quality of the instructions in IFT.Chen etal. (2023) and Zhou etal. (2023) argue that IFT requires high-quality training examples and use different proxies for quality to create the AlpaGasus and LIMA datasets.However, our experiments demonstrate that a simple heuristic for selecting training instructions, such as the length of the response, leads to better-performing models.It is important to note that length alone is not sufficient. For example, the LIMA training examples are on average twice as long as those in Alpaca-1k-longest. Additionally, we emphasize that length does not necessarily reflect quality, as illustrated by the lower scores given by GPT-3.5-Turbo to the examples in our Alpaca-1k-longest (Fig.13). This suggests that other factors come into play when determining the effectiveness of IFT datasets. As a result, it remains uncertain which specific components in the fine-tuning dataset are crucial for achieving the best model performance.

IFT can improve factuality.Gudibande etal. (2023) show the possibility of fine-tuning LLMs to imitate the style of ChatGPT. They achieve this by using ChatGPT’s responses as an IFT dataset, which can consist of up to 150 million tokens. Remarkably, both human evaluators and LLM-as-a-judge evaluators rate the responses generated by these fine-tuned models nearly as high as those generated by ChatGPT. However, this fine-tuning approach does not enhance, and in some cases even diminishes, the performance of these models on NLP benchmarks compared to the base model. A similar observation is made by Jha etal. (2023), who suggest that LIMA-1k (when used to fine-tune the MPT models from MosaicML (2023)) does not yield the same level of performance as Alpaca-52k on tasks that do not rely on automated evaluation by an LLM.In contrast, we demonstrate that IFT can lead to both a stronger preference from various LLMs serving as judges and improved performance on Open LLM tasks. However, it is key to carefully select the instruction dataset for this purpose.The question of systematically constructing optimal IFT datasets remains an open challenge.

Role of length bias of LLMs as judges.It is a relevant question whether training on the longest examples simply exploits the bias of LLMs as judges for longer replies rather than leading to inherently better models.Therefore, we have conducted experiments specifically designed to exclude this scenario, like equating the response length of different LLMs and the human evaluation.Moreover, we benchmark the ability of the models both with head-to-head comparisons, single-score evaluation, and on factual knowledge datasets.Since all these analyses show that our models outperform the baselines across datasets, tasks and evaluation methods, we consider them conclusive evidence that our results are not due to a length bias of the LLMs as judges.

Conclusions.In this work we have shown that using reply length as a heuristic can effectively pre-select instructions for LLMs alignment in SFT. Moreover, a straightforward refinement step is enough to create a dataset of only 1k instruction-response pairs which yields competitive results compared to complex alignment methods like RLHF and DPO.Thus, this approach constitutes an inexpensive yet strong baseline for future works on alignment.Our analysis also challenges the current understanding of high-quality IFT datasets and their impact on fine-tuned model performance in standard NLP benchmarks.We emphasize that a major aspect of alignment concerns mitigation of safety risks and ethical use of LLMs.We have not explored this aspect here, as it demands task-specific approaches.

Impact statement

This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here.

Acknowledgements

We thank the anonymous reviewers of ICML for insightful comments that have helped to improve the quality of the paper.M.A. was supported by the Google Fellowship and Open Phil AI Fellowship.

References

  • Abid etal. (2019)Abid, A., Abdalla, A., Abid, A., Khan, D., Alfozan, A., and Zou, J.Gradio: Hassle-free sharing and testing of ml models in the wild.arXiv preprint arXiv:1906.02569, 2019.
  • Anil etal. (2023)Anil, R., Dai, A.M., Firat, O., Johnson, M., Lepikhin, D., Passos, A., Shakeri, S., Taropa, E., Bailey, P., Chen, Z., etal.Palm 2 technical report.arXiv preprint arXiv:2305.10403, 2023.
  • Askell etal. (2021)Askell, A., Bai, Y., Chen, A., Drain, D., Ganguli, D., Henighan, T., Jones, A., Joseph, N., Mann, B., DasSarma, N., etal.A general language assistant as a laboratory for alignment.arXiv preprint arXiv:2112.00861, 2021.
  • Bai etal. (2022a)Bai, Y., Jones, A., Ndousse, K., Askell, A., Chen, A., DasSarma, N., Drain, D., Fort, S., Ganguli, D., Henighan, T., etal.Training a helpful and harmless assistant with reinforcement learning from human feedback.arXiv preprint arXiv:2204.05862, 2022a.
  • Bai etal. (2022b)Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A., McKinnon, C., etal.Constitutional ai: Harmlessness from ai feedback.arXiv preprint arXiv:2212.08073, 2022b.
  • Baumgartner etal. (2020)Baumgartner, J., Zannettou, S., Keegan, B., Squire, M., and Blackburn, J.The pushshift reddit dataset.In Proceedings of the international AAAI conference on web and social media, volume14, pp. 830–839, 2020.
  • Brown etal. (2020)Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., etal.Language models are few-shot learners.Advances in neural information processing systems, 33:1877–1901, 2020.
  • Cao etal. (2023)Cao, Y., Kang, Y., Wang, C., and Sun, L.Instruction mining: When data mining meets large language model finetuning, 2023.
  • Chen etal. (2023)Chen, L., Li, S., Yan, J., Wang, H., Gunaratna, K., Yadav, V., Tang, Z., Srinivasan, V., Zhou, T., Huang, H., etal.Alpagasus: Training a better alpaca with fewer data.arXiv preprint arXiv:2307.08701, 2023.
  • Chen etal. (2022)Chen, M., Chu, Z., Wiseman, S., and Gimpel, K.Summscreen: A dataset for abstractive screenplay summarization.In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 8602–8615, 2022.
  • Chiang etal. (2023)Chiang, W.-L., Li, Z., Lin, Z., Sheng, Y., Wu, Z., Zhang, H., Zheng, L., Zhuang, S., Zhuang, Y., Gonzalez, J.E., Stoica, I., and Xing, E.P.Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality, March 2023.URL https://lmsys.org/blog/2023-03-30-vicuna/.
  • Chung etal. (2022)Chung, H.W., Hou, L., Longpre, S., Zoph, B., Tay, Y., Fedus, W., Li, Y., Wang, X., Dehghani, M., Brahma, S., etal.Scaling instruction-finetuned language models.arXiv preprint arXiv:2210.11416, 2022.
  • Clark etal. (2018)Clark, P., Cowhey, I., Etzioni, O., Khot, T., Sabharwal, A., Schoenick, C., and Tafjord, O.Think you have solved question answering? try arc, the ai2 reasoning challenge.arXiv preprint arXiv:1803.05457, 2018.
  • Geng etal. (2023)Geng, X., Gudibande, A., Liu, H., Wallace, E., Abbeel, P., Levine, S., and Song, D.Koala: A dialogue model for academic research.Blog post, April 2023.URL https://bair.berkeley.edu/blog/2023/04/03/koala/.
  • Gudibande etal. (2023)Gudibande, A., Wallace, E., Snell, C., Geng, X., Liu, H., Abbeel, P., Levine, S., and Song, D.The false promise of imitating proprietary llms.arXiv preprint arXiv:2305.15717, 2023.
  • Hendrycks etal. (2020)Hendrycks, D., Burns, C., Basart, S., Zou, A., Mazeika, M., Song, D., and Steinhardt, J.Measuring massive multitask language understanding.In International Conference on Learning Representations, 2020.
  • Huang etal. (2022)Huang, J., Gu, S.S., Hou, L., Wu, Y., Wang, X., Yu, H., and Han, J.Large language models can self-improve.arXiv preprint arXiv:2210.11610, 2022.
  • Huang etal. (2021)Huang, L., Cao, S., Parulian, N., Ji, H., and Wang, L.Efficient attentions for long document summarization.In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1419–1436, 2021.
  • Ivison etal. (2023)Ivison, H., Wang, Y., Pyatkin, V., Lambert, N., Peters, M., Dasigi, P., Jang, J., Wadden, D., Smith, N.A., Beltagy, I., etal.Camels in a changing climate: Enhancing lm adaptation with tulu 2.arXiv preprint arXiv:2311.10702, 2023.
  • Jain etal. (2023)Jain, N., Chiang, P.-y., Wen, Y., Kirchenbauer, J., Chu, H.-M., Somepalli, G., Bartoldson, B.R., Kailkhura, B., Schwarzschild, A., Saha, A., etal.Neftune: Noisy embeddings improve instruction finetuning.arXiv preprint arXiv:2310.05914, 2023.
  • Jha etal. (2023)Jha, A., Havens, S., Dohmann, J., Trott, A., and Portes, J.Limit: Less is more for instruction tuning across evaluation paradigms.arXiv preprint arXiv:2311.13133, 2023.
  • Jiang etal. (2023)Jiang, A.Q., Sablayrolles, A., Mensch, A., Bamford, C., Chaplot, D.S., Casas, D. d.l., Bressand, F., Lengyel, G., Lample, G., Saulnier, L., etal.Mistral 7b.arXiv preprint arXiv:2310.06825, 2023.
  • Lee etal. (2023)Lee, H., Phatale, S., Mansoor, H., Lu, K., Mesnard, T., Bishop, C., Carbune, V., and Rastogi, A.Rlaif: Scaling reinforcement learning from human feedback with ai feedback.arXiv preprint arXiv:2309.00267, 2023.
  • Li etal. (2023a)Li, M., Chen, L., Chen, J., He, S., Huang, H., Gu, J., and Zhou, T.Reflection-tuning: Data recycling improves llm instruction-tuning.arXiv preprint arXiv:2310.11716, 2023a.
  • Li etal. (2023b)Li, X., Zhang, T., Dubois, Y., Taori, R., Gulrajani, I., Guestrin, C., Liang, P., and Hashimoto, T.B.Alpacaeval: An automatic evaluator of instruction-following models.https://github.com/tatsu-lab/alpaca_eval, 2023b.
  • Lin etal. (2022)Lin, S., Hilton, J., and Evans, O.Truthfulqa: Measuring how models mimic human falsehoods.In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 3214–3252, 2022.
  • Liu etal. (2023)Liu, W., Zeng, W., He, K., Jiang, Y., and He, J.What makes good data for alignment? a comprehensive study of automatic data selection in instruction tuning.In The Twelfth International Conference on Learning Representations, 2023.
  • Lu etal. (2023)Lu, K., Yuan, H., Yuan, Z., Lin, R., Lin, J., Tan, C., Zhou, C., and Zhou, J.# instag: Instruction tagging for analyzing supervised fine-tuning of large language models.In The Twelfth International Conference on Learning Representations, 2023.
  • MosaicML (2023)MosaicML.Introducing mpt-7b: A new standard for open-source, commercially usable LLMs, 2023.URL www.mosaicml.com/blog/mpt-7b.www.mosaicml.com/blog/mpt-7b, accessed: 2023-08-02.
  • Mukherjee etal. (2023)Mukherjee, S., Mitra, A., Jawahar, G., Agarwal, S., Palangi, H., and Awadallah, A.Orca: Progressive learning from complex explanation traces of gpt-4.arXiv preprint arXiv:2306.02707, 2023.
  • OpenAI (2023)OpenAI.Gpt-4 technical report.arXiv preprint arXiv:2303.08774, 2023.
  • Ouyang etal. (2022)Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., etal.Training language models to follow instructions with human feedback.Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
  • Pan etal. (2023)Pan, L., Saxon, M., Xu, W., Nathani, D., Wang, X., and Wang, W.Y.Automatically correcting large language models: Surveying the landscape of diverse self-correction strategies.arXiv preprint arXiv:2308.03188, 2023.
  • Peng etal. (2023)Peng, B., Li, C., He, P., Galley, M., and Gao, J.Instruction tuning with gpt-4.arXiv preprint arXiv:2304.03277, 2023.
  • Rafailov etal. (2023)Rafailov, R., Sharma, A., Mitchell, E., Ermon, S., Manning, C.D., and Finn, C.Direct preference optimization: Your language model is secretly a reward model.arXiv preprint arXiv:2305.18290, 2023.
  • Sakaguchi etal. (2021)Sakaguchi, K., Bras, R.L., Bhagavatula, C., and Choi, Y.Winogrande: An adversarial winograd schema challenge at scale.Communications of the ACM, 64(9):99–106, 2021.
  • Shaham etal. (2022)Shaham, U., Segal, E., Ivgi, M., Efrat, A., Yoran, O., Haviv, A., Gupta, A., Xiong, W., Geva, M., Berant, J., etal.Scrolls: Standardized comparison over long language sequences.In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 12007–12021, 2022.
  • Singhal etal. (2023)Singhal, P., Goyal, T., Xu, J., and Durrett, G.A long way to go: Investigating length correlations in rlhf.arXiv preprint arXiv:2310.03716, 2023.
  • Taori etal. (2023)Taori, R., Gulrajani, I., Zhang, T., Dubois, Y., Li, X., Guestrin, C., Liang, P., and Hashimoto, T.B.Stanford alpaca: An instruction-following llama model.https://github.com/tatsu-lab/stanford_alpaca, 2023.
  • Teknium (2023)Teknium.Openhermes 2.5: An open dataset of synthetic data for generalist llm assistants, 2023.URL https://huggingface.co/datasets/teknium/OpenHermes-2.5.
  • Touvron etal. (2023)Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., etal.Llama 2: Open foundation and fine-tuned chat models.arXiv preprint arXiv:2307.09288, 2023.
  • Wang etal. (2022)Wang, Y., Kordi, Y., Mishra, S., Liu, A., Smith, N.A., Khashabi, D., and Hajishirzi, H.Self-instruct: Aligning language model with self generated instructions.arXiv preprint arXiv:2212.10560, 2022.
  • Wang etal. (2023)Wang, Y., Ivison, H., Dasigi, P., Hessel, J., Khot, T., Chandu, K.R., Wadden, D., MacMillan, K., Smith, N.A., Beltagy, I., etal.How far can camels go? exploring the state of instruction tuning on open resources.arXiv preprint arXiv:2306.04751, 2023.
  • Wei etal. (2022)Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q.V., Zhou, D., etal.Chain-of-thought prompting elicits reasoning in large language models.Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
  • Xu etal. (2023)Xu, C., Sun, Q., Zheng, K., Geng, X., Zhao, P., Feng, J., Tao, C., and Jiang, D.Wizardlm: Empowering large language models to follow complex instructions.arXiv preprint arXiv:2304.12244, 2023.
  • Xu etal. (2022)Xu, H., Chen, Y., Du, Y., Shao, N., Wang, Y., Li, H., and Yang, Z.Zeroprompt: scaling prompt-based pretraining to 1,000 tasks improves zero-shot generalization.arXiv preprint arXiv:2201.06910, 2022.
  • Yuan etal. (2024)Yuan, W., Pang, R.Y., Cho, K., Sukhbaatar, S., Xu, J., and Weston, J.Self-rewarding language models.arXiv preprint arXiv:2401.10020, 2024.
  • Zellers etal. (2019)Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., and Choi, Y.Hellaswag: Can a machine really finish your sentence?In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 4791–4800, 2019.
  • Zeng etal. (2023)Zeng, Z., Yu, J., Gao, T., Meng, Y., Goyal, T., and Chen, D.Evaluating large language models at evaluating instruction following.arXiv preprint arXiv:2310.07641, 2023.
  • Zhao etal. (2023)Zhao, Y., Yu, B., Hui, B., Yu, H., Huang, F., Li, Y., and Zhang, N.L.A preliminary study of the intrinsic relationship between complexity and alignment.arXiv preprint arXiv:2308.05696, 2023.
  • Zheng etal. (2023)Zheng, L., Chiang, W.-L., Sheng, Y., Zhuang, S., Wu, Z., Zhuang, Y., Lin, Z., Li, Z., Li, D., Xing, E., etal.Judging llm-as-a-judge with mt-bench and chatbot arena.arXiv preprint arXiv:2306.05685, 2023.
  • Zhong etal. (2021)Zhong, M., Yin, D., Yu, T., Zaidi, A., Mutuma, M., Jha, R., Hassan, A., Celikyilmaz, A., Liu, Y., Qiu, X., etal.Qmsum: A new benchmark for query-based multi-domain meeting summarization.In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 5905–5921, 2021.
  • Zhou etal. (2023)Zhou, C., Liu, P., Xu, P., Iyer, S., Sun, J., Mao, Y., Ma, X., Efrat, A., Yu, P., YU, L., Zhang, S., Ghosh, G., Lewis, M., Zettlemoyer, L., and Levy, O.LIMA: Less is more for alignment.In Thirty-seventh Conference on Neural Information Processing Systems, 2023.

Appendix A Experimental details

A.1 IFT datasets

This section contains a list of instruction fine-tuning datasets that appear in our experiments, along with relevant information.

  • Alpaca(Taori etal., 2023) contains 52k synthetic examples generated by explicitly giving the requirement instruction generation to the text-davinci-003 model. Although the created dataset is intended to be varied, a thorough examination reveals that it is heavily US-centric. It is also discovered that the original version has numerous issues that affect its quality and suitability for training a trustworthy language model. These issues includes hallucinations, merged instructions, empty outputs, impractical instructions like generating images, wrong answers, and non-sensical instructions, etc.

  • AlpaGasus-1k/9k(Chen etal., 2023) contains 1k/9k high-quality examples filtered from the original Alpaca-52k dataset. It suggests implementing data selection by means of strong LLMs, such as ChatGPT, to automatically detect and filter out low-quality data. By doing this, they leave out problematic samples, which endanger the effectiveness of refined models.

  • Recycled-Alpaca(Li etal., 2023a) comprises of 52k enhanced examples based on Alpaca-52k. Given the initial basic dataset, a high-quality version of each data point is generated using an Oracle model, such as chatGPT. However, a typical issue with using LLMs as judges is the inability to produce different results. To address this potential issue, inspired by Chain-of-Thought prompting, numerous particular criterias are proposed for the Oracle model to follow, and then strong target LMs respond to those precise requirements with critical responses. The responses to these criteria can then be used as bridges (chains of thought) to create new, satisfied instruction-response combinations.

  • LIMA(Zhou etal., 2023) collects a dataset of 1000 prompts and responses for training, with the outputs stylistically aligned but the inputs different. It also provides an open-source test set of 300 prompts and a development set of 50. Curated from multiple sources, LIMA is primarily divided among community Q&A websites like Stack Exchange, wikiHow, and the Pushshift Reddit Dataset(Baumgartner etal., 2020), as well as manually created examples. In terms of Q&A communities, frequently upvoted answers on Reddit are typically hilarious or trolling, requiring more manual effort to align responses that adhere to the proper style. In contrast, answers from Stack Exchange and wikiHow are well-aligned with the behavior of a helpful chat assistant. Human-authored examples are used to boost the diversity of dataset.

  • Evol-Instruct (WizardLM)(Xu etal., 2023) contains 70k training examples with varying complexity and 218 test instances. The training dataset is initially initialized using Alpaca’s 52k instruction data. After iteratively completing M=4𝑀4M=4italic_M = 4 evolutions, the dataset has 250k instructions. More specifically, for each instruction in each round of evolution, one evolving prompt from total six new prompts (i.e., five from in-depth evolving and one from in-breadth evolving) is selected with equal probability. Then, ChatGPT is used to produce answers for each instruction, yielding 52×4×3=624k5243624k52\times 4\times 3=624\text{k}52 × 4 × 3 = 624 k instruction-response pairs. Finally, the Evol-Instruct dataset is created by picking a subset of 70k instructions. 218 test instructions are collected from diverse sources including online opensource projects, platforms, and forums. This test set is primarily a union of 29 distinct skills identified among real-world human instructions, such as Coding Generation & Debugging, Reasoning, Math, Writing, Complex Formats, Extensive Disciplines, and so on.

  • Vicuna(Chiang etal., 2023) divides 80 test instructions into 8 question categories, including Fermi problems, commonsense, roleplay scenarios, coding/math/writing tasks, counterfactual, knowledge, and generic, to evaluate various aspects of a chatbot’s performance. Vicuna has been demonstrated to mostly include instructions of low difficulty and complexity(Xu etal., 2023).

  • Self-Instruct(Wang etal., 2022) has 252 human-authored test instructions with 1 handcrafted output per instruction. Self-Instrction test set is created to better reflect the practical value of instruction-following models. The authors were motivated to curate instructions of different domains ranging from email writing and social media to productivity tools and programming. Authors also deliberately diversify the styles and formats of tasks, such as including instructions of different lengths and considering input/output that takes the form of bullet points, tables, codes, equations, etc.

  • Koala(Geng etal., 2023) consists of 180 real user queries that were posted on the Internet. These user-initiated queries cover a wide range of subjects, typically have a conversational tone, and are probably more indicative of the practical applications of chat-based systems. Queries with a BLEU score of more than 20% with any example from our training set are filtered away in order to reduce the possibility of test-set leaking. Prompts pertaining to code and languages other than English are also excluded because the crowd workers, who make up the pool of raters, are unable to accurately examine the answers to these questions.

A.2 Training hyperparameters

This section lists the hyperparameters necessary for reproducing our work. Our experiments are built upon FastChat framework(Zheng etal., 2023). In particular, we follow the training configuration as reported inTaori etal. (2023) to fine-tune the base model on full IFT datasets like Alpaca-52k and Evol-Instruct-70k, while we refer to LIMA(Zhou etal., 2023) and AlpaGasus(Chen etal., 2023) when fine-tuning the base model on IFT datasets with 1k and 9k training examples, respectively.In addition to existing experimental setups in prior work, we adopt the recently proposed NEFTune augmentation for our (Refined-)Alpaca-1k-longest experiments. We have neftune_noise_level set to 5 for Llama-2-7B, while for Mistral-7B-v0.1 and Llama-2-13B it is set to 3. It should be noted that we use 4 ×\times× 40G A100 to finetune Llama-2-7B and 4 ×\times× 80G A100 to finetune Mistral-7B-v0.1 and Llama-2-13B. We present the detailed training hyperparameters in Table3.

DatasetsData Size# GPUsEpochsLRLR SchedulerBatch SizeContext Win. Len.WDWarmup Rate
Llama-2-7B
Evol-Instruct-70k70k432e-5Cosine1285120.00.3
Alpaca-52k52k432e-5Cosine1285120.00.3
AlpaGasus-9k9k432e-5Cosine1285120.00.3
Alpaca-9k-longest9k432e-5Cosine1285120.00.3
AlpaGasus-1k1k4151e-5Linear12820480.10.0
LIMA-1k1k4151e-5Linear12820480.10.0
Alpaca-1k-longest1k4151e-5Linear12820480.10.0
Evol-Instruct-AlpaGasus-1k1k4151e-5Linear12820480.10.0
Evol-Instruct-1k-longest1k4151e-5Linear12820480.10.0
Mistral-7B-v0.1
Alpaca-52k52k434e-6Cosine1285120.00.3
AlpaGasus-1k1k4152e-6Linear12820480.10.0
LIMA-1k1k4152e-6Linear12820480.10.0
Alpaca-1k-longest1k4152e-6Linear12820480.10.0
Llama-2-13B
Alpaca-52k52k451e-5Cosine1285120.00.3
AlpaGasus-1k1k4151e-5Linear12820480.10.0
LIMA-1k1k4151e-5Linear12820480.10.0
Alpaca-1k-longest1k4151e-5Linear12820480.10.0

A.3 Evaluation details

Evaluation metrics for head-to-head comparisons.

Since automated evaluation based on powerful LLMs offers superior scalability, explainability and reproducibility than human evaluation, we apply an LLM that has high human preference agreement as the judge to evaluate the target model (e.g., Llama-2-7B fine-tuned on Alpaca-1k-longest) and compare it with a baseline model (e.g., GPT-4-Turbo). We append both models’ outputs in the input instruction to the LLM judge, followed by a request to the judge which prompts the model to rate the responses with a score between 1 and 10. Since there exists position bias within LLM-based automated evaluation(Zheng etal., 2023), we run evaluation on both orders (i.e., placing the response of the target model before/after the baseline model’s response) and calculate the win rate (tie is allowed).

LLM-as-a-judge.

Given their good agreement with human evaluators shown in LLMBar(Zeng etal., 2023), we decide to adopt GPT-4 (i.e., GPT-4-0613) and PaLM2 (i.e., text-bison@002) as the LLM judges to appropriately assess the instruction-following performance of instruction-tuned models.

Evaluation prompt for GPT4- and PaLM2-as-a-judge.

We adopt the same evaluation prompt for both GPT4- and PaLM2-as-a-judge as what AlpaGasus(Chen etal., 2023) uses, which is also the prompt for evaluation used in the original Vicuna work(Chiang etal., 2023). We provide the detailed form of the prompt in Fig.9.

Human evaluation.

Model-based head-to-head evaluation is notorious for its implicit preference for more verbose and engaging answers. Thus we conduct a human-based evaluation, in which the rule emphasizes that human annotators should choose the better answer from two candidates based solely on relevancy and helpfulness, and ignore potential superficial features, such as an engaging tone and response length. In particular, we sample 100 random instructions from evaluation datasets we use in model-based head-to-head comparisons and generate the responses for the baseline model, Llama-2-7B-Alpaca-52k, and the target model, Llama-2-7B-Alpaca-1k-longest. To improve the efficiency of human annotation and reach more annotators, we designed a demo for the user study (see the full template in Fig.11), building with Gradio(Abid etal., 2019), an interactive design tool.

AlpacaEval 2.0.

We apply the AlpacaEval 2.0 benchmark in our experiments since it provides transferable comparisons, which is impossible to achieve in head-to-head evaluation. AlpacaEval 2.0 provides 805 test instructions, on which we generate new responses using the target model, and then calculate the score by competing with the baseline model (i.e., GPT-4-Turbo) judged by a designated automatic evaluator.

MT-Bench.

The test dataset of this benchmark(Zheng etal., 2023) covers 8 common categories of user prompts: coding, math, reasoning, extraction, roleplay, writing, humanities/social science, and STEM. It contains 80 questions, all of which are high-quality and challenging, designed to assess models’ abilities to engage in multi-turn conversations and follow instructions.

Open LLM Leaderboard.

Several multiclass classification datasets are used to compute the models ranking: ARC(Clark etal., 2018), MMLU(Hendrycks etal., 2020), TruthfulQA(Lin etal., 2022), Winogrande(Sakaguchi etal., 2021), HellaSwag(Zellers etal., 2019). The combination of datasets widely measures an LLM’s capacity to react to factual queries and reasoning challenges, and we use this benchmark to compare the model’s change in factual capabilities before and after instruction fine-tuning.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (18)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (19)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (20)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (21)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (22)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (23)

Appendix B Additional results

B.1 Scores of Alpaca-1k-longest according to GPT-3.5-Turbo

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (24)

In Fig.13 we show the score distribution from Chen etal. (2023) for the 1k longest examples compared to those of AlpaGasus-1k (i.e. that highest scoring ones): we see that the overlap between the two datasets is minimal, and most of the longest examples have score of 3.5.Interestingly, this suggests that GPT-3.5-Turbo prefers longer responses when used as a judge, e.g. in the AlpacaEval 2.0 benchmark, while favors different features when asked to score the quality of the instruction-response pairs in Alpaca.

B.2 PaLM-2-as-a-judge details

We present detailed preference evaluation results using PaLM2-as-a-judge on an array of Llama-2-7B-based models in Fig.14. Moreover, we show the improvement given by the refined dataset in Fig.15. In both cases the observations are consistent with what obtained with GPT-4 as judge (see Fig.2 and Fig.4 respectively).

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (25)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (26)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (27)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (28)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (29)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (30)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (31)

B.3 Preference evaluation on Mistral-7B-v0.1 and LLaMA-2-13B

This section contains the average preference evaluation results on Mistral-7B-v0.1 model and Llama-2-13B model over 5 evaluation sets (i.e., LIMA, Vicuna, Koala, WizardLM, and Self-Instruct) as shown in Fig.16 and Fig.17.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (32)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (33)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (34)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (35)

B.4 Open LLM results on Mistral-7B-v0.1, LLaMA-2-13B, Evol-Instruct-70k

This section contains the evaluation results of Mistral-7B-v0.1 model and Llama-2-13B model on (Fig.18) and of Llama-2-7B fine-tuned on Evol-Instruct-based datasets (Fig.19) over the Open LLM benchmark.

Base model: Mistral-7B-v0.1
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (36)
Base model: Llama-2-13B
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (37)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (38)

B.5 Empirical proof of the intuition behind utilizing longer examples for IFT

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (39)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (40)

To support the claim that longer responses are harder to fit, we compare the progress of the training loss over epochs of 1k-longest to 1k-shortest and 1k-random, i.e., the subsets containing the 1k shortest or 1k arbitrarily long responses, respectively. For both Alpaca and Evol-Instruct, we fine-tune Llama-2-7B on each split: Fig.20(a) and Fig.20(b) show that the loss, normalized by its initial value to make it comparable across training sets, converges more slowly when using 1k-longest compared to other subsets. This confirms our hypothesis that longer instructions are harder to fit and thus provide more supervision signals and lead to better generalization. Moreover, Table5and Table6 show the resulting evaluation results using PaLM2 as the judge and MT-Bench, respectively, both of which align with our intuition.

We verify the intuition that fine-tuning using longer examples makes the model more effective in capturing long-distance semantic connections on three long-form summarization tasks: GovReport(Huang etal., 2021), SummScreen(Chen etal., 2022), and QMSum(Zhong etal., 2021), from the SCROLLS(Shaham etal., 2022) benchmark. We measure the quality of the generated summary using the ROUGE-1, ROUGE-2, and ROUGE-L scores. Detailed results are shown in Table4. Our 1k-longest model always achieves the best (i.e., in SummScreen) or the 2nd best (i.e., in GovReport and QMSum) performance. It validates that fine-tuning using long examples maintains strong summarization capability, while showing superior performance on long-form tasks.

GovReportSummScreenQMSum
ModelsROUGE-1ROUGE-2ROUGE-LAvg.ROUGE-1ROUGE-2ROUGE-LAvg.ROUGE-1ROUGE-2ROUGE-LAvg.
Llama-2-7B (base)4.584.584.584.581.661.661.661.663.223.223.223.223.153.153.153.1513.1813.1813.1813.182.062.062.062.069.779.779.779.778.348.348.348.3419.1619.1619.1619.165.375.375.375.3714.8914.8914.8914.8913.1413.1413.1413.14
Alpaca-52k20.9120.9120.9120.918.608.608.608.6013.0513.0513.0513.0514.1914.1914.1914.1923.8023.8023.8023.803.373.373.373.3714.2814.2814.2814.2813.8213.8213.8213.8226.9426.9426.9426.946.2818.3517.19
AlpaGasus-1k23.028.9213.6815.2126.40¯¯26.40\underline{26.40}under¯ start_ARG 26.40 end_ARG3.58¯¯3.58\underline{3.58}under¯ start_ARG 3.58 end_ARG15.06¯¯15.06\underline{15.06}under¯ start_ARG 15.06 end_ARG15.01¯¯15.01\underline{15.01}under¯ start_ARG 15.01 end_ARG27.09¯¯27.09\underline{27.09}under¯ start_ARG 27.09 end_ARG5.745.745.745.7418.1018.1018.1018.1016.9816.9816.9816.98
Alpaca-1k-longest22.20¯¯22.20\underline{22.20}under¯ start_ARG 22.20 end_ARG8.85¯¯8.85\underline{8.85}under¯ start_ARG 8.85 end_ARG13.35¯¯13.35\underline{13.35}under¯ start_ARG 13.35 end_ARG14.80¯¯14.80\underline{14.80}under¯ start_ARG 14.80 end_ARG26.563.8115.4515.2727.405.80¯¯5.80\underline{5.80}under¯ start_ARG 5.80 end_ARG18.16¯¯18.16\underline{18.16}under¯ start_ARG 18.16 end_ARG17.12¯¯17.12\underline{17.12}under¯ start_ARG 17.12 end_ARG

Model1k-longest winsTie1k-longest loses
Base dataset: Alpaca-52k
1k-shortest97.02.62.62.62.60.40.40.40.4
1k-random72.118.018.018.018.09.99.99.99.9
Base dataset: Evol-Instruct-70k
1k-shortest93.44.94.94.94.91.71.71.71.7
1k-random39.729.129.129.129.131.231.231.231.2
Base dataset: Open-Hermes-1M
1k-shortest95.93.53.53.53.50.60.60.60.6
1k-random84.310.410.410.410.45.35.35.35.3
ModelAlpaca-52kEvol-Instruct-70kOpen-Hermes-1M
1k-shortest1.781.781.781.782.342.342.342.341.461.461.461.46
1k-random3.743.743.743.744.034.034.034.034.064.064.064.06
1k-longest3.964.274.18

B.6 Diversity of training examples

To examine the effect of data diversity, we test our approach on the Open-Hermes-2.5(Teknium, 2023) dataset, which includes data (around 1M instructions) from different sources. Table7 shows that selecting the 1k-longest with stratified sampling, i.e., preserving diverse data sources, results in slightly better performance on the MT-Bench than uniform sampling. This shows the importance of preserving the diversity of the original dataset. However, while this is simple to do for Open-Hermes-2.5 (where the source of data is available), it is not straightforward to control on other datasets, such as Alpaca without manual inspections. Finally, the same observation holds with Mistral-7B-v0.1 as the base model, where our 1k-longest with stratified sampling achieves results very close to the entire dataset, which is 1000x larger.

ModelMT-Bench Score
Base model: Llama-2-7B
Open-Hermes-1k-longest (uniform sampling)4.184.184.184.18
Open-Hermes-1k-longest (stratified sampling)4.314.314.314.31
Base model: Mistral-7B-v0.1
Open-Hermes-1k-longest (uniform sampling)6.836.836.836.83
Open-Hermes-1k-longest (stratified sampling)6.976.976.976.97
Open-Hermes-1M7.227.227.227.22

Appendix C Comparison to additional baselines

C.1 AlpaGasus-9k

In this section, we validate the advantage of length heuristics by comparing Alpaca-9k-longest with AlpaGasus-9k, which is the best filtered subset from Alpaca-52k in the AlpaGasus paper(Chen etal., 2023). The detailed experimental results are shown in Fig.21(a), where Alpaca-9k-longest consistently outperforms AlpaGasus-9k in 5 evaluation sets. We further show comparisons between Alpaca-1k-longest and AlpaGasus-9k in Fig.21(b), which also supports our main claim: length is a strong criterion for constructing instruction fine-tuning dataset. Details of experimental setup can be seen in Table3.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (41)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (42)

C.2 Reflection-tuning

In this section, we show the advantage of proposed introspection technique by comparing it with reflection-tuning(Li etal., 2023a) on Llama-2-7B and Llama-2-13B models. We present experimental results on the Open LLM benchmark and AlpacaEval 2.0 in Table8.

Models# SFT dataARCHellaSwagMMLUTruthfulQAWinograndeAverageAlpacaEval 2.0Avg. Length
Llama-2-7B052.9952.9952.9952.9978.6478.6478.6478.6446.5646.5646.5646.5638.9738.9738.9738.9773.7273.7273.7273.7258.1858.1858.1858.18//
Llama-2-7B-Alpaca-52k52k53.9253.9253.9253.9278.8278.8278.8278.8247.0547.0547.0547.0540.3240.3240.3240.3271.8271.8271.8271.8258.3958.3958.3958.392.742.742.742.74586586586586
Llama-2-7B-Recycled-Alpaca-52k*52k53.9253.9253.9253.9277.6877.6877.6877.6847.5545.5545.5545.5545.5571.8271.8271.8271.8259.3059.3059.3059.305.935.935.935.931470147014701470
Llama-2-7B-Refined-Alpaca-1k-L1k56.7480.2346.8246.8246.8246.8249.5972.4561.176.001732173217321732
Llama-2-13B059.6459.6459.6459.6482.1582.1582.1582.1555.6355.6355.6355.6336.9236.9236.9236.9276.0976.0976.0976.0962.0962.0962.0962.09//
Llama-2-13B-Alpaca-52k52k59.7359.7359.7359.7383.0883.0883.0883.0853.8753.8753.8753.8739.9839.9839.9839.9872.7772.7772.7772.7761.2461.2461.2461.243.903.903.903.90556556556556
Llama-2-13B-Recycled-Alpaca-52k*52k58.7058.7058.7058.7080.8080.8080.8080.8053.1153.1153.1153.1143.12????
Llama-2-13B-Refined-Alpaca-1k-L1k61.9583.8855.8641.7441.7441.7441.7475.8563.868.441646164616461646

Appendix D Case study

This section consists of ten test instructions and corresponding responses of Llama-2-7B (Fig.22 and Fig.23), Mistral-7B-v0.1 (Fig.24), and Llama-2-13B (Fig.25 and Fig.26) models fine-tuned on Alpaca-1k-longest, AlpaGasus-1k, Alpaca-52k, and LIMA-1k datasets. Details of training hyperparameters are shown in Table3. We add detailed comments for qualitative analysis on responses generated by Llama-2-7B in SectionD.1. We omit detailed analysis for Mistral-7B-v0.1 and Llama-2-13B since we make similar observations as for Llama-2-7B.

D.1 Detailed comments on Llama-2-7B examples

Example #1: generate an itinerary in Switzerland.

  • Alpaca-1k-longest provides a well-structured and detailed itinerary for a 5-day trip to Switzerland, starting from Basel. It includes a variety of activities, such as visiting museums, hiking, exploring towns, and enjoying local cuisine. It also suggests different modes of transportation, such as trains and cable cars, which are common in Switzerland. Its answer is relevant, accurate, and helpful. However it mentions a “famous Meierihne cheese”, which does not exist at all. We believe this hallucination happens because of the knowledge capabilities of the base model.

  • AlpaGasus-1k also provides a well-structured response and includes a variety of activities, it is slightly less detailed than Alpaca-1k-longest’s response. For example, in Interlaken, AlpaGasus-1k suggests visiting popular hiking destinations but did not provide any information about what one might see or do there. However, AlpaGasus-1k does a good job of suggesting a variety of activities and destinations, making the itinerary interesting and diverse.

  • Alpaca-52k’s answer is less detailed and less helpful. The assistant suggested visiting the same cities on multiple days, which is not efficient or practical for a 5-day trip. The assistant also did not provide specific activities or places to visit in each city, which makes the answer less useful for someone planning a trip.

  • LIMA-1k’s answer is cut off and does not cover the full 5 days. It also repeats the same dining and nightlife options for each day, which is not very helpful or realistic.

Example #2: give an inspiring speech as a pirate captain.

  • Alpaca-1k-longest provides excellent responses to this question. It uses appropriate pirate language and provides motivating speeches that would encourage a pirate crew to search for hidden treasure. The response is relevant, accurate, and detailed, providing a vivid picture of the adventure and potential rewards.

  • AlpaGasus-1k’s response is shorter and less detailed, but still motivational and in line with the question.

  • Alpaca-52k’s response is also motivational and uses appropriate language, but is less detailed and less vivid in its description of the journey and the treasure.

  • LIMA-1k also provides excellent responses to this question. It uses appropriate pirate language and provides motivating speeches that would encourage a pirate crew to search for hidden treasure. The response is relevant, accurate, and detailed, providing a vivid picture of the adventure and potential rewards.

Example #3: write a code snippet to validate an email address.

  • Alpaca-1k-longest provides a correct regular expression for validating an email address in Python and also explained what each part of the expression does. The explanation was clear and concise, making it easy to understand how the regular expression works.

  • AlpaGasus-1k also provides a correct regular expression for validating an email address. However, there is no explanation or context provided, which might make it difficult for someone unfamiliar with regular expressions to understand.

  • Alpaca-52k’s answer is also correct and accurate, but lacks the detailed explanation

  • LIMA-1k’s regular expression is incorrect and does not match the standard email format. The explanation provided by LIMA-1k is also incorrect and confusing, as it does not correctly explain what each part of the regular expression does.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (43)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (44)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (45)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (46)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (47)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (48)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (49)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (50)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (51)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (52)

Appendix E Data contamintation on LIMA-1k

With over 240k how-to articles covering a wide range of topics, wikiHow is an online publication in the style of a wiki, where articles are frequently regarded as high-quality content. LIMA(Zhou etal., 2023) contains 200 wikiHow examples. The article’s title serves as a prompt (e.g., “How to Cook an Omelet?”) and the body text as an answer. HellaSwag(Zellers etal., 2019) from Open LLM leaderboard also includes wikiHow articles to enhance the content diversity. By cross validating the evaluation set of the HellaSwag task and the training set of LIMA, we find that the style and format of 200 wikiHow examples in LIMA are highly similar to that of in HellaSwag evaluation set. Also, surprisingly, we notice that multiple examples (e.g., “How to get a free room upgrade in las vegas?”, “How to teach a child to use scissors?”, “How to handle poking wires on braces?”, etc.) appear in both datasets, which is a strong signal of data contamination. The performance of LIMA-1k model on the HellaSwag task is also suspiciously higher than the other baselines as shown in Fig.27.

A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (53)
A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning (2024)

References

Top Articles
Latest Posts
Article information

Author: Kareem Mueller DO

Last Updated:

Views: 5342

Rating: 4.6 / 5 (66 voted)

Reviews: 81% of readers found this page helpful

Author information

Name: Kareem Mueller DO

Birthday: 1997-01-04

Address: Apt. 156 12935 Runolfsdottir Mission, Greenfort, MN 74384-6749

Phone: +16704982844747

Job: Corporate Administration Planner

Hobby: Mountain biking, Jewelry making, Stone skipping, Lacemaking, Knife making, Scrapbooking, Letterboxing

Introduction: My name is Kareem Mueller DO, I am a vivacious, super, thoughtful, excited, handsome, beautiful, combative person who loves writing and wants to share my knowledge and understanding with you.