FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions (2024)

Orion Wellerι𝜄{}^{\hskip 0.70004pt\hskip 0.70004pt{\color[rgb]{0,0,1}\boldsymbol{\iota}}}start_FLOATSUPERSCRIPT bold_italic_ι end_FLOATSUPERSCRIPTBenjamin Changι𝜄{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\iota}}start_FLOATSUPERSCRIPT bold_italic_ι end_FLOATSUPERSCRIPTSean MacAvaneyλ𝜆{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\lambda}}start_FLOATSUPERSCRIPT bold_italic_λ end_FLOATSUPERSCRIPTKyle Loα𝛼{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\alpha}}start_FLOATSUPERSCRIPT bold_italic_α end_FLOATSUPERSCRIPT
Arman Cohanγα𝛾𝛼{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\gamma\hskip 0.70004pt\alpha}}start_FLOATSUPERSCRIPT bold_italic_γ bold_italic_α end_FLOATSUPERSCRIPTBenjamin Van Durmeι𝜄{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\iota}}start_FLOATSUPERSCRIPT bold_italic_ι end_FLOATSUPERSCRIPTDawn Lawrieι𝜄{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\iota}}start_FLOATSUPERSCRIPT bold_italic_ι end_FLOATSUPERSCRIPTLuca Soldainiα𝛼{}^{\hskip 0.70004pt\color[rgb]{0,0,1}\boldsymbol{\alpha}}start_FLOATSUPERSCRIPT bold_italic_α end_FLOATSUPERSCRIPT
ι𝜄{}^{\color[rgb]{0,0,1}\iota\hskip 0.70004pt}start_FLOATSUPERSCRIPT italic_ι end_FLOATSUPERSCRIPTJohns Hopkins University α𝛼{}^{\color[rgb]{0,0,1}\alpha\hskip 0.70004pt}start_FLOATSUPERSCRIPT italic_α end_FLOATSUPERSCRIPTAllen Institute for AI
λ𝜆{}^{\color[rgb]{0,0,1}\lambda\hskip 0.70004pt}start_FLOATSUPERSCRIPT italic_λ end_FLOATSUPERSCRIPTUniversity of Glasgow γ𝛾{}^{\color[rgb]{0,0,1}\gamma\hskip 0.70004pt}start_FLOATSUPERSCRIPT italic_γ end_FLOATSUPERSCRIPTYale University
oweller@cs.jhu.edu

Abstract

Modern Language Models (LMs) are capable of following long and complex instructions that enablea large and diverse set of user requests.While Information Retrieval (IR) models use these LMs as the backbone of their architectures,virtually none of them allow users to provide detailed instructions alongside queries, thus limiting their ability to satisfy complex information needs.In this work, we study the use of instructions in IR systems.First, we introduce our dataset FollowIR, which contains a rigorous instruction evaluation benchmark as well as a training set for helping IR models learn to better follow real-world instructions.FollowIRrepurposes detailed instructions—also known as narratives—developed for professional assessors to evaluate retrieval systems.In particular, we build our benchmark from three collections curated for shared tasks at the Text REtrieval Conference (TREC).These collections contains hundreds to thousands of labeled documents per query, making them suitable for our exploration.Through this process, we can measure how well IR models follow instructions, through a new pairwise evaluation framework.Our results indicate that existing retrieval models fail to correctly use instructions, using them for basic keywords and struggling to understand long-form information.However, we show that it is possible for IR models to learn to follow complex instructions: our new FollowIR-7B model has significant improvements after fine-tuning on our training set.

1 Introduction

Modern language models (LMs) are extensively tuned to be able to follow user instructions faithfully(Chung etal., 2022; Ouyang etal., 2022a; Rafailov etal., 2023; Wang etal., 2023b; Ivison etal., 2023) and safely(Bai etal., 2022; Bianchi etal., 2024).Through these capabilities, LMs are able to successfully tackle a broad range of tasks(Chiang etal., 2024; Liang etal., 2023; Yang etal., 2023; Jimenez etal., 2024; Zeng etal., 2023), even when not explicitly fine-tuned for them.

In contrast to the broader LM community, information retrieval (IR) practitioners and researchers have yet to fully exploit instruction-tuned models.Thanks to their ability to effectively estimate semantic similarity between query and documents, LMs have been adopted as the main backbone of neural retrieval architectures(Karpukhin etal., 2020; Khattab & Zaharia, 2020; Reimers & Gurevych, 2019).However, the vast majority of these systems are fine-tuned to operate exclusively as text spans similarity estimators(Khattab & Zaharia, 2020; Izacard etal., 2021; Nogueira & Cho, 2019; Pradeep etal., 2023; Ma etal., 2023).Moving past these ad-hoc search systems to retrieve with instructions would enable support for complex information needs.For example, imagine a researcher seeking to identify papers that must contain numerous qualities to be relevant (from a given venue, using a particular class of methods, etc.) while also making sure to avoid conditions that would make the document not-relevant (using negative sentiment, using datasets from a given domain, etc.).

Recent work has started to move towards search with instructions, but this topic is still understudied with only a handful of papers(Su etal., 2022; Asai etal., 2022; Muennighoff etal., 2024).In particular, we find their use of instructions to be narrow:instructions are typically short (fewer than 10 words) and repetitive (only one instruction per dataset e.g., Su etal. (2022); Asai etal. (2022); Li & Li (2023); Xiao etal. (2023)).Further, these works lack evaluation datasets that explicitly measure instruction following—instead focusing on standard ad-hoc retrieval benchmarks.

To address these gaps we introduce FollowIR, which consists of (1) a benchmark that explicitly measures the instruction following ability of retrieval models, and (2) training data that includes diverse and realistic instructions.Our key intuition is to leverage instructions developed for professional annotators of IR systems in order to study the capabilities of instruction-following IR models.These instructions are used by annotators to judge document relevance for a given query.Fortunately, the IR field is rich with such data, as these instructions—also known as narratives—are created for all queries in any well-constructed IR dataset.In particular, we use narratives developed for shared tasks at the Text REtrieval Conference111trec.nist.gov (TREC).These instructions are thorough and complex, including minute details about what makes a document relevant vs not-relevant.Thus if annotators can use these TREC instructions to annotate document relevance, so should instruction-following retrieval models (example query and instruction pairs are shown in Figures1 and 2).

We use three deeply-judged222i.e., that a large number of documents have been judged relevant or non-relevant, see Section2.1 for more. TREC collections as the basis of our evaluation set: TREC Robust 2004 (Voorhees, 2005), TREC Common Core 2017 (Allan etal., 2017), and TREC News 2021 (Soboroff etal., 2020).These collections have been thoroughly annotated in order to evaluate recall in retrieval, with hundreds to thousands of documents judged as relevant or not-relevant.We take the instructions given to the professional annotators and alter them slightly, manually re-annotating the relevant documents.We then have paired instructions, which can be used to test how models respond to changing instructions; that is, we measure if models update the set of relevant documents to match to the new altered instructions.

FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions (1)

As there are no existing methods to compare pairwise queries in IR, we develop a new evaluation framework to do so, measuring rank-wise score changes (which we call p𝑝pitalic_p-MRR) of documents given a pair of different instructions with the same query.Results on FollowIRindicate that current models generally fail to follow instructions in retrieval unless they are 3B+ parameters or have not been trained for retrieval. Our analysis shows that these failures are due to two phenomena: (1) models are not used to long instructions, and (2) models use instruction for keyword search rather than as a definition of relevance.

To further progress in building retrieval models that can understand instructions, we build a training set of real-world human-used instructions and fine-tune a model on them (FollowIR-7B). Our results show marked improvement on FollowIRfor both standard IR metrics and for p𝑝pitalic_p-MRR, indicating a starting point for future progress on instruction following.

In summary, we contribute the following: (1) a benchmark for evaluating instruction following in retrieval (FollowIR) consisting of human annotations on top of three already highly-judged corpora, (2) analysis of why current models fail to understand instructions, and (3) training data for teaching retrieval models to follow instructions along with a new open-sourced IR model, FollowIR-7B, that can handle long instructions in IR.333Links to the code, data, and models are available at https://github.com/orionw/FollowIR

FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions (2)

2 Related Work

2.1 TREC Conferences

The United States National Institute of Science and Technology (NIST) created the TREC organization in 1993. Each year TREC sponsors many tracks, or shared tasks, on a given dataset. These tracks range from a variety of topics: anywhere from standard ad-hoc retrieval on news (Soboroff etal., 2018; Soboroff, 2021) to more complex domains such as legal retrieval (Baron etal., 2006; Oard etal., 2008), or retrieval-augmented generation/report-generation (Lawrie etal., 2024).

As part of this process, NIST sponsors annotations for these collections. Typically, this is done by pooling a set of results (runs) from a variety of retrieval models and then annotating them in rank order until funding runs out. To help facilitate annotation, track organizers provide a narrative (or instruction) for each query that will be given to the annotators—however, IR models are only ever given the query. As evaluating total recall would require annotating every document in the collection for every query (which is not feasible for collections with millions of documents), recall error is tested using post-hoc sampling and annotation. Although not every query and document pair can be evaluated, recall for queries is very high. We build off the rigorous evaluation done at TREC by using several of their collections to build FollowIR.

2.2 Instructions for LMs

Instruction-following LMs have been popularized by models such as InstructGPT(Ouyang etal., 2022a), FLAN(Wei etal., 2022), and T0(Sanh etal., 2022).They have become a large area of interest for the natural language processing community (Touvron etal., 2023a; Jiang etal., 2023; Groeneveld etal., 2024; Black etal., 2022). There has been much work in evaluating if they can generalize to new instructions (Weller etal., 2020; Wang etal., 2022c; Ouyang etal., 2022b), if we can train them to follow instructions without human-annotated data (Wang etal., 2022b; Qin etal., 2023), and applying them to various domains (Zhao etal., 2021; Singhal etal., 2023; Shaghaghian etal., 2020). As the IR community uses LMs in their pipelines, we seek to broaden the scope of IR to include instructions, aligning it with the broader NLP community.

2.3 Instructions for Retrieval

Using instructions in retrieval is a nascent area of exploration. Su etal. (2022) and Asai etal. (2022) were two of the earliest works that trained a retrieval model to use instructions along with the query. However, these instructions are typically very short, such as “Retrieve a Wikipedia paragraph that answers this question." Recent work incorporates instructions in smaller models (Xiao etal., 2023; Chen etal., 2023, 2024) as well as others which use Llama (Touvron etal., 2023a; Weller etal., 2023) or Mistral (Jiang etal., 2023) as the backbone of a larger retrieval model that can use instructions: GritLM (Muennighoff etal., 2024) trains Mistral to do both generation and embedding, while Wang etal. (2023a) uses Mistral for embeddings only.

Despite this flurry of activity, these efforts do not have an explicit instruction-related retrieval benchmark to evaluate on. Instead, they evaluate on standard retrieval benchmark suites such as MTEB (Muennighoff etal., 2022) and BEIR (Thakur etal., 2021) which do not contain instructions. Thus, these newer instruction retrieval models hand-write a few instructions, where typically each instruction is applied to an entire dataset, irrespective of the query. This makes these instructions generic: focused only on the task format, format of the “document" (paragraph, sentence, etc.), and the broad domain. Note that because of this, no current instructions contain any extra background information or negation (Weller etal., 2024) which are commonly found in real-world instructions (see Figure1 for an example of these differences).

In work concurrent to ours, Oh etal. (2024) also propose a dataset to evaluate instructions in retrieval models. Their dataset uses the MS MARCO collection(Nguyen etal., 2016), and differs in several crucial aspects: it only has one relevant document per query (e.g., sparsely judged), is GPT-4 generated and validated, focuses on the background of the user (“I am a school teaching looking for …"), and evaluates using the lowest score over N𝑁Nitalic_N instructions for the same query (measuring robustness). In contrast, we use highly-judged corpora to ensure we can measure recall, use professionally generated instructions, have human-validated relevance judgements, propose a new paired evaluation protocol, and provide a training dataset and model for teaching instruction-following.

Dataset# Q𝑄Qitalic_Q|Q|𝑄|Q|| italic_Q ||I|𝐼|I|| italic_I |Rel. D/Q# Q𝑄Qitalic_Q|I|𝐼|I|| italic_I |Rel. D/Q
TREC News ’21 (Soboroff etal., 2020)5015.340.150.13246.919.2
TREC Core ’17 (Allan etal., 2017)5016.644.0180.02053.532.7
TREC Robust ’04 (Voorhees, 2005)24911.968.269.95275.219.8

3 Building FollowIR

We derive FollowIRfrom three TREC collections: TREC News 2021 (derived from the Washington Post v4 corpus; Soboroff etal., 2020), TREC Robust 2004 (from news articles in Disks 4 and 5 collections; Voorhees, 2005), and TREC Common Core 2017 (from the New York Times Annotated corpus; Allan etal., 2017).Each of these was professionally assessed to include hundreds of annotations per query (see Table1), with 50-180 relevant documents per query on average (and many more not-relevant annotations).

Each of these TREC tracks includes instructions for the professional annotators that we now also give to the models. Although using these alone can provide some indication of how well models can follow instructions, it doesn’t explicitly test their instruction following ability. To more carefully isolate this in our benchmark, we test whether models can respond to small changes in the instruction.

To accomplish this, we ask two expert annotators to modify the TREC instructions. However, doing this in a naive way would require re-annotating all the document judgements, a non-trivial task requiring immense annotation efforts.444NIST’s budget is $1–2 million USD/year: trec.nist.gov/pubs/2010.economic.impact.pdfInstead, we task the annotators with making instructions more specific by including additional constraints that narrow the relevance definition.These transformations cause some previously relevant documents to become non-relevant without introducing any new relevant documents from the pool.Therefore, only those documents that were deemed relevant by the original TREC assessors need to be re-annotated.This makes the annotation tractable, with only dozens or hundreds of documents to re-annotate per query instead of a collection of thousands.

We annotate a subset of the original TREC queries due to cost and overlap: we sample 50 queries from TREC Robust 2004 that do not overlap with TREC Common Core (as Common Core used 50 queries from Robust04 on a new collection), and 30 queries from TREC News 2021. Table1 shows dataset statistics of judged documents and the final benchmark size.Annotators were asked to change the instructions so that the number of relevant documents was cut roughly in half, thus including a sizeable number of changed relevance judgements.We note that although the number of queries seems small by NLP standards, 30-50 queries is both effective (Webber etal., 2008) and standard in the IR community due to the expense of careful annotation over many documents per query.

Due to differences in retriever quality, if we evaluate by searching over the full collection, each model will retrieve a different number of relevant documents. However, because we evaluate instruction following based on changing the document relevance, models that do poorly in the initial retrieval will have fewer documents which change relevance in the instruction-following evaluation. To rectify this, we instead turn to a reranking task where we include all relevant documents, and use a pool of five models555We use BM25, BGE-base, E5-base-v2, TART-Contriever, and INSTRUCTOR-xl. to select the top non-relevant documents. To be able to freely distribute the data due to fair-use laws, we chunk the documents into 400-word passages with 200-word overlap and select the highest scoring passages using MaxP (Dai & Callan, 2019). This enables us to distribute our data, which we do by extending the MTEB evaluation framework (Muennighoff etal., 2022).

3.1 Evaluation Metrics for FollowIR

Our benchmark provides two ways of measuring instruction following: (1) standard retrieval metrics when using the instructions with the queries and (2) pairwise evaluation of instruction following. For (1), we use typical IR evaluation metrics but use the instruction along with the query: these metrics are mean average precision (MAP) for Core17/Robust04 and normalized discounted cumulative gain at 5 (nDCG@5) for News21. For (2) we use our novel pairwise evaluation metric that measures the delta in scores when following the modified instructions instead of the original.666Note that we do not show standard retrieval results on the modified instruction’s relevant document set, as standard retrieval scores cannot be directly compared across different query relevance annotations (qrels).

Our new pairwise evaluation metric, p𝑝pitalic_p-MRR, measures rank-wise changes between queries. In developing this metric we had the following desiderata: it should compare the results of the original instruction to those of the new instruction, it should have a standardized range from worst possible change in instruction-following score (i.e., 11-1- 1) to best possible instruction-following score (i.e., 1111) with an option for no change when using different instructions (i.e., 00), and finally should take into account the document rank so that changes from rank 1 to rank 2 are more prominent than changes from rank 99 to 100. Given the above qualifications, we use the following equation applied to each changed relevance document per query (where MRR is mean reciprocal rank, Rogsubscript𝑅𝑜𝑔R_{og}italic_R start_POSTSUBSCRIPT italic_o italic_g end_POSTSUBSCRIPT is the rank of the document when using the original instruction and Rnewsubscript𝑅𝑛𝑒𝑤R_{new}italic_R start_POSTSUBSCRIPT italic_n italic_e italic_w end_POSTSUBSCRIPT is the new rank):

p-MRR={MRRogMRRnew1ifRog>Rnew1MRRnewMRRogotherwisep-MRRcases𝑀𝑅subscript𝑅𝑜𝑔𝑀𝑅subscript𝑅𝑛𝑒𝑤1ifRog>Rnew1𝑀𝑅subscript𝑅𝑛𝑒𝑤𝑀𝑅subscript𝑅𝑜𝑔otherwise\text{$p$-MRR}=\begin{cases}\frac{MRR_{og}}{MRR_{new}}-1&\text{if $R_{og}>R_{%new}$}\\[10.0pt]1-\frac{MRR_{new}}{MRR_{og}}&\text{otherwise}\end{cases}italic_p -MRR = { start_ROW start_CELL divide start_ARG italic_M italic_R italic_R start_POSTSUBSCRIPT italic_o italic_g end_POSTSUBSCRIPT end_ARG start_ARG italic_M italic_R italic_R start_POSTSUBSCRIPT italic_n italic_e italic_w end_POSTSUBSCRIPT end_ARG - 1 end_CELL start_CELL if italic_R start_POSTSUBSCRIPT italic_o italic_g end_POSTSUBSCRIPT > italic_R start_POSTSUBSCRIPT italic_n italic_e italic_w end_POSTSUBSCRIPT end_CELL end_ROW start_ROW start_CELL 1 - divide start_ARG italic_M italic_R italic_R start_POSTSUBSCRIPT italic_n italic_e italic_w end_POSTSUBSCRIPT end_ARG start_ARG italic_M italic_R italic_R start_POSTSUBSCRIPT italic_o italic_g end_POSTSUBSCRIPT end_ARG end_CELL start_CELL otherwise end_CELL end_ROW(1)

For the final score, we average first within a given query and then over all queries in the corpora—i.e., macro-averaging across queries, to handle the different number of relevant documents per query.

4 Evaluating Instruction Following

In this section we describe the models we evaluate, their results on FollowIR, and ablations performed to better understand the behavior of current models.

4.1 Evaluation Settings

We evaluate a wide variety of IR models (trained with and without instructions), including neural models ranging from 100 million to 7 billion parameters. We evaluate on the original TREC instructions in the FollowIRbenchmark and then on the new instructions, showing both standard IR metrics and the new pairwise metric p𝑝pitalic_p-MRR. We group models into four categories:

No Instructions in Training

These retrieval models did not see instructions in training and typically aren’t given them: this includes Contriever (Izacard etal., 2021), E5 (Wang etal., 2022a), MonoBERT Nogueira etal. (2019), MonoT5 (Nogueira etal., 2020), and BM25 (Robertson etal., 1995).

Robust04News21Core17Average
ModelMAPp𝑝pitalic_p-MRRnDCGp𝑝pitalic_p-MRRMAPp𝑝pitalic_p-MRRScorep𝑝pitalic_p-MRR
No-Instruction IRE5-base-v213.4 -6.720.9 -2.014.0 -2.916.1 -3.9
Contriever19.7 -6.122.9 -2.815.3 -2.519.3 -3.8
MonoBERT21.0 -9.425.1 -0.818.4 -0.221.5 -3.5
BM2512.1 -3.119.3 -2.18.1 -1.113.2 -2.1
MonoT5-base15.7 -6.211.0 +5.012.2 -4.113.0 -1.8
E5-large-v217.4 -4.224.3 +0.917.0 +0.119.6 -1.1
MonoT5-3B27.3 +4.016.5 +1.818.2 +1.820.7 +2.5
Instruction-IRTART-Contriever14.3 -9.021.8 -3.013.3 -3.016.5 -5.0
INSTRUCTOR-base17.2 -10.422.1 -1.815.5 -1.118.3 -4.4
E5-mistral23.1 -9.627.8 -0.918.3 +0.123.1 -3.5
BGE-base16.8 -6.520.0 -0.114.6 -2.717.1 -3.1
INSTRUCTOR-xl19.7 -8.126.1 -0.916.8 +0.720.9 -2.8
BGE-large17.5 -7.822.3 +0.615.0 +0.118.3 -2.4
GritLM-7B28.6 -1.724.4 -1.020.8 +2.624.6 -0.0
TART-FLAN-T5-xl24.6 -0.712.8 +2.017.0 +2.818.1 +1.4
APIsOpenAI v3 Large27.2 -5.827.2 -2.021.6 -0.225.3 -2.7
Cohere v3 English22.3 -3.628.3 +0.220.6 +2.823.7 -0.2
Google Gecko23.3 -2.429.5 +3.923.2 +5.425.3 +2.3
Instruct LMsFLAN-T5-base6.4 +5.36.1 -0.16.5 -3.36.3 +0.6
Llama-2-7B-chat6.3 +2.01.7 +0.25.4 +2.84.5 +1.7
FLAN-T5-large14.7 +3.98.0 +8.911.4 +1.311.4 +4.7
GritLM-Reranker9.7 +6.110.2 +3.49.8 +8.69.9 +6.0
Mistral-7B-instruct23.2 +12.627.2 +4.819.7 +13.023.4 +10.1
FollowIR-7B24.8 +13.729.6 +6.320.0 +16.524.8 +12.2

Instructions in IR Training

Most retrieval models using instructions received roughly one instruction per retrieval dataset, which generally defined the domain (e.g., “Financial"), document size (sentence, passage, etc.), and task format. This includes INSTRUCTOR models (Su etal., 2022), the bi-encoder TART model trained from Contriever (Asai etal., 2022), the reranker TART trained from FLAN-T5 (Chung etal., 2022), E5 Mistral-Instruct (Wang etal., 2023a), and GritLM (Muennighoff etal., 2024). We also include BGE models (Xiao etal., 2023) in this category, although they are trained with only one instruction total for each broad task (retrieval, clustering, etc.).

API Models

We use three of the best performing API embedding models: Cohere’s v3 English, Google’s Gecko (Lee etal., 2024) and OpenAI’s Text-Embedding-v3-Large. It is mostly unknown what these models’ training procedures were—including if they were trained on instructions or not—thus we place them in a distinct category. However, we note that Google’s model did explicitly train with instructions, as mentioned in their technical report.

Instruction-Tuned LMs

We also evaluate several instruction-tuned LMs to be used as rerankers, including FLAN-T5 (Chung etal., 2022), Llama v2 (Touvron etal., 2023b), and Mistral-Instruct-v0.2 (Jiang etal., 2023). We evaluate these models in the same fashion as MonoT5 rerankers, comparing the true and false tokens. Note that these models were not trained on any retrieval-specific data.

4.2 Results

Table2 shows the main results on FollowIR, with the standard IR score shown (either MAP or nDCG@5) as well as the pairwise evaluation metric, p𝑝pitalic_p-MRR.

No-Instruction IR Models

We see that the no-instruction models range widely in standard IR metrics (in terms of nDCG@5 and MAP) but generally have negative scores for p𝑝pitalic_p-MRR(up to 3.93.9-3.9- 3.9). The only non-instruction model to score positively on average is MonoT5-3B (+2.5 p𝑝pitalic_p-MRR).

Instruction IR Models

We again see that these models have generally negative scores, with the exception being GritLM (with scores averaging roughly zero) and TART-FLAN-T5-xl which has slightly positive scores for two of the three datasets (with an average of +1.4 p𝑝pitalic_p-MRR).

FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions (3)

API Models

We see that the API models perform strongly in terms of standard IR metrics, with OpenAI’s and Google’s models performing the highest overall. However, Cohere’s and OpenAI’s models perform poorly at instruction-following with negative scores (0.20.2-0.2- 0.2 and 2.72.7-2.7- 2.7 on average, respectively) whereas Google Gecko has positive scores (+2.3) likely a result of training on more instruction-focused data.

Instruct-Tuned LMs

In contrast to the previous results, all instruction-tuned LMs show positive results for instruction following, although they have the widest range of performance using standard IR metrics (ranging from very poor scores to some of the higher scores). We see that the best performing model in this category is FollowIR–7B, which we describe in more detail in Section5.

Overall

We see that the only models that show positive results at following instructions are either IR models with over 3B parameters or those that have been explicitly trained to follow instructions (e.g. FLAN-T5),without any retrieval-specific supervision.This aligns with work in the natural language processing community which has shown that the instruction-following ability improves with scale (Brown etal., 2020) and supervised instruction-tuning (Longpre etal., 2023).

4.3 Analysis

Why do so many models fail to correctly follow instructions when they do well on typical IR metrics such as nDCG and MAP? We answer this question by ablating several components that may impact results: (1) whether IR models are not used to text that cannot be used for simple keyword search (i.e. instructions) and (2) whether they are unused to the length of the longer instructions (as current instruction IR models have been trained on much shorter instructions).

SciFactNFCorpusFiQA
ModelOGΔΔ\Deltaroman_Δ w/Key.OGΔΔ\Deltaroman_Δ w/Key.OGΔΔ\Deltaroman_Δ w/Key.
No-InstructionBM2567.9-1.732.2-5.123.6-1.6
E5-base-v271.9-2.735.4-2.539.9-0.4
Contriever64.9+0.431.7+0.024.5-3.2
MonoT5-base73.1-0.635.6-0.941.2-0.3
Uses InstructionTART-Contriever67.6-0.333.4-5.331.8-0.4
INSTRUCTOR-base57.8+1.031.6-0.439.2-0.1
BGE-base73.2-0.535.5+0.040.8-2.3
TART-FLAN-xl74.2+1.633.9+0.439.6-0.3
INSTRUCTOR-xl62.4+0.236.0-0.646.9+0.8
E5-Mistral77.1-5.138.8+0.356.7-6.5

To test these, we compare the original query-only result to those where we additionally give the model either the full instruction, a shorter instruction, or keywords from the instruction. We gather these short instructions and keywords by prompting GPT-4-Turbo-1106 to generate them from the original full instruction (for TREC data) or otherwise use the original short instructions given by the authors of the model (for BEIR data). For the full prompt text, please see AppendixE.

We show results for these ablations in Table3, where positive scores indicate that adding information improves the model while negative scores indicate a drop in performance. We see a consistent trend where models that did poorly on longer instructions perform better on keywords and shorter instructions than with the full instruction. However, models that are able to follow instructions see better results with the additional information, on average.

These results show that models are (1) using the instruction text as keywords (as performance is higher when using only keywords) and (2) are not able to utilize the extra information in the instructions (as they generally decrease in performance with this additional information).

We also confirm that these results hold on datasets outside of TREC collections and show results on three BEIR datasets: SciFact, NFCorpus, and FiQA. We show in Table3 the original score (using the short instructions from their papers) and the change in score when using just keywords from the instruction (again extracted from GPT-4). We show results only for models which performed poorly for instruction-following.We see that the scores for keywords vs the short instruction are generally similar, with most models seeing a change of around ±plus-or-minus\pm± 1 point, except for the strongest of the non-instruction-following models, E5-Mistral, seeing a larger drop on some datasets.

Overall

We find overall (on both TREC and BEIR datasets) that models use instructions for keyword matching and are unused to longer instructions that may contain slightly less relevant words.

5 Teaching Instruction Following

Is it possible to improve model performance in following instructions? We show that fine-tuning on a training set of longer instructions can provide a method for doing so.

We start by gathering a training set to teach models. We collect all TREC narratives (i.e., instructions) from tasks not in FollowIR, consisting of 1836 pairs of queries and narratives. However, we note that this does not provide any positive or negative documents for fine-tuning.

In order to obtain documents for training, we prompt GPT-3.5-Turbo-1106 to generate relevant and not-relevant documents, generating roughly two relevant and non-relevant instances per query. However, these synthetic documents are noisy and contains errors w.r.t. the labels—to remedy this, we perform a round of filtering and use the best performing open-source model from Table2 (Mistral-7B-Instruct-v0.2) to score each of the generated documents according to the instruction. We then filter the documents according to whether Mistral correctly predicts the generated label, and finally balance the relevant and non-relevant samples, choosing only one relevant and non-relevant document per query. Our total is similar-to\sim1800 training instances on similar-to\sim1200 unique query/instructions pairs.

We then train our instruction-following model, FollowIR-7B, by fine-tuning Mistral-7B-Instruct-v0.2 on our data using the Llama-Factory framework (Hiyouga, 2023) with LoRA (Hu etal., 2021). Full training hyperparameter details are found in AppendixA.

ModelRobustness@10
BM2526.9
TART-Contriever47.5
RepLLaMa52.6
E5-Mistral55.4
Mistral-7B-instruct35.3
FollowIR-7B71.5

When we evaluate this model on FollowIR(Table2), we find that the scores consistently improve. Compared to the original Mistral-7B-Instruct-v0.2, our model improves on both standard IR metrics (+6.0% relative improvement) and on instruction following (+20.8% relative improvement). We also show that this improvement holds on the concurrent InstructIR dataset (Table4), where FollowIR-7B scores double the base Mistral-7B scores (71.5 Robustness@10 vs 35.3) and is the top scoring model overall. Thus, we can see that it is possible to train IR models to be better instruction followers.

6 Conclusion

Despite the use of LMs as the backbone of neural retrieval models, most existing IR models do not take instructions that define document relevance.Further, there is no existing resource that measures how well retrieval models can follow instructions.We build a new benchmark that explicitly measures the instruction following ability of retrieval models and find that nearly all retrieval models do not follow instructions, with the exception of larger models (3B+ parameters) or instruction-tuned LMs that typically are not used for retrieval.However, we show that it is possible to improve their instruction following ability, and build and release a training corpus for teaching retrieval models to follow instructions.Our new model, FollowIR-7B, shows improvement on both standard retrieval metrics as well as in instruction following.We hope that these resources will allow the community to develop more capable instruction-following retrieval models that can quickly adapt to a relevance definition given flexible natural language text.

7 Limitations

Reranking vs Full Retrieval

As our setup for evaluating instruction following requires evaluating the documents which changed relevance, we cannot use the full collection for retrieval (as each retriever finds different relevant documents by design). Further, due to licensing restrictions, we cannot distribute the full corpora from the TREC tracks—thus we distribute passages due to fair use laws. However, we show full corpus retrieval results for a subset of models in AppendixC and note similar trends in terms of the lack of instruction following.

Possible Errors

Our work is built on the TREC document collections and judgements, as well as new annotation efforts. We do not check for potential errors in the TREC annotations, and our newly gathered annotations may have small errors. Despite these caveats, we see that our dataset still provides a useful evaluation setup for measuring instruction following.

References

  • Allan etal. (2017)James Allan, Donna Harman, Evangelos Kanoulas, Dan Li, Christophe VanGysel, and EllenM Voorhees.Trec 2017 common core track overview.In TREC, 2017.
  • Asai etal. (2022)Akari Asai, Timo Schick, Patrick Lewis, Xilun Chen, Gautier Izacard, Sebastian Riedel, Hannaneh Hajishirzi, and Wen-tau Yih.Task-aware retrieval with instructions.arXiv preprint arXiv:2211.09260, 2022.
  • Bai etal. (2022)Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, Nicholas Joseph, Saurav Kadavath, Jackson Kernion, Tom Conerly, Sheer El-Showk, Nelson Elhage, Zac Hatfield-Dodds, Danny Hernandez, Tristan Hume, Scott Johnston, Shauna Kravec, Liane Lovitt, Neel Nanda, Catherine Olsson, Dario Amodei, Tom Brown, Jack Clark, Sam McCandlish, Chris Olah, Ben Mann, and Jared Kaplan.Training a helpful and harmless assistant with reinforcement learning from human feedback, 2022.
  • Baron etal. (2006)JasonR Baron, DavidD Lewis, and DouglasW Oard.Trec 2006 legal track overview.In TREC. Citeseer, 2006.
  • Bianchi etal. (2024)Federico Bianchi, Mirac Suzgun, Giuseppe Attanasio, Paul Röttger, Dan Jurafsky, Tatsunori Hashimoto, and James Zou.Safety-tuned llamas: Lessons from improving the safety of large language models that follow instructions, 2024.
  • Black etal. (2022)Sid Black, Stella Biderman, Eric Hallahan, QuentinG. Anthony, Leo Gao, Laurence Golding, Horace He, Connor Leahy, Kyle McDonell, Jason Phang, MichaelMartin Pieler, USVSNSai Prashanth, Shivanshu Purohit, Laria Reynolds, Jonathan Tow, Benqi Wang, and Samuel Weinbach.Gpt-neox-20b: An open-source autoregressive language model.ArXiv, abs/2204.06745, 2022.URL https://api.semanticscholar.org/CorpusID:248177957.
  • Brown etal. (2020)Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, JaredD Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, etal.Language models are few-shot learners.Advances in neural information processing systems, 33:1877–1901, 2020.
  • Chen etal. (2024)Haonan Chen, Zhicheng Dou, Kelong Mao, Jiongnan Liu, and Ziliang Zhao.Generalizing conversational dense retrieval via llm-cognition data augmentation.arXiv preprint arXiv:2402.07092, 2024.
  • Chen etal. (2023)Jianlv Chen, sh*tao Xiao, Peitian Zhang, Kun Luo, Defu Lian, and Zheng Liu.Bge m3-embedding: Multi-lingual, multi-functionality, multi-granularity text embeddings through self-knowledge distillation, 2023.
  • Chiang etal. (2024)Wei-Lin Chiang, Lianmin Zheng, Ying Sheng, AnastasiosNikolas Angelopoulos, Tianle Li, Dacheng Li, Hao Zhang, Banghua Zhu, Michael Jordan, JosephE. Gonzalez, and Ion Stoica.Chatbot arena: An open platform for evaluating llms by human preference, 2024.
  • Chung etal. (2022)HyungWon Chung, LeHou, Shayne Longpre, Barret Zoph, YiTay, William Fedus, Yunxuan Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson, ShixiangShane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, Aakanksha Chowdhery, Alex Castro-Ros, Marie Pellat, Kevin Robinson, Dasha Valter, Sharan Narang, Gaurav Mishra, Adams Yu, Vincent Zhao, Yanping Huang, Andrew Dai, Hongkun Yu, Slav Petrov, EdH. Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, QuocV. Le, and Jason Wei.Scaling instruction-finetuned language models, 2022.
  • Dai & Callan (2019)Zhuyun Dai and Jamie Callan.Deeper text understanding for ir with contextual neural language modeling.In Proceedings of the 42nd international ACM SIGIR conference on research and development in information retrieval, pp. 985–988, 2019.
  • Groeneveld etal. (2024)Dirk Groeneveld, IzBeltagy, Pete Walsh, Aksh*ta Bhagia, Rodney Kinney, Oyvind Tafjord, A.Jha, Hamish Ivison, Ian Magnusson, Yizhong Wang, Shane Arora, David Atkinson, Russell Authur, KhyathiRaghavi Chandu, Arman Cohan, Jennifer Dumas, Yanai Elazar, Yuling Gu, Jack Hessel, Tushar Khot, William Merrill, JacobDaniel Morrison, Niklas Muennighoff, Aakanksha Naik, Crystal Nam, MatthewE. Peters, Valentina Pyatkin, Abhilasha Ravichander, Dustin Schwenk, Saurabh Shah, Will Smith, Emma Strubell, Nishant Subramani, Mitchell Wortsman, Pradeep Dasigi, Nathan Lambert, Kyle Richardson, Luke Zettlemoyer, Jesse Dodge, Kyle Lo, Luca Soldaini, NoahA. Smith, and Hanna Hajishirzi.Olmo: Accelerating the science of language models.2024.URL https://api.semanticscholar.org/CorpusID:267365485.
  • Hiyouga (2023)Hiyouga.Llama factory.https://github.com/hiyouga/LLaMA-Factory, 2023.
  • Hu etal. (2021)EdwardJ Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, LuWang, and Weizhu Chen.Lora: Low-rank adaptation of large language models.arXiv preprint arXiv:2106.09685, 2021.
  • Ivison etal. (2023)Hamish Ivison, Yizhong Wang, Valentina Pyatkin, Nathan Lambert, Matthew Peters, Pradeep Dasigi, Joel Jang, David Wadden, NoahA. Smith, IzBeltagy, and Hannaneh Hajishirzi.Camels in a changing climate: Enhancing lm adaptation with tulu 2, 2023.
  • Izacard etal. (2021)Gautier Izacard, Mathilde Caron, Lucas Hosseini, Sebastian Riedel, Piotr Bojanowski, Armand Joulin, and Edouard Grave.Unsupervised dense information retrieval with contrastive learning.arXiv preprint arXiv:2112.09118, 2021.
  • Jiang etal. (2023)AlbertQiaochu Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, DevendraSingh Chaplot, Diego deLasCasas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, L’elioRenard Lavaud, Marie-Anne Lachaux, Pierre Stock, TevenLe Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, and WilliamEl Sayed.Mistral 7b.ArXiv, abs/2310.06825, 2023.URL https://api.semanticscholar.org/CorpusID:263830494.
  • Jimenez etal. (2024)CarlosE Jimenez, John Yang, Alexander Wettig, Shunyu Yao, Kexin Pei, Ofir Press, and KarthikR Narasimhan.SWE-bench: Can language models resolve real-world github issues?In The Twelfth International Conference on Learning Representations, 2024.URL https://openreview.net/forum?id=VTF8yNQM66.
  • Karpukhin etal. (2020)Vladimir Karpukhin, Barlas Oğuz, Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih.Dense passage retrieval for open-domain question answering.arXiv preprint arXiv:2004.04906, 2020.
  • Khattab & Zaharia (2020)Omar Khattab and Matei Zaharia.Colbert: Efficient and effective passage search via contextualized late interaction over bert.In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, pp. 39–48, 2020.
  • Lawrie etal. (2024)Dawn Lawrie, Sean MacAvaney, James Mayfield, Paul McNamee, DouglasW. Oard, Luca Soldaini, and Eugene Yang.Overview of the TREC 2023 NeuCLIR track.2024.
  • Lee etal. (2024)Jinhyuk Lee, Zhuyun Dai, Xiaoqi Ren, Blair Chen, Daniel Cer, JeremyR Cole, Kai Hui, Michael Boratko, Rajvi Kapadia, Wen Ding, etal.Gecko: Versatile text embeddings distilled from large language models.arXiv preprint arXiv:2403.20327, 2024.
  • Li & Li (2023)Xianming Li and Jing Li.Angle-optimized text embeddings.arXiv preprint arXiv:2309.12871, 2023.
  • Liang etal. (2023)Percy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, Benjamin Newman, Binhang Yuan, Bobby Yan, CeZhang, Christian Cosgrove, ChristopherD. Manning, Christopher Ré, Diana Acosta-Navas, DrewA. Hudson, Eric Zelikman, Esin Durmus, Faisal Ladhak, Frieda Rong, Hongyu Ren, Huaxiu Yao, Jue Wang, Keshav Santhanam, Laurel Orr, Lucia Zheng, Mert Yuksekgonul, Mirac Suzgun, Nathan Kim, Neel Guha, Niladri Chatterji, Omar Khattab, Peter Henderson, Qian Huang, Ryan Chi, SangMichael Xie, Shibani Santurkar, Surya Ganguli, Tatsunori Hashimoto, Thomas Icard, Tianyi Zhang, Vishrav Chaudhary, William Wang, Xuechen Li, Yifan Mai, Yuhui Zhang, and Yuta Koreeda.Holistic evaluation of language models, 2023.
  • Longpre etal. (2023)Shayne Longpre, LeHou, TuVu, Albert Webson, HyungWon Chung, YiTay, Denny Zhou, QuocV Le, Barret Zoph, Jason Wei, etal.The flan collection: Designing data and methods for effective instruction tuning.In International Conference on Machine Learning, pp. 22631–22648. PMLR, 2023.
  • Ma etal. (2023)Xueguang Ma, Liang Wang, Nan Yang, Furu Wei, and Jimmy Lin.Fine-tuning llama for multi-stage text retrieval.arXiv preprint arXiv:2310.08319, 2023.
  • Muennighoff etal. (2022)Niklas Muennighoff, Nouamane Tazi, Loïc Magne, and Nils Reimers.Mteb: Massive text embedding benchmark.arXiv preprint arXiv:2210.07316, 2022.
  • Muennighoff etal. (2024)Niklas Muennighoff, Hongjin Su, Liang Wang, Nan Yang, Furu Wei, Tao Yu, Amanpreet Singh, and Douwe Kiela.Generative representational instruction tuning.arXiv preprint arXiv:2402.09906, 2024.
  • Nguyen etal. (2016)Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and LiDeng.MS MARCO: A human generated machine reading comprehension dataset.CoRR, abs/1611.09268, 2016.URL http://arxiv.org/abs/1611.09268.
  • Nogueira & Cho (2019)Rodrigo Nogueira and Kyunghyun Cho.Passage re-ranking with bert.arXiv preprint arXiv:1901.04085, 2019.
  • Nogueira etal. (2019)Rodrigo Nogueira, Wei Yang, Kyunghyun Cho, and Jimmy Lin.Multi-stage document ranking with bert.arXiv preprint arXiv:1910.14424, 2019.
  • Nogueira etal. (2020)Rodrigo Nogueira, Zhiying Jiang, and Jimmy Lin.Document ranking with a pretrained sequence-to-sequence model.arXiv preprint arXiv:2003.06713, 2020.
  • Oard etal. (2008)DouglasW Oard, Björn Hedin, Stephen Tomlinson, and JasonR Baron.Overview of the trec 2008 legal track.In TREC, pp. 500–277, 2008.
  • Oh etal. (2024)Hanseok Oh, Hyunji Lee, Seonghyeon Ye, Haebin Shin, Hansol Jang, Changwook Jun, and Minjoon Seo.Instructir: A benchmark for instruction following of information retrieval models.arXiv preprint arXiv:2402.14334, 2024.
  • Ouyang etal. (2022a)Long Ouyang, Jeffrey Wu, XuJiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, PaulF Christiano, Jan Leike, and Ryan Lowe.Training language models to follow instructions with human feedback.In S.Koyejo, S.Mohamed, A.Agarwal, D.Belgrave, K.Cho, and A.Oh (eds.), Advances in Neural Information Processing Systems, volume35, pp. 27730–27744. Curran Associates, Inc., 2022a.URL https://proceedings.neurips.cc/paper_files/paper/2022/file/b1efde53be364a73914f58805a001731-Paper-Conference.pdf.
  • Ouyang etal. (2022b)Long Ouyang, Jeffrey Wu, XuJiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, etal.Training language models to follow instructions with human feedback.Advances in Neural Information Processing Systems, 35:27730–27744, 2022b.
  • Pradeep etal. (2023)Ronak Pradeep, Sahel Sharifymoghaddam, and Jimmy Lin.Rankzephyr: Effective and robust zero-shot listwise reranking is a breeze!arXiv preprint arXiv:2312.02724, 2023.
  • Qin etal. (2023)Yujia Qin, Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu, Yankai Lin, Xin Cong, Xiangru Tang, Bill Qian, Sihan Zhao, Lauren Hong, Runchu Tian, Ruobing Xie, Jie Zhou, Mark Gerstein, Dahai Li, Zhiyuan Liu, and Maosong Sun.Toolllm: Facilitating large language models to master 16000+ real-world apis, 2023.
  • Rafailov etal. (2023)Rafael Rafailov, Archit Sharma, Eric Mitchell, ChristopherD Manning, Stefano Ermon, and Chelsea Finn.Direct preference optimization: Your language model is secretly a reward model.In A.Oh, T.Neumann, A.Globerson, K.Saenko, M.Hardt, and S.Levine (eds.), Advances in Neural Information Processing Systems, volume36, pp. 53728–53741. Curran Associates, Inc., 2023.URL https://proceedings.neurips.cc/paper_files/paper/2023/file/a85b405ed65c6477a4fe8302b5e06ce7-Paper-Conference.pdf.
  • Reimers & Gurevych (2019)Nils Reimers and Iryna Gurevych.Sentence-bert: Sentence embeddings using siamese bert-networks.arXiv preprint arXiv:1908.10084, 2019.
  • Robertson etal. (1995)StephenE Robertson, Steve Walker, Susan Jones, MichelineM Hanco*ck-Beaulieu, Mike Gatford, etal.Okapi at trec-3.Nist Special Publication Sp, 109:109, 1995.
  • Sanh etal. (2022)Victor Sanh, Albert Webson, Colin Raffel, Stephen Bach, Lintang Sutawika, Zaid Alyafeai, Antoine Chaffin, Arnaud Stiegler, Arun Raja, Manan Dey, MSaiful Bari, Canwen Xu, Urmish Thakker, ShanyaSharma Sharma, Eliza Szczechla, Taewoon Kim, Gunjan Chhablani, Nihal Nayak, Debajyoti Datta, Jonathan Chang, Mike Tian-Jian Jiang, Han Wang, Matteo Manica, Sheng Shen, ZhengXin Yong, Harsh*t Pandey, Rachel Bawden, Thomas Wang, Trishala Neeraj, Jos Rozen, Abheesht Sharma, Andrea Santilli, Thibault Fevry, JasonAlan Fries, Ryan Teehan, TevenLe Scao, Stella Biderman, Leo Gao, Thomas Wolf, and AlexanderM Rush.Multitask prompted training enables zero-shot task generalization.In International Conference on Learning Representations, 2022.URL https://openreview.net/forum?id=9Vrb9D0WI4.
  • Shaghaghian etal. (2020)Shohreh Shaghaghian, LunaYue Feng, Borna Jafarpour, and Nicolai Pogrebnyakov.Customizing contextualized language models for legal document reviews.In 2020 IEEE International Conference on Big Data (Big Data), pp. 2139–2148. IEEE, 2020.
  • Singhal etal. (2023)Karan Singhal, Shekoofeh Azizi, Tao Tu, SSara Mahdavi, Jason Wei, HyungWon Chung, Nathan Scales, Ajay Tanwani, Heather Cole-Lewis, Stephen Pfohl, etal.Large language models encode clinical knowledge.Nature, 620(7972):172–180, 2023.
  • Soboroff (2021)Ian Soboroff.Overview of trec 2021.In 30th Text REtrieval Conference. Gaithersburg, Maryland, 2021.
  • Soboroff etal. (2018)Ian Soboroff, Shudong Huang, and Donna Harman.Trec 2018 news track overview.In TREC, volume 409, pp. 410, 2018.
  • Soboroff etal. (2020)Ian Soboroff, Shudong Huang, and Donna Harman.Trec 2020 news track overview.In TREC, 2020.
  • Su etal. (2022)Hongjin Su, Weijia Shi, Jungo Kasai, Yizhong Wang, Yushi Hu, Mari Ostendorf, Wen-tau Yih, NoahA. Smith, Luke Zettlemoyer, and Tao Yu.One embedder, any task: Instruction-finetuned text embeddings.2022.URL https://arxiv.org/abs/2212.09741.
  • Thakur etal. (2021)Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, and Iryna Gurevych.Beir: A heterogenous benchmark for zero-shot evaluation of information retrieval models.arXiv preprint arXiv:2104.08663, 2021.
  • Touvron etal. (2023a)Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, etal.Llama: Open and efficient foundation language models.arXiv preprint arXiv:2302.13971, 2023a.
  • Touvron etal. (2023b)Hugo Touvron, Louis Martin, KevinR. Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, DanielM. Bikel, Lukas Blecher, CristianCantón Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, AnthonyS. Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, IsabelM. Kloumann, A.V. Korenev, PunitSingh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, EricMichael Smith, R.Subramanian, Xia Tan, Binh Tang, Ross Taylor, Adina Williams, JianXiang Kuan, Puxin Xu, Zhengxu Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, andThomas Scialom.Llama 2: Open foundation and fine-tuned chat models.ArXiv, abs/2307.09288, 2023b.URL https://api.semanticscholar.org/CorpusID:259950998.
  • Voorhees (2005)EllenM Voorhees.The trec robust retrieval track.In ACM SIGIR Forum, volume39, pp. 11–20. ACM New York, NY, USA, 2005.
  • Wang etal. (2022a)Liang Wang, Nan Yang, Xiaolong Huang, Binxing Jiao, Linjun Yang, Daxin Jiang, Rangan Majumder, and Furu Wei.Text embeddings by weakly-supervised contrastive pre-training.arXiv preprint arXiv:2212.03533, 2022a.
  • Wang etal. (2023a)Liang Wang, Nan Yang, Xiaolong Huang, Linjun Yang, Rangan Majumder, and Furu Wei.Improving text embeddings with large language models.arXiv preprint arXiv:2401.00368, 2023a.
  • Wang etal. (2022b)Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, NoahA Smith, Daniel Khashabi, and Hannaneh Hajishirzi.Self-instruct: Aligning language model with self generated instructions.arXiv preprint arXiv:2212.10560, 2022b.
  • Wang etal. (2022c)Yizhong Wang, Swaroop Mishra, Pegah Alipoormolabashi, Yeganeh Kordi, Amirreza Mirzaei, Anjana Arunkumar, Arjun Ashok, ArutSelvan Dhanasekaran, Atharva Naik, David Stap, etal.Super-naturalinstructions: Generalization via declarative instructions on 1600+ nlp tasks.arXiv preprint arXiv:2204.07705, 2022c.
  • Wang etal. (2023b)Yizhong Wang, Hamish Ivison, Pradeep Dasigi, Jack Hessel, Tushar Khot, KhyathiRaghavi Chandu, David Wadden, Kelsey MacMillan, NoahA. Smith, IzBeltagy, and Hannaneh Hajishirzi.How far can camels go? exploring the state of instruction tuning on open resources, 2023b.
  • Webber etal. (2008)William Webber, Alistair Moffat, and Justin Zobel.Statistical power in retrieval experimentation.In Proceedings of the 17th ACM conference on Information and knowledge management, pp. 571–580, 2008.
  • Wei etal. (2022)Jason Wei, Maarten Bosma, VincentY. Zhao, Kelvin Guu, AdamsWei Yu, Brian Lester, Nan Du, AndrewM. Dai, and QuocV. Le.Finetuned language models are zero-shot learners, 2022.
  • Weller etal. (2020)Orion Weller, Nicholas Lourie, Matt Gardner, and MatthewE Peters.Learning from task descriptions.arXiv preprint arXiv:2011.08115, 2020.
  • Weller etal. (2023)Orion Weller, Kyle Lo, David Wadden, Dawn Lawrie, Benjamin VanDurme, Arman Cohan, and Luca Soldaini.When do generative query and document expansions fail? a comprehensive study across methods, retrievers, and datasets.arXiv preprint arXiv:2309.08541, 2023.
  • Weller etal. (2024)Orion Weller, DawnJ Lawrie, and BenjaminVan Durme.Nevir: Negation in neural information retrieval.Conference of the European Chapter of the Association for Computational Linguistics, 2024.URL https://api.semanticscholar.org/CorpusID:258676146.
  • Xiao etal. (2023)sh*tao Xiao, Zheng Liu, Peitian Zhang, and Niklas Muennighoff.C-pack: Packaged resources to advance general chinese embedding, 2023.
  • Yang etal. (2023)Rui Yang, Lin Song, Yanwei Li, Sijie Zhao, Yixiao Ge, Xiu Li, and Ying Shan.Gpt4tools: Teaching large language model to use tools via self-instruction.In A.Oh, T.Neumann, A.Globerson, K.Saenko, M.Hardt, and S.Levine (eds.), Advances in Neural Information Processing Systems, volume36, pp. 71995–72007. Curran Associates, Inc., 2023.URL https://proceedings.neurips.cc/paper_files/paper/2023/file/e393677793767624f2821cec8bdd02f1-Paper-Conference.pdf.
  • Zeng etal. (2023)Zhiyuan Zeng, Jiatong Yu, Tianyu Gao, YuMeng, Tanya Goyal, and Danqi Chen.Evaluating large language models at evaluating instruction following, 2023.
  • Zhao etal. (2021)Ming Zhao, Peter Anderson, Vihan Jain, SuWang, Alexander Ku, Jason Baldridge, and Eugene Ie.On the evaluation of vision-and-language navigation instructions.arXiv preprint arXiv:2101.10504, 2021.

Appendix A Hyperparameters for Fine-Tuning Mistral

Appendix B Hyperparameters for Inference

We use default parameters for inference, taken from the original code of the authors of the papers we use (from their MTEB evaluations).

Appendix C Full Retrieval Results

In Table4 we show results for models searching on the full collections of the TREC tasks included in FollowIR. Note that because each model retrieves different relevant documents, the instruction-following evaluation has a different set of instances that each model is evaluated on (as it can only be evaluated on documents it retrieved that then become not-relevant).

Robust04News21Core17
(mAP)(nDCG@5)(mAP)
ModelOGΔΔ\Deltaroman_ΔOGΔΔ\Deltaroman_ΔOGΔΔ\Deltaroman_Δ
No
Instruct
BM2521.4-1.230.1+5.316.8-0.2
E5-base-v222.7-7.033.6+1.819.7-3.0
Contriever19.2-7.722.5+9.022.6-7.6
Uses
Instruct
TART-Contriever25.5-10.140.0-5.022.6-7.6
BGE-base23.6-3.136.5-7.823.0-2.1
INSTRUCTOR-base22.5-2.233.3-2.820.0-0.2
INSTRUCTOR-XL30.4-3.138.1-0.129.9-2.8

Appendix D Keywords used for BEIR experiments

GPT-4-Turbo-1106 extracted the following keywords (Table5) from the instructions these models used, which generated the results in Table3.

ModelDatasetKeywords
BM25/Contriever/E5/MonoT5FiQAFinance web
BM25/Contriever/E5/MonoT5SciFactscience paper verify
BM25/Contriever/E5/MonoT5NFCorpusmedicine relevant
TART-dualFiQAfinancial web
TART-dualSciFactscientific paper verify
TART-dualNFCorpusscientific paper paragraph
INSTRUCTOR-baseFiQAfinancial supporting:
INSTRUCTOR-baseSciFactscientific supporting passage:
INSTRUCTOR-baseNFCorpusmedicine relevant
BGE-baseFiQArelevant passages:
BGE-baseSciFactrelevant passages:
BGE-baseNFCorpusrelevant passages:
INSTRUCTOR-xlFiQAfinance supporting:
INSTRUCTOR-xlSciFactscientific supporting passages:
INSTRUCTOR-xlNFCorpusnutrition facts public medical:
E5-MistralFiQAfinancial replies
E5-MistralSciFactscientific
E5-MistralNFCorpusretrieve relevant
TART-T5-FLAN-xlFiQAfinancial web
TART-T5-FLAN-xlSciFactscientific paper verify
TART-T5-FLAN-xlNFCorpusScientific paper paragraph

Appendix E Prompts Used

We use these prompts for generating the short instructions, the keywords, and the synthetic documents. The examples used in the prompt for the “Full Instructions to Short Instructions" prompt were partially created by the authors, as only the short instructions were provided by TART/INSTRUCTOR.

FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions (2024)

References

Top Articles
Latest Posts
Article information

Author: Stevie Stamm

Last Updated:

Views: 5618

Rating: 5 / 5 (60 voted)

Reviews: 91% of readers found this page helpful

Author information

Name: Stevie Stamm

Birthday: 1996-06-22

Address: Apt. 419 4200 Sipes Estate, East Delmerview, WY 05617

Phone: +342332224300

Job: Future Advertising Analyst

Hobby: Leather crafting, Puzzles, Leather crafting, scrapbook, Urban exploration, Cabaret, Skateboarding

Introduction: My name is Stevie Stamm, I am a colorful, sparkling, splendid, vast, open, hilarious, tender person who loves writing and wants to share my knowledge and understanding with you.