You can't pick your neighbors, or can you? When and how to rely on retrieval in the NN-LM

A Drozdov, S Wang, R Rahimi, A McCallum… - arXiv preprint arXiv …, 2022 - arxiv.org
arXiv preprint arXiv:2210.15859, 2022arxiv.org
Retrieval-enhanced language models (LMs), which condition their predictions on text
retrieved from large external datastores, have recently shown significant perplexity
improvements compared to standard LMs. One such approach, the $ k $ NN-LM,
interpolates any existing LM's predictions with the output of a $ k $-nearest neighbors model
and requires no additional training. In this paper, we explore the importance of lexical and
semantic matching in the context of items retrieved by $ k $ NN-LM. We find two trends:(1) …
Retrieval-enhanced language models (LMs), which condition their predictions on text retrieved from large external datastores, have recently shown significant perplexity improvements compared to standard LMs. One such approach, the NN-LM, interpolates any existing LM's predictions with the output of a -nearest neighbors model and requires no additional training. In this paper, we explore the importance of lexical and semantic matching in the context of items retrieved by NN-LM. We find two trends: (1) the presence of large overlapping -grams between the datastore and evaluation set plays an important factor in strong performance, even when the datastore is derived from the training data; and (2) the NN-LM is most beneficial when retrieved items have high semantic similarity with the query. Based on our analysis, we define a new formulation of the NN-LM that uses retrieval quality to assign the interpolation coefficient. We empirically measure the effectiveness of our approach on two English language modeling datasets, Wikitext-103 and PG-19. Our re-formulation of the NN-LM is beneficial in both cases, and leads to nearly 4% improvement in perplexity on the Wikitext-103 test set.
arxiv.org