Abstract

Current query expansion models use pseudo-relevance feedback to improve first-pass retrieval effectiveness; however, this fails when the initial results are not relevant. Instead of building a language model from retrieved results, we propose Generative Relevance Feedback (GRF) that builds probabilistic feedback models from long-form text generated from Large Language Models. We study the effective methods for generating text by varying the zero-shot generation subtasks: queries, entities, facts, news articles, documents, and essays. We evaluate GRF on document retrieval benchmarks covering a diverse set of queries and document collections, and the results show that GRF methods significantly outperform previous PRF methods. Specifically, we improve MAP between 5-19% and nDCG@10 17-24% compared to RM3 expansion and achieve state-of-the-art recall across all datasets.

Department(s)

Computer Science

Comments

Engineering and Physical Sciences Research Council, Grant EP/V025708/1

Keywords and Phrases

Document Retrieval; Pseudo-Relevance Feedback; Text Generation

International Standard Book Number (ISBN)

978-145039408-6

Document Type

Article - Conference proceedings

Document Version

Citation

File Type

text

Language(s)

English

Rights

© 2024 Association for Computing Machinery, All rights reserved.

Publication Date

19 Jul 2023

Share

 
COinS