Abstract
Entity-oriented search systems often learn vector representations of entities via the introductory paragraph from the Wikipedia page of the entity. As such representations are the same for every query, our hypothesis is that the representations are not ideal for IR tasks. In this work, we present BERT Entity Representations (BERT-ER) which are query-specific vector representations of entities obtained from text that describes how an entity is relevant for a query. Using BERT-ER in a downstream entity ranking system, we achieve a performance improvement of 13-42% (Mean Average Precision) over a system that uses the BERT embedding of the introductory paragraph from Wikipedia on two large-scale test collections. Our approach also outperforms entity ranking systems using entity embeddings from Wikipedia2Vec, ERNIE, and E-BERT. We show that our entity ranking system using BERT-ER can increase precision at the top of the ranking by promoting relevant entities to the top. With this work, we release our BERT models and query-specific entity embeddings fine-tuned for the entity ranking task.
Recommended Citation
S. Chatterjee and L. Dietz, "Bert-Er: Query-Specific Bert Entity Representations for Entity Ranking," SIGIR 2022 - Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1466 - 1477, Association for Computing Machinery, Jul 2022.
The definitive version is available at https://doi.org/10.1145/3477495.3531944
Department(s)
Computer Science
Publication Status
Public Access
Keywords and Phrases
bert; entity ranking; query-specific entity representations
International Standard Book Number (ISBN)
978-145038732-3
Document Type
Article - Conference proceedings
Document Version
Citation
File Type
text
Language(s)
English
Rights
© 2024 Association for Computing Machinery, All rights reserved.
Publication Date
06 Jul 2022
Comments
National Science Foundation, Grant 1846017