Title
CEDR: Contextualized Embeddings for Document Ranking
Abstract
Although considerable attention has been given to neural ranking architectures recently, far less attention has been paid to the term representations that are used as input to these models. In this work, we investigate how two pretrained contextualized language modes (ELMo and BERT) can be utilized for ad-hoc document ranking. Through experiments on TREC benchmarks, we find that several existing neural ranking architectures can benefit from the additional context provided by contextualized language models. Furthermore, we propose a joint approach that incorporates BERTu0027s classification vector into existing neural models and show that it outperforms state-of-the-art ad-hoc ranking baselines. We call this joint approach CEDR (Contextualized Embeddings for Document Ranking). We also address practical challenges in using these models for ranking, including the maximum input length imposed by BERT and runtime performance impacts of contextualized language models.
Year
DOI
Venue
2019
10.1145/3331184.3331317
Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval
Keywords
Field
DocType
contextualized word embeddings, neural ranking
Information retrieval,Ranking,Computer science,Baseline (configuration management),Language model
Conference
ISBN
Citations 
PageRank 
978-1-4503-6172-9
22
0.93
References 
Authors
0
4
Name
Order
Citations
PageRank
Sean MacAvaney16614.50
Andrew Yates2514.61
Arman Cohan313918.25
Nazli Goharian446049.93