← smac.pub home

CEDR: Contextualized Embeddings for Document Ranking

pdf bibtex code slides poster 374 citations short conference paper

Authors: Sean MacAvaney, Andrew Yates, Arman Cohan, Nazli Goharian

Appeared in: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2019)

Links/IDs:
DOI 10.1145/3331184.3331317 DBLP conf/sigir/MacAvaneyYCG19 arXiv 1904.07094 Google Scholar 7wWfoDgAAAAJ:eQOLeE2rZwMC Semantic Scholar 1ec78c0ec945572673fabd50bf263870fe9d3601 smac.pub sigir2019-cedr

Abstract:

Although considerable attention has been given to neural ranking architectures recently, far less attention has been paid to the term representations that are used as input to these models. In this work, we investigate how two pretrained contextualized language modes (ELMo and BERT) can be utilized for ad-hoc document ranking. Through experiments on TREC benchmarks, we find that several existing neural ranking architectures can benefit from the additional context provided by contextualized language models. Furthermore, we propose a joint approach that incorporates BERT's classification vector into existing neural models and show that it outperforms state-of-the-art ad-hoc ranking baselines. We also address practical challenges in using these models for ranking, including the maximum input length imposed by BERT and runtime performance impacts of contextualized language models.

BibTeX @inproceedings{macavaney:sigir2019-cedr, author = {MacAvaney, Sean and Yates, Andrew and Cohan, Arman and Goharian, Nazli}, title = {CEDR: Contextualized Embeddings for Document Ranking}, booktitle = {Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval}, year = {2019}, url = {https://arxiv.org/abs/1904.07094}, doi = {10.1145/3331184.3331317}, pages = {1101--1104} }

Tweets