About This Document
- sl:arxiv_author :
- sl:arxiv_firstAuthor : Jinhyuk Lee
- sl:arxiv_num : 2109.08133
- sl:arxiv_published : 2021-09-16T17:42:45Z
- sl:arxiv_summary : Dense retrieval methods have shown great promise over sparse retrieval
methods in a range of NLP problems. Among them, dense phrase retrieval-the most
fine-grained retrieval unit-is appealing because phrases can be directly used
as the output for question answering and slot filling tasks. In this work, we
follow the intuition that retrieving phrases naturally entails retrieving
larger text blocks and study whether phrase retrieval can serve as the basis
for coarse-level retrieval including passages and documents. We first observe
that a dense phrase-retrieval system, without any retraining, already achieves
better passage retrieval accuracy (+3-5% in top-5 accuracy) compared to passage
retrievers, which also helps achieve superior end-to-end QA performance with
fewer passages. Then, we provide an interpretation for why phrase-level
supervision helps learn better fine-grained entailment compared to
passage-level supervision, and also show that phrase retrieval can be improved
to achieve competitive performance in document-retrieval tasks such as entity
linking and knowledge-grounded dialogue. Finally, we demonstrate how phrase
filtering and vector quantization can reduce the size of our index by 4-10x,
making dense phrase retrieval a practical and versatile solution in
multi-granularity retrieval.@en
- sl:arxiv_title : Phrase Retrieval Learns Passage Retrieval, Too@en
- sl:arxiv_updated : 2021-09-16T17:42:45Z
- sl:bookmarkOf : https://arxiv.org/abs/2109.08133
- sl:creationDate : 2021-09-30
- sl:creationTime : 2021-09-30T14:50:09Z
- sl:relatedDoc : http://www.semanlink.net/doc/2021/09/princeton_nlp_densephrases_acl
Documents with similar tags (experimental)