About This Document
- sl:arxiv_author :
- sl:arxiv_firstAuthor : Ledell Wu
- sl:arxiv_num : 1911.03814
- sl:arxiv_published : 2019-11-10T01:01:45Z
- sl:arxiv_summary : This paper introduces a conceptually simple, scalable, and highly effective
BERT-based entity linking model, along with an extensive evaluation of its
accuracy-speed trade-off. We present a two-stage zero-shot linking algorithm,
where each entity is defined only by a short textual description. The first
stage does retrieval in a dense space defined by a bi-encoder that
independently embeds the mention context and the entity descriptions. Each
candidate is then re-ranked with a cross-encoder, that concatenates the mention
and entity text. Experiments demonstrate that this approach is state of the art
on recent zero-shot benchmarks (6 point absolute gains) and also on more
established non-zero-shot evaluations (e.g. TACKBP-2010), despite its relative
simplicity (e.g. no explicit entity embeddings or manually engineered mention
tables). We also show that bi-encoder linking is very fast with nearest
neighbour search (e.g. linking with 5.9 million candidates in 2 milliseconds),
and that much of the accuracy gain from the more expensive cross-encoder can be
transferred to the bi-encoder via knowledge distillation. Our code and models
are available at https://github.com/facebookresearch/BLINK.@en
- sl:arxiv_title : Scalable Zero-shot Entity Linking with Dense Entity Retrieval@en
- sl:arxiv_updated : 2020-09-29T08:13:47Z
- sl:bookmarkOf : https://arxiv.org/abs/1911.03814
- sl:creationDate : 2020-05-02
- sl:creationTime : 2020-05-02T11:43:47Z
Documents with similar tags (experimental)