About This Document
- sl:arxiv_author :
- sl:arxiv_firstAuthor : Fabio Petroni
- sl:arxiv_num : 2009.02252
- sl:arxiv_published : 2020-09-04T15:32:19Z
- sl:arxiv_summary : Challenging problems such as open-domain question answering, fact checking,
slot filling and entity linking require access to large, external knowledge
sources. While some models do well on individual tasks, developing general
models is difficult as each task might require computationally expensive
indexing of custom knowledge sources, in addition to dedicated infrastructure.
To catalyze research on models that condition on specific information in large
textual resources, we present a benchmark for knowledge-intensive language
tasks (KILT). All tasks in KILT are grounded in the same snapshot of Wikipedia,
reducing engineering turnaround through the re-use of components, as well as
accelerating research into task-agnostic memory architectures. We test both
task-specific and general baselines, evaluating downstream performance in
addition to the ability of the models to provide provenance. We find that a
shared dense vector index coupled with a seq2seq model is a strong baseline,
outperforming more tailor-made approaches for fact checking, open-domain
question answering and dialogue, and yielding competitive results on entity
linking and slot filling, by generating disambiguated text. KILT data and code
are available at https://github.com/facebookresearch/KILT.@en
- sl:arxiv_title : KILT: a Benchmark for Knowledge Intensive Language Tasks@en
- sl:arxiv_updated : 2021-05-27T15:20:59Z
- sl:bookmarkOf : https://arxiv.org/abs/2009.02252
- sl:creationDate : 2022-01-23
- sl:creationTime : 2022-01-23T18:25:25Z
Documents with similar tags (experimental)