Fine-Grained Distillation for Long Document Retrieval
Zhou, Yucheng, Shen, Tao, Geng, Xiubo, Tao, Chongyang, Long, Guodong, Xu, Can, Jiang, Daxin
–arXiv.org Artificial Intelligence
Long document retrieval aims to fetch query-relevant documents from a large-scale collection, where knowledge distillation has become de facto to improve a retriever by mimicking a heterogeneous yet powerful cross-encoder. However, in contrast to passages or sentences, retrieval on long documents suffers from the scope hypothesis that a long document may cover multiple topics. This maximizes their structure heterogeneity and poses a granular-mismatch issue, leading to an inferior distillation efficacy. In this work, we propose a new learning framework, fine-grained distillation (FGD), for long-document retrievers. While preserving the conventional dense retrieval paradigm, it first produces global-consistent representations crossing different fine granularity and then applies multi-granular aligned distillation merely during training. In experiments, we evaluate our framework on two long-document retrieval benchmarks, which show state-of-the-art performance.
arXiv.org Artificial Intelligence
Dec-20-2022
- Country:
- Africa > Ethiopia (0.04)
- Asia
- China > Hong Kong (0.04)
- Middle East > Jordan (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Europe
- North America
- Canada (0.04)
- Dominican Republic (0.04)
- United States
- California > Los Angeles County
- Long Beach (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Maryland > Baltimore (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- California > Los Angeles County
- Oceania > Australia
- Queensland (0.04)
- Genre:
- Research Report (0.64)
- Industry:
- Education (0.46)
- Technology: