knowledge candidate
609c5e5089a9aa967232aba2a4d03114-AuthorFeedback.pdf
Effect of UniLM: We observe obvious performance drop when using fine-tuned UniLM2 with fixed top-1 retrieved knowledge (-parameterized posterior in Table 3). (aligned with the columns of Table 1). Our model performs implicit knowledge selection on the input K14 knowledgesentences(concatenatedinasequence)inanend-to-endwaylikeDRD[52]. F1 on the validation set increases until the number of knowledge reaches 10, but22 stays stable when the number increases from 10 to 30. Intesttime,knowledge28 selection module is mainly to shorten the input sequence of knowledge candidates, so the performance drop is not29 significant.
Multimodal Reranking for Knowledge-Intensive Visual Question Answering
Wen, Haoyang, Zhuang, Honglei, Zamani, Hamed, Hauptmann, Alexander, Bendersky, Michael
Knowledge-intensive visual question answering requires models to effectively use external knowledge to help answer visual questions. A typical pipeline includes a knowledge retriever and an answer generator. However, a retriever that utilizes local information, such as an image patch, may not provide reliable question-candidate relevance scores. Besides, the two-tower architecture also limits the relevance score modeling of a retriever to select top candidates for answer generator reasoning. In this paper, we introduce an additional module, a multi-modal reranker, to improve the ranking quality of knowledge candidates for answer generation. Our reranking module takes multi-modal information from both candidates and questions and performs cross-item interaction for better relevance score modeling. Experiments on OK-VQA and A-OKVQA show that multi-modal reranker from distant supervision provides consistent improvements. We also find a training-testing discrepancy with reranking in answer generation, where performance improves if training knowledge candidates are similar to or noisier than those used in testing.
- North America > United States > Washington > King County > Seattle (0.14)
- Asia > Taiwan > Taiwan Province > Taipei (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (15 more...)
CET2: Modelling Topic Transitions for Coherent and Engaging Knowledge-Grounded Conversations
Xu, Lin, Zhou, Qixian, Fu, Jinlan, Ng, See-Kiong
Knowledge-grounded dialogue systems aim to generate coherent and engaging responses based on the dialogue contexts and selected external knowledge. Previous knowledge selection methods tend to rely too heavily on the dialogue contexts or over-emphasize the new information in the selected knowledge, resulting in the selection of repetitious or incongruous knowledge and further generating repetitive or incoherent responses, as the generation of the response depends on the chosen knowledge. To address these shortcomings, we introduce a Coherent and Engaging Topic Transition (CET2) framework to model topic transitions for selecting knowledge that is coherent to the context of the conversations while providing adequate knowledge diversity for topic development. Our CET2 framework considers multiple factors for knowledge selection, including valid transition logic from dialogue contexts to the following topics and systematic comparisons between available knowledge candidates. Extensive experiments on two public benchmarks demonstrate the superiority and the better generalization ability of CET2 on knowledge selection. This is due to our well-designed transition features and comparative knowledge selection strategy, which are more transferable to conversations about unseen topics. Analysis of fine-grained knowledge selection accuracy also shows that CET2 can better balance topic entailment (contextual coherence) and development (knowledge diversity) in dialogue than existing approaches.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Singapore (0.05)
- North America > United States > Washington > King County > Seattle (0.04)
- (7 more...)
- Textiles, Apparel & Luxury Goods (0.68)
- Education (0.46)
Diverse and Faithful Knowledge-Grounded Dialogue Generation via Sequential Posterior Inference
Xu, Yan, Kong, Deqian, Xu, Dehong, Ji, Ziwei, Pang, Bo, Fung, Pascale, Wu, Ying Nian
The capability to generate responses with diversity and faithfulness using factual knowledge is paramount for creating a human-like, trustworthy dialogue system. Common strategies either adopt a two-step paradigm, which optimizes knowledge selection and response generation separately, and may overlook the inherent correlation between these two tasks, or leverage conditional variational method to jointly optimize knowledge selection and response generation by employing an inference network. In this paper, we present an end-to-end learning framework, termed Sequential Posterior Inference (SPI), capable of selecting knowledge and generating dialogues by approximately sampling from the posterior distribution. Unlike other methods, SPI does not require the inference network or assume a simple geometry of the posterior distribution. This straightforward and intuitive inference procedure of SPI directly queries the response generation model, allowing for accurate knowledge selection and generation of faithful responses. In addition to modeling contributions, our experimental results on two common dialogue datasets (Wizard of Wikipedia and Holl-E) demonstrate that SPI outperforms previous strong baselines according to both automatic and human evaluation metrics.
- Asia > China > Hong Kong (0.04)
- Oceania > Australia (0.04)
- North America > United States > Hawaii > Honolulu County > Honolulu (0.04)
- (3 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models (0.68)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.68)
Kformer: Knowledge Injection in Transformer Feed-Forward Layers
Yao, Yunzhi, Huang, Shaohan, Dong, Li, Wei, Furu, Chen, Huajun, Zhang, Ningyu
Recent days have witnessed a diverse set of knowledge injection models for pre-trained language models (PTMs); however, most previous studies neglect the PTMs' own ability with quantities of implicit knowledge stored in parameters. A recent study has observed knowledge neurons in the Feed Forward Network (FFN), which are responsible for expressing factual knowledge. In this work, we propose a simple model, Kformer, which takes advantage of the knowledge stored in PTMs and external knowledge via knowledge injection in Transformer FFN layers. Empirically results on two knowledge-intensive tasks, commonsense reasoning (i.e., SocialIQA) and medical question answering (i.e., MedQA-USMLE), demonstrate that Kformer can yield better performance than other knowledge injection technologies such as concatenation or attention-based injection. We think the proposed simple model and empirical findings may be helpful for the community to develop more powerful knowledge injection methods. Code available in https://github.com/zjunlp/Kformer.
- North America > United States (0.14)
- Asia > Taiwan (0.04)
- Asia > China > Zhejiang Province > Ningbo (0.04)
- (3 more...)