Kweon, Wonbin
Improving Scientific Document Retrieval with Concept Coverage-based Query Set Generation
Kang, SeongKu, Jin, Bowen, Kweon, Wonbin, Zhang, Yu, Lee, Dongha, Han, Jiawei, Yu, Hwanjo
In specialized fields like the scientific domain, constructing large-scale human-annotated datasets poses a significant challenge due to the need for domain expertise. Recent methods have employed large language models to generate synthetic queries, which serve as proxies for actual user queries. However, they lack control over the content generated, often resulting in incomplete coverage of academic concepts in documents. We introduce Concept Coverage-based Query set Generation (CCQGen) framework, designed to generate a set of queries with comprehensive coverage of the document's concepts. A key distinction of CCQGen is that it adaptively adjusts the generation process based on the previously generated queries. We identify concepts not sufficiently covered by previous queries, and leverage them as conditions for subsequent query generation. This approach guides each new query to complement the previous ones, aiding in a thorough understanding of the document. Extensive experiments demonstrate that CCQGen significantly enhances query quality and retrieval performance.
Uncertainty Quantification and Decomposition for LLM-based Recommendation
Kweon, Wonbin, Jang, Sanghwan, Kang, SeongKu, Yu, Hwanjo
Instruction-tuned for recommendation, we demonstrate that LLMs often exhibit uncertainty LLMs [4, 29, 64, 66] have shown remarkable performance for the in their recommendations. To ensure the trustworthy zero-shot ranking task [23, 25], and can be further fine-tuned with use of LLMs in generating recommendations, we emphasize the the user history logged on the system [2, 19, 81]. Recent methods importance of assessing the reliability of recommendations generated [10, 70, 79, 80] adopt the retrieval-augmented generation paradigm by LLMs. We start by introducing a novel framework for [3, 27], where LLMs are employed to generate ranking lists with candidates estimating the predictive uncertainty to quantitatively measure the retrieved by candidate generators. This approach exhibits reliability of LLM-based recommendations. We further propose to state-of-the-art recommendation performance over conventional decompose the predictive uncertainty into recommendation uncertainty sequential recommenders [31, 63], facilitating better online updates and prompt uncertainty, enabling in-depth analyses of and avoiding hallucination. the primary source of uncertainty. Through extensive experiments, While LLMs have been widely employed in real-world applications we (1) demonstrate predictive uncertainty effectively indicates the that can influence human behavior, there is a lack of exploration reliability of LLM-based recommendations, (2) investigate the origins in assessing the reliability of the LLM-based recommendation. of uncertainty with decomposed uncertainty measures, and Indeed, despite their superior performance, we demonstrate recommendations (3) propose uncertainty-aware prompting for a lower predictive generated by LLMs are highly volatile depending on uncertainty and enhanced recommendation. Our source code and the prompting details (e.g., word choice, number of user histories, model weights are available at https://github.com/WonbinKweon/
Verbosity-Aware Rationale Reduction: Effective Reduction of Redundant Rationale via Principled Criteria
Jang, Joonwon, Kim, Jaehee, Kweon, Wonbin, Yu, Hwanjo
Large Language Models (LLMs) rely on generating extensive intermediate reasoning units (e.g., tokens, sentences) to enhance final answer quality across a wide range of complex tasks. While generating multiple reasoning paths or iteratively refining rationales proves effective for improving performance, these approaches inevitably result in significantly higher inference costs. In this work, we propose a novel sentence-level rationale reduction training framework that leverages likelihood-based criteria, verbosity, to identify and remove redundant reasoning sentences. Unlike previous approaches that utilize token-level reduction, our sentence-level reduction framework maintains model performance while reducing generation length. This preserves the original reasoning abilities of LLMs and achieves an average 17.15% reduction in generation costs across various models and tasks.
Rectifying Demonstration Shortcut in In-Context Learning
Jang, Joonwon, Jang, Sanghwan, Kweon, Wonbin, Jeon, Minjin, Yu, Hwanjo
Large language models (LLMs) are able to solve various tasks with only a few demonstrations utilizing their in-context learning (ICL) abilities. However, LLMs often rely on their pre-trained semantic priors of demonstrations rather than on the input-label relationships to proceed with ICL prediction. In this work, we term this phenomenon as the 'Demonstration Shortcut'. While previous works have primarily focused on improving ICL prediction results for predefined tasks, we aim to rectify the Demonstration Shortcut, thereby enabling the LLM to effectively learn new input-label relationships from demonstrations. To achieve this, we introduce In-Context Calibration, a demonstration-aware calibration method. We evaluate the effectiveness of the proposed method in two settings: (1) the Original ICL Task using the standard label space and (2) the Task Learning setting, where the label space is replaced with semantically unrelated tokens. In both settings, In-Context Calibration demonstrates substantial improvements, with results generalized across three LLM families (OPT, GPT, and Llama2) under various configurations.
Doubly Calibrated Estimator for Recommendation on Data Missing Not At Random
Kweon, Wonbin, Yu, Hwanjo
Recommender systems often suffer from selection bias as users tend to rate their preferred items. The datasets collected under such conditions exhibit entries missing not at random and thus are not randomized-controlled trials representing the target population. To address this challenge, a doubly robust estimator and its enhanced variants have been proposed as they ensure unbiasedness when accurate imputed errors or predicted propensities are provided. However, we argue that existing estimators rely on miscalibrated imputed errors and propensity scores as they depend on rudimentary models for estimation. We provide theoretical insights into how miscalibrated imputation and propensity models may limit the effectiveness of doubly robust estimators and validate our theorems using real-world datasets. On this basis, we propose a Doubly Calibrated Estimator that involves the calibration of both the imputation and propensity models. To achieve this, we introduce calibration experts that consider different logit distributions across users. Moreover, we devise a tri-level joint learning framework, allowing the simultaneous optimization of calibration experts alongside prediction and imputation models. Through extensive experiments on real-world datasets, we demonstrate the superiority of the Doubly Calibrated Estimator in the context of debiased recommendation tasks.
Distillation from Heterogeneous Models for Top-K Recommendation
Kang, SeongKu, Kweon, Wonbin, Lee, Dongha, Lian, Jianxun, Xie, Xing, Yu, Hwanjo
Recent recommender systems have shown remarkable performance by using an ensemble of heterogeneous models. However, it is exceedingly costly because it requires resources and inference latency proportional to the number of models, which remains the bottleneck for production. Our work aims to transfer the ensemble knowledge of heterogeneous teachers to a lightweight student model using knowledge distillation (KD), to reduce the huge inference costs while retaining high accuracy. Through an empirical study, we find that the efficacy of distillation severely drops when transferring knowledge from heterogeneous teachers. Nevertheless, we show that an important signal to ease the difficulty can be obtained from the teacher's training trajectory. This paper proposes a new KD framework, named HetComp, that guides the student model by transferring easy-to-hard sequences of knowledge generated from the teachers' trajectories. To provide guidance according to the student's learning state, HetComp uses dynamic knowledge construction to provide progressively difficult ranking knowledge and adaptive knowledge transfer to gradually transfer finer-grained ranking information. Our comprehensive experiments show that HetComp significantly improves the distillation quality and the generalization of the student model.
Obtaining Calibrated Probabilities with Personalized Ranking Models
Kweon, Wonbin, Kang, SeongKu, Yu, Hwanjo
For personalized ranking models, the well-calibrated probability of an item being preferred by a user has great practical value. While existing work shows promising results in image classification, probability calibration has not been much explored for personalized ranking. In this paper, we aim to estimate the calibrated probability of how likely a user will prefer an item. We investigate various parametric distributions and propose two parametric calibration methods, namely Gaussian calibration and Gamma calibration. Each proposed method can be seen as a post-processing function that maps the ranking scores of pre-trained models to well-calibrated preference probabilities, without affecting the recommendation performance. We also design the unbiased empirical risk minimization framework that guides the calibration methods to learning of true preference probability from the biased user-item interaction dataset. Extensive evaluations with various personalized ranking models on real-world datasets show that both the proposed calibration methods and the unbiased empirical risk minimization significantly improve the calibration performance.