lamer
Diagnosing and Remedying Knowledge Deficiencies in LLMs via Label-free Curricular Meaningful Learning
Xiong, Kai, Ding, Xiao, Du, Li, Ying, Jiahao, Liu, Ting, Qin, Bing, Cao, Yixin
Large Language Models (LLMs) are versatile and demonstrate impressive generalization ability by mining and learning information from extensive unlabeled text. However, they still exhibit reasoning mistakes, often stemming from knowledge deficiencies, which can affect their trustworthiness and reliability. Although users can provide diverse and comprehensive queries, obtaining sufficient and effective feedback is demanding. Furthermore, evaluating LLMs comprehensively with limited labeled samples is difficult. This makes it a challenge to diagnose and remedy the deficiencies of LLMs through rich label-free user queries. To tackle this challenge, we propose a label-free curricular meaningful learning framework (LaMer). LaMer first employs relative entropy to automatically diagnose and quantify the knowledge deficiencies of LLMs in a label-free setting. Next, to remedy the diagnosed knowledge deficiencies, we apply curricular meaningful learning: first, we adopt meaningful learning to adaptively synthesize augmentation data according to the severity of the deficiencies, and then design a curricular deficiency remedy strategy to remedy the knowledge deficiencies of LLMs progressively. Experiments show that LaMer efficiently and effectively diagnoses and remedies knowledge deficiencies in LLMs, improving various LLMs across seven out-of-distribution (OOD) reasoning and language understanding benchmarks, achieving comparable results to baselines with just 40\% training data. LaMer even surpasses methods that rely on labeled datasets for deficiency diagnosis. In application, our label-free method can offer an effective knowledge deficiency diagnostic tool for efficient LLM development.
Large Language Models are Strong Zero-Shot Retriever
Shen, Tao, Long, Guodong, Geng, Xiubo, Tao, Chongyang, Zhou, Tianyi, Jiang, Daxin
In this work, we propose a simple method that applies a large language model (LLM) to large-scale retrieval in zero-shot scenarios. Our method, the Language language model as Retriever (LameR), is built upon no other neural models but an LLM, while breaking brute-force combinations of retrievers with LLMs and lifting the performance of zero-shot retrieval to be very competitive on benchmark datasets. Essentially, we propose to augment a query with its potential answers by prompting LLMs with a composition of the query and the query's in-domain candidates. The candidates, regardless of correct or wrong, are obtained by a vanilla retrieval procedure on the target collection. As a part of the prompts, they are likely to help LLM generate more precise answers by pattern imitation or candidate summarization. Even if all the candidates are wrong, the prompts at least make LLM aware of in-collection patterns and genres. Moreover, due to the low performance of a self-supervised retriever, the LLM-based query augmentation becomes less effective as the retriever bottlenecks the whole pipeline. Therefore, we propose to leverage a non-parametric lexicon-based method (e.g., BM25) as the retrieval module to capture query-document overlap in a literal fashion. As such, LameR makes the retrieval procedure transparent to the LLM, thus circumventing the performance bottleneck.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- (10 more...)
This Laser-Firing Truck Could Help Make Hot Cities More Livable
When you go on a road trip, you pack snacks and drinks and make sure you have good music to queue. Climate scientist Katia Lamer, on the other hand, packs party balloons loaded with atmospheric sensors, then climbs into a laser-firing observatory on wheels. Lamer--director of operations at the Brookhaven National Laboratory's Center for Multiscale Applied Sensing--recently completed a 1,700-mile road trip from Upton, New York, to Houston, Texas, in a specially designed science truck while taking a bevy of measurements, from air temperature to humidity to wind. The big plan: better understanding the complex climate dynamics of cities, where conditions can vary wildly not only from neighborhood to neighborhood, but door to door. "The big difference with urban environments is that they're much more heterogeneous than natural environments. What that means is that there are more elements, like individual buildings, that create these canyons," says Lamer, referring to the corridors between structures.
- North America > United States > Texas > Harris County > Houston (0.26)
- North America > United States > New York (0.26)