single prompt
Joint Extraction Matters: Prompt-Based Visual Question Answering for Multi-Field Document Information Extraction
Visual question answering (VQA) has emerged as a flexible approach for extracting specific pieces of information from document images. However, existing work typically queries each field in isolation, overlooking potential dependencies across multiple items. This paper investigates the merits of extracting multiple fields jointly versus separately. Through experiments on multiple large vision language models and datasets, we show that jointly extracting fields often improves accuracy, especially when the fields share strong numeric or contextual dependencies. We further analyze how performance scales with the number of requested items and use a regression based metric to quantify inter field relationships. Our results suggest that multi field prompts can mitigate confusion arising from similar surface forms and related numeric values, providing practical methods for designing robust VQA systems in document information extraction tasks.
- North America > United States > Washington > King County > Seattle (0.04)
- North America > Mexico > Mexico City > Mexico City (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Information Extraction (0.91)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.71)
Strategic Insights in Human and Large Language Model Tactics at Word Guessing Games
Rikters, Matīss, Reinsone, Sanita
At the beginning of 2022, a simplistic word-guessing game took the world by storm and was further adapted to many languages beyond the original English version. In this paper, we examine the strategies of daily word-guessing game players that have evolved during a period of over two years. A survey gathered from 25% of frequent players reveals their strategies and motivations for continuing the daily journey. We also explore the capability of several popular open-access large language model systems and open-source models at comprehending and playing the game in two different languages. Results highlight the struggles of certain models to maintain correct guess length and generate repetitions, as well as hallucinations of non-existent words and inflections.
Predicting the Quality of Revisions in Argumentative Writing
Liu, Zhexiong, Litman, Diane, Wang, Elaine, Matsumura, Lindsay, Correnti, Richard
The ability to revise in response to feedback is critical to students' writing success. In the case of argument writing in specific, identifying whether an argument revision (AR) is successful or not is a complex problem because AR quality is dependent on the overall content of an argument. For example, adding the same evidence sentence could strengthen or weaken existing claims in different argument contexts (ACs). To address this issue we developed Chain-of-Thought prompts to facilitate ChatGPT-generated ACs for AR quality predictions. The experiments on two corpora, our annotated elementary essays and existing college essays benchmark, demonstrate the superiority of the proposed ACs over baselines.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (6 more...)
- Health & Medicine > Therapeutic Area (0.46)
- Education > Curriculum > Subject-Specific Education (0.34)
3 Tips to reduce OpenAI GPT-3's costs by Smart Prompting
GPT-3's highest and the most accurate model Davinci costs 6 cents for every 1000 tokens. So it isn't really inexpensive to operate at scale in a production app. So beyond designing prompts, it is essential to even master the craft of smart prompting, that is to reduce the number of tokens in the input prompt. In this tutorial, we will see a few techniques to reduce the number of tokens in a given prompt from my experience of building supermeme.ai, And remember every 1000 tokens reduced is 6-cents (0.06$) saved, so at scale this is huge.
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.83)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.83)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.41)