Goto

Collaborating Authors

 single prompt


Joint Extraction Matters: Prompt-Based Visual Question Answering for Multi-Field Document Information Extraction

Loem, Mengsay, Hosaka, Taiju

arXiv.org Artificial Intelligence

Visual question answering (VQA) has emerged as a flexible approach for extracting specific pieces of information from document images. However, existing work typically queries each field in isolation, overlooking potential dependencies across multiple items. This paper investigates the merits of extracting multiple fields jointly versus separately. Through experiments on multiple large vision language models and datasets, we show that jointly extracting fields often improves accuracy, especially when the fields share strong numeric or contextual dependencies. We further analyze how performance scales with the number of requested items and use a regression based metric to quantify inter field relationships. Our results suggest that multi field prompts can mitigate confusion arising from similar surface forms and related numeric values, providing practical methods for designing robust VQA systems in document information extraction tasks.


Strategic Insights in Human and Large Language Model Tactics at Word Guessing Games

Rikters, Matīss, Reinsone, Sanita

arXiv.org Artificial Intelligence

At the beginning of 2022, a simplistic word-guessing game took the world by storm and was further adapted to many languages beyond the original English version. In this paper, we examine the strategies of daily word-guessing game players that have evolved during a period of over two years. A survey gathered from 25% of frequent players reveals their strategies and motivations for continuing the daily journey. We also explore the capability of several popular open-access large language model systems and open-source models at comprehending and playing the game in two different languages. Results highlight the struggles of certain models to maintain correct guess length and generate repetitions, as well as hallucinations of non-existent words and inflections.


Predicting the Quality of Revisions in Argumentative Writing

Liu, Zhexiong, Litman, Diane, Wang, Elaine, Matsumura, Lindsay, Correnti, Richard

arXiv.org Artificial Intelligence

The ability to revise in response to feedback is critical to students' writing success. In the case of argument writing in specific, identifying whether an argument revision (AR) is successful or not is a complex problem because AR quality is dependent on the overall content of an argument. For example, adding the same evidence sentence could strengthen or weaken existing claims in different argument contexts (ACs). To address this issue we developed Chain-of-Thought prompts to facilitate ChatGPT-generated ACs for AR quality predictions. The experiments on two corpora, our annotated elementary essays and existing college essays benchmark, demonstrate the superiority of the proposed ACs over baselines.


3 Tips to reduce OpenAI GPT-3's costs by Smart Prompting

#artificialintelligence

GPT-3's highest and the most accurate model Davinci costs 6 cents for every 1000 tokens. So it isn't really inexpensive to operate at scale in a production app. So beyond designing prompts, it is essential to even master the craft of smart prompting, that is to reduce the number of tokens in the input prompt. In this tutorial, we will see a few techniques to reduce the number of tokens in a given prompt from my experience of building supermeme.ai, And remember every 1000 tokens reduced is 6-cents (0.06$) saved, so at scale this is huge.