corrective feedback
- North America > United States > Arizona > Maricopa County > Tempe (0.04)
- North America > United States > Colorado > Larimer County > Fort Collins (0.04)
- Europe > Czechia > Prague (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- North America > United States > Colorado > Denver County > Denver (0.04)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
What to Say and When to Say it: Live Fitness Coaching as a T estbed for Situated Interaction Sunny Panchal
Vision-language models have shown impressive progress in recent years. However, existing models are largely limited to turn-based interactions, where each turn must be stepped (i.e., prompted) by the user. Open-ended, asynchronous interactions, where an AI model may proactively deliver timely responses or feedback based on the unfolding situation in real-time, are an open challenge.
- North America > United States > California > San Diego County > San Diego (0.04)
- Asia > Indonesia > Bali (0.04)
- Health & Medicine > Consumer Health (0.93)
- Law (0.93)
- Information Technology (0.67)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.95)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
- North America > United States > California > San Diego County > San Diego (0.04)
- Asia > Indonesia > Bali (0.04)
- Health & Medicine > Consumer Health (0.93)
- Law (0.93)
- Information Technology (0.67)
- North America > United States > Arizona > Maricopa County > Tempe (0.04)
- North America > United States > Colorado > Larimer County > Fort Collins (0.04)
- Europe > Czechia > Prague (0.04)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.51)
Large Language Model-Driven Dynamic Assessment of Grammatical Accuracy in English Language Learner Writing
Jaganov, Timur, Blake, John, Villegas, Julián, Carr, Nicholas
This study investigates the potential for Large Language Models (LLMs) to scale-up Dynamic Assessment (DA). To facilitate such an investigation, we first developed DynaWrite-a modular, microservices-based grammatical tutoring application which supports multiple LLMs to generate dynamic feedback to learners of English. Initial testing of 21 LLMs, revealed GPT-4o and neural chat to have the most potential to scale-up DA in the language learning classroom. Further testing of these two candidates found both models performed similarly in their ability to accurately identify grammatical errors in user sentences. However, GPT-4o consistently outperformed neural chat in the quality of its DA by generating clear, consistent, and progressively explicit hints. Real-time responsiveness and system stability were also confirmed through detailed performance testing, with GPT-4o exhibiting sufficient speed and stability. This study shows that LLMs can be used to scale-up dynamic assessment and thus enable dynamic assessment to be delivered to larger groups than possible in traditional teacher-learner settings.
- Asia > Japan (0.05)
- Oceania > Australia > Victoria > Melbourne (0.04)
- Europe > United Kingdom > England > West Midlands > Birmingham (0.04)
- (3 more...)
- Research Report > New Finding (1.00)
- Instructional Material (1.00)
- Education > Curriculum > Subject-Specific Education (0.49)
- Education > Assessment & Standards > Student Performance (0.46)
- Education > Educational Setting > Higher Education (0.46)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- North America > United States > Colorado > Denver County > Denver (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
version of our paper, we shall clarify the details in Section 3 (R2), and make intuition in the methods section much
We thank the reviewers for the detailed comments, suggestions, and a positive assessment of our work. We will correct for color schemes in all figures (R1). We have also made captions of figures cleaner (R3). We have added a description of the setup to the paper. In Fig 5 (left), DisCor actually outperforms Unif( s,a) on these environments.
Annotating Errors in English Learners' Written Language Production: Advancing Automated Written Feedback Systems
Coyne, Steven, Galvan-Sosa, Diana, Spring, Ryan, Guerraoui, Camélia, Zock, Michael, Sakaguchi, Keisuke, Inui, Kentaro
Recent advances in natural language processing (NLP) have contributed to the development of automated writing evaluation (AWE) systems that can correct grammatical errors. However, while these systems are effective at improving text, they are not optimally designed for language learning. They favor direct revisions, often with a click-to-fix functionality that can be applied without considering the reason for the correction. Meanwhile, depending on the error type, learners may benefit most from simple explanations and strategically indirect hints, especially on generalizable grammatical rules. To support the generation of such feedback, we introduce an annotation framework that models each error's error type and generalizability. For error type classification, we introduce a typology focused on inferring learners' knowledge gaps by connecting their errors to specific grammatical patterns. Following this framework, we collect a dataset of annotated learner errors and corresponding human-written feedback comments, each labeled as a direct correction or hint. With this data, we evaluate keyword-guided, keyword-free, and template-guided methods of generating feedback using large language models (LLMs). Human teachers examined each system's outputs, assessing them on grounds including relevance, factuality, and comprehensibility. We report on the development of the dataset and the comparative performance of the systems investigated.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.28)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Europe > France > Provence-Alpes-Côte d'Azur > Bouches-du-Rhône > Marseille (0.04)
- (13 more...)