Towards an AI-Driven Video-Based American Sign Language Dictionary: Exploring Design and Usage Experience with Learners
Hassan, Saad, Bohacek, Matyas, Kim, Chaelin, Crochet, Denise
–arXiv.org Artificial Intelligence
Searching for unfamiliar American Sign Language (ASL) signs is challenging for learners because, unlike spoken languages, they cannot type a text-based query to look up an unfamiliar sign. Advances in isolated sign recognition have enabled the creation of video-based dictionaries, allowing users to submit a video and receive a list of the closest matching signs. Previous HCI research using Wizard-of-Oz prototypes has explored interface designs for ASL dictionaries. Building on these studies, we incorporate their design recommendations and leverage state-of-the-art sign-recognition technology to develop an automated video-based dictionary. We also present findings from an observational study with twelve novice ASL learners who used this dictionary during video-comprehension and question-answering tasks. Our results address human-AI interaction challenges not covered in previous WoZ research, including recording and resubmitting signs, unpredictable outputs, system latency, and privacy concerns. These insights offer guidance for designing and deploying video-based ASL dictionary systems.
arXiv.org Artificial Intelligence
Apr-9-2025
- Country:
- Europe (1.00)
- North America > United States (1.00)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Education > Curriculum > Subject-Specific Education (0.37)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning (0.46)
- Natural Language (0.48)
- Vision > Video Understanding (0.34)
- Data Science > Data Mining (0.34)
- Human Computer Interaction (1.00)
- Information Management > Search (0.34)
- Artificial Intelligence
- Information Technology