Multimodal Fusion with LLMs for Engagement Prediction in Natural Conversation
Ma, Cheng Charles, Joo, Kevin Hyekang, Vail, Alexandria K., Bhattacharya, Sunreeta, García, Álvaro Fernández, Baker-Matsuoka, Kailana, Mathew, Sheryl, Holt, Lori L., De la Torre, Fernando
–arXiv.org Artificial Intelligence
Over the past decade, wearable computing devices (``smart glasses'') have undergone remarkable advancements in sensor technology, design, and processing power, ushering in a new era of opportunity for high-density human behavior data. Equipped with wearable cameras, these glasses offer a unique opportunity to analyze non-verbal behavior in natural settings as individuals interact. Our focus lies in predicting engagement in dyadic interactions by scrutinizing verbal and non-verbal cues, aiming to detect signs of disinterest or confusion. Leveraging such analyses may revolutionize our understanding of human communication, foster more effective collaboration in professional environments, provide better mental health support through empathetic virtual interactions, and enhance accessibility for those with communication barriers. In this work, we collect a dataset featuring 34 participants engaged in casual dyadic conversations, each providing self-reported engagement ratings at the end of each conversation. We introduce a novel fusion strategy using Large Language Models (LLMs) to integrate multiple behavior modalities into a ``multimodal transcript'' that can be processed by an LLM for behavioral reasoning tasks. Remarkably, this method achieves performance comparable to established fusion techniques even in its preliminary implementation, indicating strong potential for further research and optimization. This fusion method is one of the first to approach ``reasoning'' about real-world human behavior through a language model. Smart glasses provide us the ability to unobtrusively gather high-density multimodal data on human behavior, paving the way for new approaches to understanding and improving human communication with the potential for important societal benefits. The features and data collected during the studies will be made publicly available to promote further research.
arXiv.org Artificial Intelligence
Sep-13-2024
- Country:
- Asia
- Azerbaijan > Karabakh Economic Region
- Shusha District > Shusha (0.04)
- China > Hong Kong (0.04)
- Indonesia > Bali (0.04)
- Middle East
- Jordan (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Singapore (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Azerbaijan > Karabakh Economic Region
- Europe
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States
- Pennsylvania > Allegheny County
- Pittsburgh (0.14)
- New York > New York County
- New York City (0.05)
- Colorado > Boulder County
- Boulder (0.04)
- District of Columbia > Washington (0.04)
- Washington > King County
- Bellevue (0.04)
- Wisconsin > Dane County
- Madison (0.04)
- California > Santa Clara County
- San Jose (0.04)
- Hawaii > Honolulu County
- Honolulu (0.04)
- Texas > Travis County
- Austin (0.14)
- Pennsylvania > Allegheny County
- Canada > Ontario
- Asia
- Genre:
- Questionnaire & Opinion Survey (1.00)
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Industry:
- Health & Medicine > Therapeutic Area (1.00)
- Technology: