Vision Language Models for Dynamic Human Activity Recognition in Healthcare Settings
Abid, Abderrazek, Ho, Thanh-Cong, Karray, Fakhri
–arXiv.org Artificial Intelligence
As generative AI continues to evolve, Vision Language Models (VLMs) have emerged as promising tools in various healthcare applications. One area that remains relatively underexplored is their use in human activity recognition (HAR) for remote health monitoring. VLMs offer notable strengths, including greater flexibility and the ability to overcome some of the constraints of traditional deep learning models. However, a key challenge in applying VLMs to HAR lies in the difficulty of evaluating their dynamic and often non-deterministic outputs. To address this gap, we introduce a descriptive caption data set and propose comprehensive evaluation methods to evaluate VLMs in HAR. Through comparative experiments with state-of-the-art deep learning models, our findings demonstrate that VLMs achieve comparable performance and, in some cases, even surpass conventional approaches in terms of accuracy. This work contributes a strong benchmark and opens new possibilities for the integration of VLMs into intelligent healthcare systems. Code and dataset are available at: https://github.com/gouga10/
arXiv.org Artificial Intelligence
Nov-18-2025
- Country:
- Asia > Middle East
- UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America
- Canada > Ontario
- Waterloo Region > Waterloo (0.04)
- United States (0.04)
- Canada > Ontario
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.87)
- Industry:
- Health & Medicine > Consumer Health (0.50)
- Information Technology > Security & Privacy (0.46)
- Technology: