Goto

Collaborating Authors

 facial keypoint


Graph-based Online Monitoring of Train Driver States via Facial and Skeletal Features

Nocentini, Olivia, Lagomarsino, Marta, Solak, Gokhan, Cho, Younggeol, Tong, Qiyi, Lorenzini, Marta, Ajoudani, Arash

arXiv.org Artificial Intelligence

--Driver fatigue poses a significant challenge to railway safety, with traditional systems like the dead-man switch offering limited and basic alertness checks. This study presents an online behavior-based monitoring system utilizing a cus-tomised Directed-Graph Neural Network (DGNN) to classify train driver's states into three categories: alert, not alert, and pathological. T o optimize input representations for the model, an ablation study was performed, comparing three feature configurations: skeletal-only, facial-only, and a combination of both. Experimental results show that combining facial and skeletal features yields the highest accuracy (80.88%) in the three-class model, outperforming models using only facial or skeletal features. Furthermore, this combination achieves over 99% accuracy in the binary alertness classification. Additionally, we introduced a novel dataset that, for the first time, incorporates simulated pathological conditions into train driver monitoring, broadening the scope for assessing risks related to fatigue and health. This work represents a step forward in enhancing railway safety through advanced online monitoring using vision-based technologies. Fatigue is a critical safety concern in railway operations, where long shifts and repetitive activities can significantly impair a driver's alertness [1]. Despite regulations aimed at ensuring adequate rest for train drivers, fatigue-related incidents remain alarmingly common. According to research by the AA Charitable Trust, one in eight drivers admits to falling asleep at the wheel, while nearly two-fifths have felt so tired that they feared they might nod off [2]. The risk is particularly high with modern automated trains operating on night shifts and monotonous routes [3].


AI for human assessment: What do professional assessors need?

Arakawa, Riku, Yakura, Hiromu

arXiv.org Artificial Intelligence

Recent organizations have started to adopt AI-based decision support tools to optimize human resource development practices, while facing various challenges of using AIs in highly contextual and sensitive domains. We present our case study that aims to help professional assessors make decisions in human assessment, in which they conduct interviews with assessees and evaluate their suitability for certain job roles. Our workshop with two industrial assessors elucidated troubles they face (i.e., maintaining stable and non-subjective observation of assessees' behaviors) and derived requirements of AI systems (i.e., extracting their nonverbal cues from interview videos in an interpretable manner). In response, we employed an unsupervised anomaly detection algorithm using multimodal behavioral features such as facial keypoints, body and head pose, and gaze. The algorithm extracts outlier scenes from the video based on behavioral features as well as informing which feature contributes to the outlierness. We first evaluated how the assessors would perceive the extracted cues and discovered that the algorithm is useful in suggesting scenes to which assessors would pay attention, thanks to its interpretability. Then, we developed an interface prototype incorporating the algorithm and had six assessors use it for their actual assessment. Their comments revealed the effectiveness of introducing unsupervised anomaly detection to enhance their feeling of confidence and objectivity of the assessment along with potential use scenarios of such AI-based systems in human assessment. Our approach, which builds on top of the idea of separating observation and interpretation in human-AI collaboration, will facilitate human decision making in highly contextual domains, such as human assessment, while keeping their trust in the system.


Visual Detection of Diver Attentiveness for Underwater Human-Robot Interaction

Enan, Sadman Sakib, Sattar, Junaed

arXiv.org Artificial Intelligence

Many underwater tasks, such as cable-and-wreckage inspection, search-and-rescue, benefit from robust human-robot interaction (HRI) capabilities. With the recent advancements in vision-based underwater HRI methods, autonomous underwater vehicles (AUVs) can communicate with their human partners even during a mission. However, these interactions usually require active participation especially from humans (e.g., one must keep looking at the robot during an interaction). Therefore, an AUV must know when to start interacting with a human partner, i.e., if the human is paying attention to the AUV or not. In this paper, we present a diver attention estimation framework for AUVs to autonomously detect the attentiveness of a diver and then navigate and reorient itself, if required, with respect to the diver to initiate an interaction. The core element of the framework is a deep neural network (called DATT-Net) which exploits the geometric relation among 10 facial keypoints of the divers to determine their head orientation. Our on-the-bench experimental evaluations (using unseen data) demonstrate that the proposed DATT-Net architecture can determine the attentiveness of human divers with promising accuracy. Our real-world experiments also confirm the efficacy of DATT-Net which enables real-time inference and allows the AUV to position itself for an AUV-diver interaction.


Pain Intensity Estimation from Mobile Video Using 2D and 3D Facial Keypoints

Lee, Matthew, Kennedy, Lyndon, Girgensohn, Andreas, Wilcox, Lynn, Lee, John Song En, Tan, Chin Wen, Sng, Ban Leong

arXiv.org Machine Learning

For the more than 300 million surgeries performed worldwide every year, managing post-surgical pain is critical for successful surgical outcomes. Pain is the most prominent post-surgical concern, with an estimated 86% of surgical patients in the United States experiencing pain after surgery, with 75% of these patients reporting at least moderate to extreme pain [1]. Higher postoperative pain is associated with more postoperative complications [2], indicating the importance of pain management. Furthermore, the use of opioid analgesics is a powerful tool for managing pain but can pose risks of adverse drug events (experienced by 10% of surgical patients), leading to prolonged length of stay, high hospitalization costs, and potentially addiction [3]. Thus, regular and careful pain assessment is important for balancing between pain relief and potential side effects of powerful opioid analgesics [4]. However, one of the challenges of pain management is accurately assessing the pain level of patients. Pain is inherently subjective, and the personal experience of pain is difficult to observe and measure objectively by those not experiencing it [5] (e.g., care providers). The standard practice used in clinical care requires patients to self-report their pain intensity level using a numeric or visual scale, such as the popular Numerical Pain Rating Scale [6].