Saha, Koustuv
AI on My Shoulder: Supporting Emotional Labor in Front-Office Roles with an LLM-based Empathetic Coworker
Swain, Vedant Das, Zhong, Qiuyue "Joy", Parekh, Jash Rajesh, Jeon, Yechan, Zimmerman, Roy, Czerwinski, Mary, Suh, Jina, Mishra, Varun, Saha, Koustuv, Hernandez, Javier
Client-Service Representatives (CSRs) are vital to organizations. Frequent interactions with disgruntled clients, however, disrupt their mental well-being. To help CSRs regulate their emotions while interacting with uncivil clients, we designed Pro-Pilot, an LLM-powered assistant, and evaluated its efficacy, perception, and use. Our comparative analyses between 665 human and Pro-Pilot-generated support messages demonstrate Pro-Pilot's ability to adapt to and demonstrate empathy in various incivility incidents. Additionally, 143 CSRs assessed Pro-Pilot's empathy as more sincere and actionable than human messages. Finally, we interviewed 20 CSRs who interacted with Pro-Pilot in a simulation exercise. They reported that Pro-Pilot helped them avoid negative thinking, recenter thoughts, and humanize clients; showing potential for bridging gaps in coworker support. Yet, they also noted deployment challenges and emphasized the irreplaceability of shared experiences. We discuss future designs and societal implications of AI-mediated emotional labor, underscoring empathy as a critical function for AI assistants in front-office roles.
SLM-Mod: Small Language Models Surpass LLMs at Content Moderation
Zhan, Xianyang, Goyal, Agam, Chen, Yilun, Chandrasekharan, Eshwar, Saha, Koustuv
Large language models (LLMs) have shown promise in many natural language understanding tasks, including content moderation. However, these models can be expensive to query in real-time and do not allow for a community-specific approach to content moderation. To address these challenges, we explore the use of open-source small language models (SLMs) for community-specific content moderation tasks. We fine-tune and evaluate SLMs (less than 15B parameters) by comparing their performance against much larger open- and closed-sourced models. Using 150K comments from 15 popular Reddit communities, we find that SLMs outperform LLMs at content moderation -- 11.5% higher accuracy and 25.7% higher recall on average across all communities. We further show the promise of cross-community content moderation, which has implications for new communities and the development of cross-platform moderation techniques. Finally, we outline directions for future work on language model based content moderation. Code and links to HuggingFace models can be found at https://github.com/AGoyal0512/SLM-Mod.
Using Audio Data to Facilitate Depression Risk Assessment in Primary Health Care
Levinson, Adam Valen, Goyal, Abhay, Man, Roger Ho Chun, Lee, Roy Ka-Wei, Saha, Koustuv, Parekh, Nimay, Altice, Frederick L., Cheung, Lam Yin, De Choudhury, Munmun, Kumar, Navin
Telehealth is a valuable tool for primary health care (PHC), where depression is a common condition. PHC is the first point of contact for most people with depression, but about 25% of diagnoses made by PHC physicians are inaccurate. Many other barriers also hinder depression detection and treatment in PHC. Artificial intelligence (AI) may help reduce depression misdiagnosis in PHC and improve overall diagnosis and treatment outcomes. Telehealth consultations often have video issues, such as poor connectivity or dropped calls. Audio-only telehealth is often more practical for lower-income patients who may lack stable internet connections. Thus, our study focused on using audio data to predict depression risk. The objectives were to: 1) Collect audio data from 24 people (12 with depression and 12 without mental health or major health condition diagnoses); 2) Build a machine learning model to predict depression risk. TPOT, an autoML tool, was used to select the best machine learning algorithm, which was the K-nearest neighbors classifier. The selected model had high performance in classifying depression risk (Precision: 0.98, Recall: 0.93, F1-Score: 0.96). These findings may lead to a range of tools to help screen for and treat depression. By developing tools to detect depression risk, patients can be routed to AI-driven chatbots for initial screenings. Partnerships with a range of stakeholders are crucial to implementing these solutions. Moreover, ethical considerations, especially around data privacy and potential biases in AI models, need to be at the forefront of any AI-driven intervention in mental health care.
Can Workers Meaningfully Consent to Workplace Wellbeing Technologies?
Chowdhary, Shreya, Kawakami, Anna, Gray, Mary L., Suh, Jina, Olteanu, Alexandra, Saha, Koustuv
Sensing technologies deployed in the workplace can unobtrusively collect detailed data about individual activities and group interactions that are otherwise difficult to capture. A hopeful application of these technologies is that they can help businesses and workers optimize productivity and wellbeing. However, given the workplace's inherent and structural power dynamics, the prevalent approach of accepting tacit compliance to monitor work activities rather than seeking workers' meaningful consent raises privacy and ethical concerns. This paper unpacks the challenges workers face when consenting to workplace wellbeing technologies. Using a hypothetical case to prompt reflection among six multi-stakeholder focus groups involving 15 participants, we explored participants' expectations and capacity to consent to these technologies. We sketched possible interventions that could better support meaningful consent to workplace wellbeing technologies by drawing on critical computing and feminist scholarship -- which reframes consent from a purely individual choice to a structural condition experienced at the individual level that needs to be freely given, reversible, informed, enthusiastic, and specific (FRIES). The focus groups revealed how workers are vulnerable to "meaningless" consent -- as they may be subject to power dynamics that minimize their ability to withhold consent and may thus experience an erosion of autonomy, also undermining the value of data gathered in the name of "wellbeing." To meaningfully consent, participants wanted changes to the technology and to the policies and practices surrounding the technology. Our mapping of what prevents workers from meaningfully consenting to workplace wellbeing technologies (challenges) and what they require to do so (interventions) illustrates how the lack of meaningful consent is a structural problem requiring socio-technical solutions.
Mental Health Coping Stories on Social Media: A Causal-Inference Study of Papageno Effect
Yuan, Yunhao, Saha, Koustuv, Keller, Barbara, Isometsรค, Erkki Tapio, Aledavood, Talayeh
A considerable amount of literature [16, 25, 49] has studied The Papageno effect concerns how media can play a positive role and re-confirmed the harmful effect of media, dubbed the "Werther in preventing and mitigating suicidal ideation and behaviors. With effect" [38], describing a spike in suicides after a heavily publicized the increasing ubiquity and widespread use of social media, individuals suicide. However, there is much less research about the beneficial often express and share lived experiences and struggles effects of media, referred to as the "Papageno effect", describing a decrease with mental health. However, there is a gap in our understanding in suicides after reporting alternatives to suicide. Niederkrotenthaler about the existence and effectiveness of the Papageno effect in social et al. explored the possible protective effect of media media, which we study in this paper. In particular, we adopt a reporting about suicide [34]. This study finds a decrease in suicides, causal-inference framework to examine the impact of exposure to if reports of suicide related content portray ways of overcoming mental health coping stories on individuals on Twitter. We obtain suicidal ideation without narrating suicidal behaviors.
Charting the Sociotechnical Gap in Explainable AI: A Framework to Address the Gap in XAI
Ehsan, Upol, Saha, Koustuv, De Choudhury, Munmun, Riedl, Mark O.
Explainable AI (XAI) systems are sociotechnical in nature; thus, they are subject to the sociotechnical gap--divide between the technical affordances and the social needs. However, charting this gap is challenging. In the context of XAI, we argue that charting the gap improves our problem understanding, which can reflexively provide actionable insights to improve explainability. Utilizing two case studies in distinct domains, we empirically derive a framework that facilitates systematic charting of the sociotechnical gap by connecting AI guidelines in the context of XAI and elucidating how to use them to address the gap. We apply the framework to a third case in a new domain, showcasing its affordances. Finally, we discuss conceptual implications of the framework, share practical considerations in its operationalization, and offer guidance on transferring it to new contexts. By making conceptual and practical contributions to understanding the sociotechnical gap in XAI, the framework expands the XAI design space.
Jointly Predicting Job Performance, Personality, Cognitive Ability, Affect, and Well-Being
Robles-Granda, Pablo, Lin, Suwen, Wu, Xian, D'Mello, Sidney, Martinez, Gonzalo J., Saha, Koustuv, Nies, Kari, Mark, Gloria, Campbell, Andrew T., De Choudhury, Munmun, Dey, Anind D., Gregg, Julie, Grover, Ted, Mattingly, Stephen M., Mirjafari, Shayan, Moskal, Edward, Striegel, Aaron, Chawla, Nitesh V.
Assessment of job performance, personalized health and psychometric measures are domains where data-driven and ubiquitous computing exhibits the potential of a profound impact in the future. Existing techniques use data extracted from questionnaires, sensors (wearable, computer, etc.), or other traits, to assess well-being and cognitive attributes of individuals. However, these techniques can neither predict individual's well-being and psychological traits in a global manner nor consider the challenges associated to processing the data available, that is incomplete and noisy. In this paper, we create a benchmark for predictive analysis of individuals from a perspective that integrates: physical and physiological behavior, psychological states and traits, and job performance. We design data mining techniques as benchmark and uses real noisy and incomplete data derived from wearable sensors to predict 19 constructs based on 12 standardized well-validated tests. The study included 757 participants who were knowledge workers in organizations across the USA with varied work roles. We developed a data mining framework to extract the meaningful predictors for each of the 19 variables under consideration. Our model is the first benchmark that combines these various instrument-derived variables in a single framework to understand people's behavior by leveraging real uncurated data from wearable, mobile, and social media sources. We verify our approach experimentally using the data obtained from our longitudinal study. The results show that our framework is consistently reliable and capable of predicting the variables under study better than the baselines when prediction is restricted to the noisy, incomplete data.
A Social Media Based Examination of the Effects of Counseling Recommendations after Student Deaths on College Campuses
Saha, Koustuv (Georgia Institute of Technology) | Weber, Ingmar (Qatar Computing Research Institute, HBKU) | Choudhury, Munmun De (Georgia Institute of Technology)
Student deaths on college campuses, whether brought about by a suicide or an uncontrollable incident, have serious repercussions for the mental wellbeing of students. Consequently, many campus administrators implement post-crisis intervention measures to promote student-centric mental health support. Information about these measures, which we refer to as "counseling recommendations", are often shared via electronic channels, including social media. However, the current ability to assess the effects of these recommendations on post-crisis psychological states is limited. We propose a causal analysis framework to examine the effects of these counseling recommendations after student deaths. We leverage a dataset from 174 Reddit campus communities and ~400M posts of ~350K users. Then we employ statistical modeling and natural language analysis to quantify the psychosocial shifts in behavioral, cognitive, and affective expression of grief in individuals who are "exposed" to (comment on) the counseling recommendations, compared to that in a matched control cohort. Drawing on crisis and psychology research, we find that the exposed individuals show greater grief, psycholinguistic, and social expressiveness, providing evidence of a healing response to crisis and thereby positive psychological effects of the counseling recommendations. We discuss the implications of our work in supporting post-crisis rehabilitation and intervention efforts on college campuses.