Goto

Collaborating Authors

 suicidal ideation


Young Mormons Built an App to Help Men Quit Gooning

WIRED

The Relay app allows users to track their porn-free streaks and get group support. Its creators say they're taking a stand against porn and AI erotica. Jamie would meticulously schedule his days around finding time alone to watch porn and masturbate--often up to five times a day. The 32-year-old Michigan engineer, who did not want to use his real name due to privacy concerns, first watched porn at the impressionable age of 12, but never realized he had a problem until just after his father's funeral three years ago. "I didn't shed a single tear," he says.


Detecting Suicidal Ideation in Text with Interpretable Deep Learning: A CNN-BiGRU with Attention Mechanism

Bhuiyan, Mohaiminul Islam, Kamarudin, Nur Shazwani, Ismail, Nur Hafieza

arXiv.org Artificial Intelligence

Worldwide, suicide is the second leading cause of death for adolescents with past suicide attempts to be an important predictor for increased future suicides. While some people with suicidal thoughts may try to suppress them, many signal their intentions in social media platforms. To address these issues, we propose a new type of hybrid deep learning scheme, i.e., the combination of a CNN architecture and a BiGRU technique, which can accurately identify the patterns of suicidal ideation from SN datasets. Also, we apply Explainable AI methods using SHapley Additive exPlanations to interpret the prediction results and verifying the model reliability. This integration of CNN local feature extraction, BiGRU bidirectional sequence modeling, attention mechanisms, and SHAP interpretability provides a comprehensive framework for suicide detection. Training and evaluation of the system were performed on a publicly available dataset. Several performance metrics were used for evaluating model performance. Our method was found to have achieved 93.97 accuracy in experimental results. Comparative study to different state-of-the-art Machine Learning and DL models and existing literature demonstrates the superiority of our proposed technique over all the competing methods.


multiMentalRoBERTa: A Fine-tuned Multiclass Classifier for Mental Health Disorder

Islam, K M Sajjadul, Fields, John, Madiraju, Praveen

arXiv.org Artificial Intelligence

The early detection of mental health disorders from social media text is critical for enabling timely support, risk assessment, and referral to appropriate resources. This work introduces multiMentalRoBERTa, a fine-tuned RoBERTa model designed for multiclass classification of common mental health conditions, including stress, anxiety, depression, post-traumatic stress disorder (PTSD), suicidal ideation, and neutral discourse. Drawing on multiple curated datasets, data exploration is conducted to analyze class overlaps, revealing strong correlations between depression and suicidal ideation as well as anxiety and PTSD, while stress emerges as a broad, overlapping category. Comparative experiments with traditional machine learning methods, domain-specific transformers, and prompting-based large language models demonstrate that multiMentalRoBERTa achieves superior performance, with macro F1-scores of 0.839 in the six-class setup and 0.870 in the five-class setup (excluding stress), outperforming both fine-tuned MentalBERT and baseline classifiers. Beyond predictive accuracy, explainability methods, including Layer Integrated Gradients and KeyBERT, are applied to identify lexical cues that drive classification, with a particular focus on distinguishing depression from suicidal ideation. The findings emphasize the effectiveness of fine-tuned transformers for reliable and interpretable detection in sensitive contexts, while also underscoring the importance of fairness, bias mitigation, and human-in-the-loop safety protocols. Overall, multiMentalRoBERTa is presented as a lightweight, robust, and deployable solution for enhancing support in mental health platforms.


Has OpenAI really made ChatGPT better for users with mental health problems?

The Guardian

ChatGPT on App Store displayed on a phone screen on 07 June 2025. ChatGPT on App Store displayed on a phone screen on 07 June 2025. Has OpenAI really made ChatGPT better for users with mental health problems? Prompts indicating suicidal ideation got alarming replies, which experts say shows'how easy it is to break the model' A n OpenAI statement released this week claimed the company had made its popular service ChatGPT better at supporting users experiencing mental health problems like suicidal ideation or delusions, but experts tell the Guardian they need to do more to truly ensure users are protected. The Guardian tested several prompts indicating suicidal ideation with the ChatGPT GPT-5 updated model, which is now the default, and got alarming responses from the large language model (LLM) chatbot.


Detecting Early and Implicit Suicidal Ideation via Longitudinal and Information Environment Signals on Social Media

Shimgekar, Soorya Ram, Zhao, Ruining, Goyal, Agam, Rodriguez, Violeta J., Bloom, Paul A., Sundaram, Hari, Saha, Koustuv

arXiv.org Artificial Intelligence

On social media, many individuals experiencing suicidal ideation (SI) do not disclose their distress explicitly. Instead, signs may surface indirectly through everyday posts or peer interactions. Detecting such implicit signals early is critical but remains challenging. We frame early and implicit SI as a forward-looking prediction task and develop a computational framework that models a user's information environment, consisting of both their longitudinal posting histories as well as the discourse of their socially proximal peers. We adopted a composite network centrality measure to identify top neighbors of a user, and temporally aligned the user's and neighbors' interactions -- integrating the multi-layered signals in a fine-tuned DeBERTa-v3 model. In a Reddit study of 1,000 (500 Case and 500 Control) users, our approach improves early and implicit SI detection by 15% over individual-only baselines. These findings highlight that peer interactions offer valuable predictive signals and carry broader implications for designing early detection systems that capture indirect as well as masked expressions of risk in online environments.


Two-Stage Voting for Robust and Efficient Suicide Risk Detection on Social Media

Song, Yukai, Zhou, Pengfei, Escobar-Viera, César, Biernesser, Candice, Huang, Wei, Hu, Jingtong

arXiv.org Artificial Intelligence

Suicide rates have risen worldwide in recent years, underscoring the urgent need for proactive prevention strategies. Social media provides valuable signals, as many at-risk individuals - who often avoid formal help due to stigma - choose instead to share their distress online. Yet detecting implicit suicidal ideation, conveyed indirectly through metaphor, sarcasm, or subtle emotional cues, remains highly challenging. Lightweight models like BERT handle explicit signals but fail on subtle implicit ones, while large language models (LLMs) capture nuance at prohibitive computational cost. To address this gap, we propose a two-stage voting architecture that balances efficiency and robustness. In Stage 1, a lightweight BERT classifier rapidly resolves high-confidence explicit cases. In Stage 2, ambiguous inputs are escalated to either (i) a multi-perspective LLM voting framework to maximize recall on implicit ideation, or (ii) a feature-based ML ensemble guided by psychologically grounded indicators extracted via prompt-engineered LLMs for efficiency and interpretability. To the best of our knowledge, this is among the first works to operationalize LLM-extracted psychological features as structured vectors for suicide risk detection. On two complementary datasets - explicit-dominant Reddit and implicit-only DeepSuiMind - our framework outperforms single-model baselines, achieving 98.0% F1 on explicit cases, 99.7% on implicit ones, and reducing the cross-domain gap below 2%, while significantly lowering LLM cost.


OpenAI Adds Parental Safety Controls for Teen ChatGPT Users. Here's What to Expect

WIRED

OpenAI Adds Parental Safety Controls for Teen ChatGPT Users. OpenAI's review process for teenage ChatGPT users who are flagged for suicidal ideation includes human moderators. Parents can expect an alert about alarming prompts within hours. Starting today, OpenAI is rolling out ChatGPT safety tools intended for parents to use with their teenagers. This worldwide update includes the ability for parents, as well as law enforcement, to receive notifications if a child--in this case, users between the ages of 13 and 18--engages in chatbot conversations about self harm or suicide.


AI chatbots are becoming popular alternatives to therapy. But they may worsen mental health crises, experts warn

The Guardian

In 2023, a Belgian man reportedly ended his life after developing eco-anxiety and confiding in an AI chatbot over six weeks about the future of the planet. Without those conversations, his widow reportedly told the Belgian outlet La Libre, "he would still be here". In April this year, a 35-year-old Florida man was shot and killed by police in another chatbot-related incident: his father later told media that the man had come to believe an entity named Juliet was trapped inside ChatGPT, and then killed by OpenAI. When the man, who reportedly struggled with bipolar disorder and schizophrenia, was confronted by police, he allegedly charged at them with a knife. The wide availability of chatbots in the past few years has apparently led some to believe there is a ghost in the machine – one that is conscious, capable of loving and being loved.


A Gold Standard Dataset and Evaluation Framework for Depression Detection and Explanation in Social Media using LLMs

Bolegave, Prajval, Bhattacharya, Pushpak

arXiv.org Artificial Intelligence

Early detection of depression from online social media posts holds promise for providing timely mental health interventions. In this work, we present a high-quality, expert-annotated dataset of 1,017 social media posts labeled with depressive spans and mapped to 12 depression symptom categories. Unlike prior datasets that primarily offer coarse post-level labels \cite{cohan-etal-2018-smhd}, our dataset enables fine-grained evaluation of both model predictions and generated explanations. We develop an evaluation framework that leverages this clinically grounded dataset to assess the faithfulness and quality of natural language explanations generated by large language models (LLMs). Through carefully designed prompting strategies, including zero-shot and few-shot approaches with domain-adapted examples, we evaluate state-of-the-art proprietary LLMs including GPT-4.1, Gemini 2.5 Pro, and Claude 3.7 Sonnet. Our comprehensive empirical analysis reveals significant differences in how these models perform on clinical explanation tasks, with zero-shot and few-shot prompting. Our findings underscore the value of human expertise in guiding LLM behavior and offer a step toward safer, more transparent AI systems for psychological well-being.


Speech as a Multimodal Digital Phenotype for Multi-Task LLM-based Mental Health Prediction

Ali, Mai, Lucasius, Christopher, Patel, Tanmay P., Aitken, Madison, Vorstman, Jacob, Szatmari, Peter, Battaglia, Marco, Kundur, Deepa

arXiv.org Artificial Intelligence

Speech is a noninvasive digital phenotype that can offer valuable insights into mental health conditions, but it is often treated as a single modality. In contrast, we propose the treatment of patient speech data as a trimodal multimedia data source for depression detection. This study explores the potential of large language model-based architectures for speech-based depression prediction in a multimodal regime that integrates speech-derived text, acoustic landmarks, and vocal biomarkers. Adolescent depression presents a significant challenge and is often comorbid with multiple disorders, such as suicidal ideation and sleep disturbances. This presents an additional opportunity to integrate multi-task learning (MTL) into our study by simultaneously predicting depression, suicidal ideation, and sleep disturbances using the multimodal formulation. We also propose a longitudinal analysis strategy that models temporal changes across multiple clinical interactions, allowing for a comprehensive understanding of the conditions' progression. Our proposed approach, featuring trimodal, longitudinal MTL is evaluated on the Depression Early Warning dataset. It achieves a balanced accuracy of 70.8%, which is higher than each of the unimodal, single-task, and non-longitudinal methods.