Goto

Collaborating Authors

 safeguard


Mind launches inquiry into AI and mental health after Guardian investigation

The Guardian

The Guardian revealed how people were being put at risk of harm by false and misleading health information in Google AI Overviews. The Guardian revealed how people were being put at risk of harm by false and misleading health information in Google AI Overviews. Exclusive: England and Wales charity to examine safeguards after Guardian exposed'very dangerous' advice on Google AI Overviews'Very dangerous': a Mind mental health expert on Google's AI summaries Mind is launching a significant inquiry into artificial intelligence and mental health after a Guardian investigation exposed how Google's AI Overviews gave people "very dangerous" medical advice. In a year-long commission, the mental health charity, which operates in England and Wales, will examine the risks and safeguards required as AI increasingly influences the lives of millions of people affected by mental health issues worldwide. The inquiry - the first of its kind globally - will bring together the world's leading doctors and mental health professionals, as well as people with lived experience, health providers, policymakers and tech companies.


ByteDance to curb AI video app after Disney legal threat

BBC News

Chinese technology giant ByteDance has pledged to curb a controversial artificial intelligence (AI) video-making tool, following threats of legal action from Disney and complaints from other entertainment giants. In the last few days, videos made using the latest version of the app Seedance have proliferated online. Many have been lauded for their realism. Disney's lawyers accused ByteDance of committing a virtual smash-and-grab of their intellectual property, including superheroes from Marvel, Star Wars and various cartoons. On Monday ByteDance told the BBC that the company respects intellectual property rights and we have heard the concerns regarding Seedance 2.0.


Use of AI to harm women has only just begun, experts warn

The Guardian

Elon Musk's AI tool, Grok, is being investigated by the UK's media regulator. Elon Musk's AI tool, Grok, is being investigated by the UK's media regulator. "Since discovering Grok AI, regular porn doesn't do it for me anymore, it just sounds absurd now," one enthusiast for the Elon Musk-owned AI chatbot wrote on Reddit. Another agreed: "If I want a really specific person, yes." If those who have been horrified by the distribution of sexualised imagery on Grok hoped that last week's belated safeguards could put the genie back in the bottle, there are many such posts on Reddit and elsewhere that tell a different story.


OpenAI tightens AI rules for teens but concerns remain

FOX News

This material may not be published, broadcast, rewritten, or redistributed. Quotes displayed in real-time or delayed by at least 15 minutes. Market data provided by Factset . Powered and implemented by FactSet Digital Solutions . Mutual Fund and ETF data provided by Refinitiv Lipper . Concerns remain over AI's impact on young people amid boom Tech expert praises New York's school cellphone ban as social media concerns rise Trump advisor details administration's push to boost AI hiring Kash Patel to close FBI's Hoover building in DC permanently Santa is'PACKING HEAT' during a traffic stop Trump has made AI a'key part' of his agenda, expert says Youth conservative movement has'never been in better hands,' podcast host says Most-used passwords of 2025 include '123456' and'password' Megan Garcia, a mother who lost her son to suicide after he allegedly became emotionally attached to an AI chatbot, discusses the dangers of the technology on'Fox News Sunday.'


Home Office admits facial recognition tech issue with black and Asian subjects

The Guardian

Facial recognition cameras being used near Arsenal's Emirates Stadium in north London before a match last month. Facial recognition cameras being used near Arsenal's Emirates Stadium in north London before a match last month. Calls for review after technology found to return more false positives for'some demographic groups' on certain settings Fri 5 Dec 2025 06.11 ESTLast modified on Fri 5 Dec 2025 06.57 Ministers are facing calls for stronger safeguards on the use of facial recognition technology after the Home Office admitted it is more likely to incorrectly identify black and Asian people than their white counterparts on some settings. Following the latest testing conducted by the National Physical Laboratory (NPL) of the technology's application within the police national database, the Home Office said it was "more likely to incorrectly include some demographic groups in its search results".


SCI: A Metacognitive Control for Signal Dynamics

Meesala, Vishal Joshua

arXiv.org Artificial Intelligence

Modern deep learning systems are typically deployed as open-loop function approximators: they map inputs to outputs in a single pass, without regulating how much computation or explanatory effort is spent on a given case. In safety-critical settings, this is brittle: easy and ambiguous inputs receive identical processing, and uncertainty is only read off retrospectively from raw probabilities. We introduce the Surgical Cognitive Interpreter (SCI), a lightweight closed-loop metacognitive control layer that wraps an existing stochastic model and turns prediction into an iterative process. SCI monitors a scalar interpretive state SP(t), here instantiated as a normalized entropy-based confidence signal, and adaptively decides whether to stop, continue sampling, or abstain. The goal is not to improve accuracy per se, but to regulate interpretive error ΔSP and expose a safety signal that tracks when the underlying model is likely to fail. We instantiate SCI around Monte Carlo dropout classifiers in three domains: vision (MNIST digits), medical time series (MIT-BIH arrhythmia), and industrial condition monitoring (rolling-element bearings). In all cases, the controller allocates more inference steps to misclassified inputs than to correct ones (up to about 3-4x on MNIST and bearings, and 1.4x on MIT-BIH). The resulting ΔSP acts as a usable safety signal for detecting misclassifications (AUROC 0.63 on MNIST, 0.70 on MIT-BIH, 0.86 on bearings). Code and reproducibility: https://github.com/vishal-1344/sci


Australia clamps downs on 'nudify' sites used for AI-generated child abuse

Al Jazeera

Australia clamps downs on'nudify' sites used for AI-generated child abuse Internet users in Australia have been blocked from accessing several websites that used artificial intelligence to create child sexual exploitation material, the country's internet regulator has announced. The three "nudify" sites withdrew from Australia following an official warning, eSafety Commissioner Julie Inman Grant said on Thursday. Grant said such "nudify" services, which allow users to make images of real people appear naked using AI, have had a "devastating" effect in Australian schools. "We took enforcement action in September because this provider failed to put in safeguards to prevent its services being used to create child sexual exploitation material and were even marketing features like undressing'any girl,' and with options for'schoolgirl' image generation and features such as'sex mode,'" Grand said in a statement. The development comes after Grant's office issued a formal warning to the United Kingdom-based company behind the sites in September, threatening civil penalties of up to 49.5 million Australian dollars ($32.2m) if it did not introduce safeguards to prevent image-based abuse.