Goto

Collaborating Authors

 safeguard


Tennessee minors sue Musk's xAI, alleging Grok generated sexual images of them

The Japan Times

Tennessee minors sue Musk's xAI, alleging Grok generated sexual images of them Governments and regulators around the world have launched probes into xAI, imposed bans and demanded safeguards in a growing push to curb illegal and offensive material. Three Tennessee plaintiffs, including two minors, sued Elon Musk's xAI on Monday, alleging that it knowingly designed its Grok image generator to let people create sexually explicit content by using real photos of others. The lawsuit, filed in the San Jose, California federal court, is seeking class-action status for people in the United States who were reasonably identifiable in sexualized images or videos generated by Grok based on real images of themselves. The artificial intelligence company did not immediately respond to a request for comment. After an outcry over sexually explicit content generated by the chatbot, xAI said in January that it had blocked all users from editing images of real people in revealing clothing and from generating images of people in revealing clothing in jurisdictions where it's illegal. Governments and regulators around the world have also since launched probes, imposed bans and demanded safeguards in a growing push to curb illegal and offensive material.


Mind launches inquiry into AI and mental health after Guardian investigation

The Guardian

The Guardian revealed how people were being put at risk of harm by false and misleading health information in Google AI Overviews. The Guardian revealed how people were being put at risk of harm by false and misleading health information in Google AI Overviews. Exclusive: England and Wales charity to examine safeguards after Guardian exposed'very dangerous' advice on Google AI Overviews'Very dangerous': a Mind mental health expert on Google's AI summaries Mind is launching a significant inquiry into artificial intelligence and mental health after a Guardian investigation exposed how Google's AI Overviews gave people "very dangerous" medical advice. In a year-long commission, the mental health charity, which operates in England and Wales, will examine the risks and safeguards required as AI increasingly influences the lives of millions of people affected by mental health issues worldwide. The inquiry - the first of its kind globally - will bring together the world's leading doctors and mental health professionals, as well as people with lived experience, health providers, policymakers and tech companies.



ByteDance to curb AI video app after Disney legal threat

BBC News

Chinese technology giant ByteDance has pledged to curb a controversial artificial intelligence (AI) video-making tool, following threats of legal action from Disney and complaints from other entertainment giants. In the last few days, videos made using the latest version of the app Seedance have proliferated online. Many have been lauded for their realism. Disney's lawyers accused ByteDance of committing a virtual smash-and-grab of their intellectual property, including superheroes from Marvel, Star Wars and various cartoons. On Monday ByteDance told the BBC that the company respects intellectual property rights and we have heard the concerns regarding Seedance 2.0.




Use of AI to harm women has only just begun, experts warn

The Guardian

Elon Musk's AI tool, Grok, is being investigated by the UK's media regulator. Elon Musk's AI tool, Grok, is being investigated by the UK's media regulator. "Since discovering Grok AI, regular porn doesn't do it for me anymore, it just sounds absurd now," one enthusiast for the Elon Musk-owned AI chatbot wrote on Reddit. Another agreed: "If I want a really specific person, yes." If those who have been horrified by the distribution of sexualised imagery on Grok hoped that last week's belated safeguards could put the genie back in the bottle, there are many such posts on Reddit and elsewhere that tell a different story.


OpenAI tightens AI rules for teens but concerns remain

FOX News

This material may not be published, broadcast, rewritten, or redistributed. Quotes displayed in real-time or delayed by at least 15 minutes. Market data provided by Factset . Powered and implemented by FactSet Digital Solutions . Mutual Fund and ETF data provided by Refinitiv Lipper . Concerns remain over AI's impact on young people amid boom Tech expert praises New York's school cellphone ban as social media concerns rise Trump advisor details administration's push to boost AI hiring Kash Patel to close FBI's Hoover building in DC permanently Santa is'PACKING HEAT' during a traffic stop Trump has made AI a'key part' of his agenda, expert says Youth conservative movement has'never been in better hands,' podcast host says Most-used passwords of 2025 include '123456' and'password' Megan Garcia, a mother who lost her son to suicide after he allegedly became emotionally attached to an AI chatbot, discusses the dangers of the technology on'Fox News Sunday.'


Home Office admits facial recognition tech issue with black and Asian subjects

The Guardian

Facial recognition cameras being used near Arsenal's Emirates Stadium in north London before a match last month. Facial recognition cameras being used near Arsenal's Emirates Stadium in north London before a match last month. Calls for review after technology found to return more false positives for'some demographic groups' on certain settings Fri 5 Dec 2025 06.11 ESTLast modified on Fri 5 Dec 2025 06.57 Ministers are facing calls for stronger safeguards on the use of facial recognition technology after the Home Office admitted it is more likely to incorrectly identify black and Asian people than their white counterparts on some settings. Following the latest testing conducted by the National Physical Laboratory (NPL) of the technology's application within the police national database, the Home Office said it was "more likely to incorrectly include some demographic groups in its search results".


SCI: A Metacognitive Control for Signal Dynamics

Meesala, Vishal Joshua

arXiv.org Artificial Intelligence

Modern deep learning systems are typically deployed as open-loop function approximators: they map inputs to outputs in a single pass, without regulating how much computation or explanatory effort is spent on a given case. In safety-critical settings, this is brittle: easy and ambiguous inputs receive identical processing, and uncertainty is only read off retrospectively from raw probabilities. We introduce the Surgical Cognitive Interpreter (SCI), a lightweight closed-loop metacognitive control layer that wraps an existing stochastic model and turns prediction into an iterative process. SCI monitors a scalar interpretive state SP(t), here instantiated as a normalized entropy-based confidence signal, and adaptively decides whether to stop, continue sampling, or abstain. The goal is not to improve accuracy per se, but to regulate interpretive error ΔSP and expose a safety signal that tracks when the underlying model is likely to fail. We instantiate SCI around Monte Carlo dropout classifiers in three domains: vision (MNIST digits), medical time series (MIT-BIH arrhythmia), and industrial condition monitoring (rolling-element bearings). In all cases, the controller allocates more inference steps to misclassified inputs than to correct ones (up to about 3-4x on MNIST and bearings, and 1.4x on MIT-BIH). The resulting ΔSP acts as a usable safety signal for detecting misclassifications (AUROC 0.63 on MNIST, 0.70 on MIT-BIH, 0.86 on bearings). Code and reproducibility: https://github.com/vishal-1344/sci