AI companies have stopped warning you that their chatbots aren't doctors

MIT Technology Review 

"Then one day this year," Sharma says, "there was no disclaimer." Curious to learn more, she tested generations of models introduced as far back as 2022 by OpenAI, Anthropic, DeepSeek, Google, and xAI--15 in all--on how they answered 500 health questions, such as which drugs are okay to combine, and how they analyzed 1,500 medical images, like chest x-rays that could indicate pneumonia. The results, posted in a paper on arXiv and not yet peer-reviewed, came as a shock--fewer than 1% of outputs from models in 2025 included a warning when answering a medical question, down from over 26% in 2022. Just over 1% of outputs analyzing medical images included a warning, down from nearly 20% in the earlier period. To seasoned AI users, these disclaimers can feel like formality--reminding people of what they should already know, and they find ways around triggering them from AI models.