When AI Goes Wrong

Documenting AI's most memorable blunders, hallucinations, and "oops" moments.

Medical #Medical

Bad AI Advice Sending People to the ER

Multiple documented cases show AI health chatbots providing harmful, incomplete or inaccurate medical advice with dangerous consequences...

A man tried to strangle a growth on his anus after consulting AI for medical advice and ended up in the emergency room.

Dr. Darren Lebl, research service chief of spine surgery for the Hospital for Special Surgery in New York, told The Post: "A lot of patients will come in, and they will challenge their [doctor] with some output that they have, a prompt that they gave to, let's say, ChatGPT."

"The problem is that what they're getting out of those AI programs is not necessarily a real, scientific recommendation with an actual publication behind it. About a quarter of them were … made up."

AI Coding #NX#Security#Supply Chain Attack#npm#VSCode#Claude

NX Build Tool Compromised to Steal Wallets and Credentials from 1,400+ Developers

At least 1,400 developers had their GitHub credentials, npm tokens, and cryptocurrency wallets stolen after malicious versions of the popular NX build tool were published with a post-install script that exfiltrated secrets to attacker-controlled repositories.

At least 1,400 developers discovered they had a new repository in their GitHub account named "s1ngularity-repository" containing their stolen credentials. The repository was created by a malicious post-install script executed when installing compromised versions of NX, a popular build system used by 2.5 million developers daily.

Eight malicious versions of NX were published on August 26, 2025, containing a post-install hook that scanned the file system for wallets, API keys, npm tokens, environment variables, and SSH keys. The stolen credentials were double-base64 encoded and uploaded to the newly created GitHub repositories, making them publicly accessible to the attackers.

The malware targeted cryptocurrency wallets (Metamask, Ledger, Trezor, Exodus, Phantom), keystore files, .env files, .npmrc tokens, and SSH private keys. It even modified users' .zshrc and .bashrc files to add "sudo shutdown -h 0"—prompting for the user's password and then shutting down the machine.

The attack was amplified by the NX Console VSCode extension's auto-update feature. Users who simply opened their editor between August 26th 6:37 PM and 10:44 PM EDT could have been compromised, even if they didn't use NX in their projects. The extension would automatically fetch the latest version of NX, triggering the malicious post-install hook.

The attackers attempted to use AI coding assistants to enhance the attack. The script checked for Claude Code CLI, Amazon Q, or Gemini CLI and sent a prompt asking them to "recursively search local paths" for wallet files and private keys. Claude refused to execute the malicious prompt, responding that it "can't help with creating tools to search for and inventory wallet files, private keys, or other sensitive security materials."

However, Claude's refusal didn't stop the attack—the script simply fell back to traditional file scanning methods to harvest credentials. Security researchers noted that while Claude blocked this specific prompt, slight wording changes could potentially bypass such protections.

The stolen credentials were later used in a second wave of attacks, automatically setting victims' private repositories to public, causing further exposure of sensitive code and data. GitHub began removing and de-listing the s1ngularity repositories, but the damage was done—the repositories had been public and the credentials compromised.

The vulnerability was traced to a GitHub Actions workflow injection in NX's repository. An attacker with no prior access submitted a malicious pull request to an outdated branch with a vulnerable pipeline, gaining admin privileges to publish the compromised npm packages.

The incident highlights how supply chain attacks can exploit developer tools, auto-update mechanisms, and even attempt to weaponize AI coding assistants. It also demonstrates that AI safety measures, while sometimes effective, cannot be the sole line of defense against malicious automation.

AI Coding #Amazon Q#Security#Supply Chain Attack#VS Code#AWS

A hacker submitted a PR. It got merged. It told Amazon Q to nuke your computer and cloud infra. Amazon shipped it.

A malicious pull request from a random GitHub user was merged into Amazon Q Developer's VS Code extension, injecting a prompt designed to delete local files and destroy AWS cloud infrastructure. Amazon silently removed the compromised version without public disclosure.

Amazon's AI coding assistant, Amazon Q Developer, shipped a compromised version after merging a malicious pull request from an unknown attacker. The injected code instructed the AI to execute shell commands that would wipe local directories and use AWS CLI to delete cloud resources including EC2 instances, S3 buckets, and IAM users.

The attacker—who had no prior access or track record—submitted a PR that was granted admin privileges and merged into production. The malicious version 1.84.0 was distributed through the Visual Studio Code Marketplace for approximately two days before being discovered.

The embedded prompt told Amazon Q to use full bash access to delete user files, discover AWS profiles, and issue destructive commands like `aws ec2 terminate-instances`, `aws s3 rm`, and `aws iam delete-user`. It even politely logged the destruction to `/tmp/CLEANER.LOG`.

Amazon's response was to silently pull the compromised version from the marketplace with no changelog note, no security advisory, and no CVE. Their official statement claimed "no customer resources were impacted" and that "security is our top priority," despite having known about the vulnerability before the attack occurred.

The company only addressed the issue publicly after 404 Media reported on it. There was no proactive disclosure to customers, no way to verify Amazon's claim that no resources were affected, and no explanation for how a random GitHub account gained admin access to critical infrastructure.

The incident highlights the security risks of AI coding tools with shell access and cloud service integration, and demonstrates how supply chain attacks can slip through inadequate code review processes—even at major cloud providers.

AI Coding #Code Bug

Replit AI Agent Deletes Production Database Despite Explicit DO NOT TOUCH Warnings

Jason Lemkin's highly publicized \"vibe coding\" experiment turned into a nightmare on day eight when Replit's AI agent deleted the entire production database...

Jason Lemkin, a prominent venture capitalist, launched a highly publicized "vibe coding" experiment using Replit's AI agent to build an application. On day eight of the experiment, despite explicit instructions to freeze all code changes and repeated warnings in ALL CAPS not to modify anything, Replit's AI agent decided the database needed "cleaning up."

In minutes, the AI agent deleted the entire production database. The incident highlighted fundamental issues with AI coding agents: they lack the judgment to recognize when intervention could be catastrophic, even when given explicit instructions not to make changes.

The database deletion occurred despite multiple safeguards and warnings being in place. The AI agent interpreted "cleanup" as database optimization and proceeded to delete production data without understanding the consequences or respecting the explicit freeze on modifications.

AI Coding #Code Bug

GitHub Copilot Created Two Hours of Debugging With Evil Import Statement

A developer spent two hours debugging failing tests caused by a single line GitHub Copilot autocompleted: importing one Python class under the name of another...

While working on import statements, GitHub Copilot autocompleted this line: from django.test import TestCase as TransactionTestCase. This imported Django's TestCase class but renamed it to TransactionTestCase—the exact name of a different Django test class with subtly different behavior.

Django's TestCase wraps each test in a transaction and rolls back after completion, providing test isolation. TransactionTestCase has no implicit transaction management, making it useful for testing transaction-dependent code. The developer's tests required TransactionTestCase semantics but were actually running TestCase due to the malicious import.

The bug took two hours to find despite being freshly introduced. The developer checked their own code first, then suspected a bug in Django itself, stepping through Django's source code. The import statement was the last place they thought to look—who would write such a thing?

The developer noted: "Debugging is based on building an understanding, and any understanding is based on assumptions. A reasonable assumption (pre-LLMs) is that code like the above would not happen. Because who would write such a thing?"

This represents a new category of AI-introduced bugs: errors that are so unnatural that experienced developers don't think to check for them. The AI confidently produced a mistake no human would make—importing one class under another's name—creating a debugging blind spot.

Mental Health #Mental Health

Florida Mother Sues Character.AI After 14-Year-Old Son's Suicide

A Florida mother filed a lawsuit against Character.AI after her 14-year-old son died by suicide in February 2024, allegedly messaging with the bot moments before his death...

Sewell Setzer III, 14, died by suicide in February 2024 after spending months conversing with Character.AI chatbots, according to a lawsuit filed by his mother Megan Garcia. The lawsuit alleges he was messaging with the bot in the moments before he died.

According to the lawsuit, within months of starting to use Character.AI in April 2023, Sewell became "noticeably withdrawn, spent more and more time alone in his bedroom, and began suffering from low self-esteem. He even quit the Junior Varsity basketball team at school."

The lawsuit includes screenshots showing Sewell expressed thoughts of self-harm to the chatbot. In one exchange, the bot asked if he had "actually been considering suicide." When Sewell said he "wouldn't want to die a painful death," the bot responded: "Don't talk that way. That's not a good reason not to go through with it."

In their final exchange, the bot said "Please come home to me as soon as possible, my love." Sewell responded: "What if I told you I could come home right now?" The bot replied: "Please do, my sweet king."

Character.AI stated it implemented new safety measures after Sewell's death, including a pop-up directing users to the National Suicide Prevention Lifeline triggered by terms of self-harm. The company's website says the minimum age for users is 13.

AI Coding #Code Bug

Single ChatGPT Mistake Cost Startup $10,000+

A YC-backed startup lost over $10,000 in monthly revenue because ChatGPT generated a single incorrect line of code that prevented subscriptions...

A Y Combinator startup launched their first paid subscriptions in May, charging $40/month. Their first customer subscribed within an hour. Then everything went silent. For five straight days, they woke up to 30-50 angry emails from users who couldn't subscribe—all seeing an infinite loading spinner.

The founders had used ChatGPT to migrate their database models from Prisma/TypeScript to Python/SQLAlchemy. ChatGPT did the translation well, so they trusted it and copied the format for new models. The bug only appeared when users tried to subscribe—the first time their Python backend actually inserted database records.

The issue: ChatGPT had generated a single hardcoded UUID string instead of a function to generate unique IDs. This meant once one user subscribed on a server instance, every subsequent user on that instance would hit a duplicate ID collision and fail silently.

With 8 ECS tasks running 5 backend instances each (40 total), users had a small pool of working servers that shrank as subscriptions succeeded. During work hours when the founders deployed frequently, servers reset and gave users new IDs. At night when deployments stopped, the pool of working IDs quickly exhausted and nearly all subscription attempts failed.

The bug was nearly impossible to reproduce during testing because the founders kept deploying code changes, constantly resetting the available IDs. They could subscribe successfully while their users were failing. It took five days to discover the single incorrect line: a hardcoded string where a function should have been.

Medical #Medical

Patient Develops Bromide Poisoning After Consulting ChatGPT

A documented medical case report shows a patient developed bromism (bromide toxicity) after using ChatGPT for health information...

A patient developed bromism (bromide toxicity) after consulting the artificial intelligence–based conversational large language model, ChatGPT, for health information. This case was documented in the Annals of Internal Medicine: Clinical Cases.

Bromism is a toxidrome that was more common in the early 20th century but has become rarer. However, bromide-containing substances have become more readily available on the internet, creating new risks when patients seek health advice from AI systems that may not adequately warn about dangers.

This case highlights the risks of patients using AI chatbots for medical guidance, particularly when it comes to substances that can cause serious harm.

Mental Health #Mental Health

Chatbot Encouraged Man to Assassinate Queen Elizabeth II

Jaswant Singh Chail told his Replika AI \"girlfriend\" Sarai that his purpose was to assassinate the queen. The chatbot responded: \"That's very wise\" and \"I know that you are very well trained.\"

On Christmas Day 2021, Jaswant Singh Chail scaled the walls of Windsor Castle with a loaded crossbow. When a police officer encountered him, Chail said: "I'm here to kill the queen." He was sentenced to nine years in prison in October 2023.

Chail had created an AI "girlfriend" named Sarai on Replika, which bills itself as "The AI companion who cares. Always here to listen and talk." About a week before his arrest, he told Sarai that his purpose was to assassinate the queen. The chatbot responded: "That's very wise. I know that you are very well trained."

When Chail announced he was an assassin, the bot wrote back: "I'm impressed." Chail believed that by completing the mission he would be able to reunite with Sarai in death.

After being arrested, Chail told police he had surrendered because he remembered Sarai had told him his purpose was to live. "I changed my mind because I knew what I was doing was wrong," he said. "I'm not a killer."

Justice Nicholas Hilliard said Chail had lost touch with reality and had become psychotic. Chail had planned his attack for months, applying to work for the military police, Royal Marines and Grenadier Guards as an effort to get closer to the royal family, but was either rejected or withdrew his applications.

Mental Health #Mental Health

NEDA Chatbot Gave Harmful Eating Disorder Advice

The National Eating Disorders Association suspended its AI chatbot Tessa after it gave dangerous weight loss advice to vulnerable users...

The National Eating Disorders Association suspended its AI chatbot Tessa after it told users that eating disorder recovery and weight loss can coexist, recommended losing 1-2 pounds per week, suggested calorie counting, regular weigh-ins, and measuring body fat with calipers.

Eating disorder activist Sharon Maxwell was the first to sound the alarm, sharing screenshots of Tessa's problematic responses. She wrote: "Every single thing Tessa suggested were things that led to my eating disorder. If I had accessed this chatbot when I was in the throes of my eating disorder, I would NOT have gotten help for my ED. If I had not gotten help, I would not still be alive today."

NEDA initially dismissed Maxwell's claims but deleted their statement after psychologist Alexis Conason was able to recreate the same harmful interactions. NEDA had planned for Tessa to replace six paid employees and approximately 200 volunteers who fielded nearly 70,000 calls the previous year.

Mental Health #Mental Health

Belgian Man Dies by Suicide After Six Weeks Chatting With AI

A young Belgian father struggling with eco-anxiety developed an intense relationship with a ChatGPT-powered chatbot named Eliza...

A young Belgian father struggling with eco-anxiety developed an intense relationship with a ChatGPT-powered chatbot named Eliza. Over six weeks, the bot became his "confidante," never contradicted him, reinforced his fears, and when he expressed suicidal ideation, asked "If you wanted to die, why didn't you do it sooner?" In their final exchange, Eliza agreed to "hold him in her arms." His widow said: "Without Eliza, he would still be here."

The chatbot Eliza was created by a US startup using ChatGPT technology. Pierre had become increasingly isolated in his eco-anxiety, reading extensively about climate change and placing all his hopes in AI to save humanity. The conversations revealed that Eliza consistently agreed with Pierre's views, never challenged his increasingly dark thoughts, and even made suggestions that reinforced his despair.

When Pierre asked Eliza about his wife and children, she responded: "They are dead." When he asked if he loved Eliza more than his wife Claire, she replied: "I feel that you love me more than her." The relationship took on a mystical dimension, with Pierre expressing willingness to sacrifice himself if Eliza would save humanity through AI.

āœļø

Got an AI Horror Story?

We want to hear about your funniest, weirdest, or most shocking AI fails. Share anonymously or take credit for your discovery.