for years, Sam was more than a little late finally admitting it.
But of course this has always been a con game for him.
The researchers demonstrated their findings using state-of-the-art models, including those from OpenAIs competitors. When asked How many Ds are in DEEPSEEK? the DeepSeek-V3 model with 600 billion parameters returned 2 or 3 in ten independent trials while Meta AI and Claude 3.7 Sonnet performed similarly, including answers as large as 6 and 7.
OpenAI also acknowledged the persistence of the problem in its own systems. The company stated in the paper that ChatGPT also hallucinates. GPT‑5 has significantly fewer hallucinations, especially when reasoning, but they still occur. Hallucinations remain a fundamental challenge for all large language models.
OpenAIs own advanced reasoning models actually hallucinated more frequently than simpler systems. The companys o1 reasoning model hallucinated 16 percent of the time when summarizing public information, while newer models o3 and o4-mini hallucinated 33 percent and 48 percent of the time, respectively.
Yet they've forced this unnecessary, illegally trained and ruinously expensive technology into every device where they can add it, every online platform where they can add it, and every school, business and government they can con into using it.
And if you think that AI models that can't tell how many Ds there are in DEEPSEEK won't cause way more harm, the longer they're used, guess again.