π¨ AI is hallucinating more, just as weβre trusting it with more critical work. New βreasoningβ models, such as OpenAIβs o3 and o4-mini, were designed to solve complex problems step-by-step. But the results?
π§ o3: 51% hallucination rate on general questions
π o4-mini: 79% hallucination on benchmark tests
π Google & DeepSeekβs models also show rising errors
β οΈ Trial-and-error learning compounds risk at each step
Why is this happening? Because these models donβt understand truth, they just predict what sounds right. And the more they βthink,β the more they misstep.
Weβre using these tools in legal, medical, and enterprise settingsβyet even their creators admit:
π§© We donβt know exactly how they work.
β
Itβs a wake-up call: accuracy, explainability, and source traceability must be the new AI benchmarks.
#AI #LLM #ResponsibleAI #AIEthics #Hallucination
https://www.nytimes.com/2025/05/05/technology/ai-hallucinations-chatgpt-google.html