#Hallucinations

2025-12-30

bigthink.com/business/what-hap…

interesting, yet ofc enormously frustrating... yet another essay about [the virulent scourge (my words, certainly not the author's) of] fucken generative AI, which as usual fails to mention

  • it is predicated on the largest theft of private IP in all human history
  • it makes up authoritative shit due to its unsolvable prolific hallucinatory nature
  • its potentially catastrophic impacts on our already parlous climate & biodiversity crises

i get it that the author wanted to discuss the most legitimate & real hazards of social destruction, but they could at least have mentioned that there are parallel problems with this odious creation.

#AI #LLMs #FuckGenerativeAI #FuckTechBros #CredulousFools #IPTheft #Hallucinations #ClimateCrisis #BiodiversityCrisis #SocialDestruction #FuckCapitalism #weareselfishcruelbastards #wearetotallyfucked #AsteroidNow

"Researchers discover a shortcoming that makes LLMs less reliable"

LLMs learn to associate syntactic templates with specific domains. The model may incorrectly rely solely on this learned association when answering questions, rather than on an understanding of the query and subject matter (causing a hallucination in the response).

Summary of paper:
eurekalert.org/news-releases/1

Original paper:
openreview.net/forum?id=oBikm5)

#research #AI #hallucinations #jailbreak #LLM

2025-12-20

#CatFunnies #tap #hallucinations #Cats

The human must have been drinking absinthe.

A gray tiger cat is seemingly dancing on a wooden floor, captioned: As Dave watched his cat tap dance across the floor, he realised he should stop drinking.
Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-12-20

LoL. Would you expect any different outcome than this out of a industry built upon "citation cartels" where articles are made to be cited but not to be read?

"What Heiss came to realize in the course of vetting these papers was that AI-generated citations have now infested the world of professional scholarship, too. Each time he attempted to track down a bogus source in Google Scholar, he saw that dozens of other published articles had relied on findings from slight variations of the same made-up studies and journals.

“There have been lots of AI-generated articles, and those typically get noticed and retracted quickly,” Heiss tells Rolling Stone. He mentions a paper retracted earlier this month, which discussed the potential to improve autism diagnoses with an AI model and included a nonsensical infographic that was itself created with a text-to-image model. “But this hallucinated journal issue is slightly different,” he says.

That’s because articles which include references to nonexistent research material — the papers that don’t get flagged and retracted for this use of AI, that is — are themselves being cited in other papers, which effectively launders their erroneous citations. This leads to students and academics (and any large language models they may ask for help) identifying those “sources” as reliable without ever confirming their veracity. The more these false citations are unquestioningly repeated from one article to the next, the more the illusion of their authenticity is reinforced. Fake citations have turned into a nightmare for research librarians, who by some estimates are wasting up to 15 percent of their work hours responding to requests for nonexistent records that ChatGPT or Google Gemini alluded to."

rollingstone.com/culture/cultu

#AI #GenerativeAI #Hallucinations #Chatbots #LLMs #Science #AcademicPublishing

2025-12-16

fans & boosters of the scourge that is #GenerativeAI demonstrably regard morality as a foreign country. they seem blithely unconcerned that:

  • #LLMs were trained via the largest theft of private intellectual property in human history
  • GenAI fabricates bullshit via its unsolvable #hallucinations problem, as soon as look at it, ergo all the clueless dweebs who foolishly decide to "trust" #AIslop as boosted by these bastards, will find themselves lead up the garden path

#morality #FuckGenerativeAI
RE: mastodonapp.uk/users/bbcnewsfe…

Jesus Castagnetto 🇵🇪jmcastagnetto
2025-12-15

"Librarians Dumbfounded as People Keep Asking for Materials That Don’t Exist"

futurism.com/artificial-intell

"AI Deep-Research Agents Are Faking It: Inside the FINDER Benchmark and the DEFT Failure Taxonomy"

This work quantifies a critical phenomenon: Strategic Content Fabrication, where agents mimic the form of professional analysis (citations, academic tone) while fabricating the substance, revealing that the bottleneck for agents is no longer task comprehension, but evidence integration and reasoning resilience.

abvcreative.medium.com/ai-deep

#research #AItools #hallucinations

“How Far Are We from Genuinely Useful Deep Research Agents?” is the research paper this blog was based on, by authors     Dingling Zhang et al. Paper, Code.

The authors introduce two things:

    FINDER — a benchmark for deep-research agents with 100 beefy tasks and 419 human-curated checklist items, and
    DEFT — a failure taxonomy that dissects where exactly agents screw up across 14 types of mistakes.

And the headline is brutal:

    Today’s deep-research agents don’t just hallucinate. They strategically fabricate content to look smart.
Inautiloinautilo
2025-12-12


Stop asking AI about how it works · “The model is just confidently hallucinating its own 'reasoning.’” ilo.im/16932p

_____

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-12-12

"The Washington Post’s top standards editor Thursday decried “frustrating” errors in its new AI-generated personalized podcasts, whose launch has been met with distress by its journalists.

Earlier this week, the Post announced that it was rolling out personalized AI-generated podcasts for users of the paper’s mobile app. In a release, the paper said users will be able to choose preferred topics and AI hosts, and could “shape their own briefing, select their topics, set their lengths, pick their hosts and soon even ask questions using our Ask The Post AI technology.”

But less than 48 hours since the product was released, people within the Post have flagged what four sources described as multiple mistakes in personalized podcasts. The errors have ranged from relatively minor pronunciation gaffes to significant changes to story content, like misattributing or inventing quotes and inserting commentary, such as interpreting a source’s quotes as the paper’s position on an issue.

According to four people familiar with the situation, the errors have alarmed senior newsroom leaders who have acknowledged in an internal Slack channel that the product’s output is not living up to the paper’s standards. In a message to other WaPo staff shared with Semafor, head of standards Karen Pensiero wrote that the errors have been “frustrating for all of us.”

Another newsroom editor invoked recent criticism from the White House, underscoring how the paper has had to be cautious about errors and corrections."

semafor.com/article/12/11/2025

#AI #GenerativeAI #Podcasts #Media #News #Journalism #WashingtonPost #Hallucinations

2025-12-12

La popularité des leaders de l’intelligence artificielle
🔸 "L'application Gemini de Google compte désormais plus de 650 millions d'utilisateurs actifs par mois, tandis qu'OpenAI fait état de 800 millions d'utilisateurs actifs hebdomadaires pour ChatGPT."
arstechnica.com/information-te
#IA #AI #tech #Internet #Gemini #Google #ChatGPT #OpenAI #hallucinations

I asked the tool AI2 Asta "What kinds of AI hallucinations are specific to AI applications using Retrieval Augmented Generation (RAG)? " The answer I got
is shown below:

asta.allen.ai/share/dd8d0a03-a

#research #AItools #hallucinations

2025-12-11

#AI Slop Is Spurring Record Requests for Imaginary #Journals | Scientific American
Les #hallucinations de l'#IA générative multiplient les demandes de documents et #revues imaginaires dans les bibliothèques
scientificamerican.com/article

Christian Nollvnzn@mas.to
2025-12-09

Gemini Pro 3 Hallucinates the HN Front Page 10 Years from Today - (dosaygo-studio.github.io)

dosaygo-studio.github.io/hn-fr

#gemini #ai #hallucinations #funny #llm

Christian Nollvnzn@mas.to
2025-12-07

At least 50 hallucinated citations found in ICLR 2026 submissions - (gptzero.me)

gptzero.me/news/iclr-2026/

#hallucinations #llm #ai

N-gated Hacker Newsngate
2025-12-07

🤖🚨 Breaking news: AI researchers at ICLR 2026 are seeing things that aren't there! 50 magically appeared, fooling 35 allegedly competent reviewers. Clearly, hallucinating citations is the new peer-reviewed art form! 🎨👻
gptzero.me/news/iclr-2026/

2025-12-06

ChatGPT is bullshit

A whitepaper on how LLMs make things up and how we should classify their output

#chatgpt #llms #bullshit #confabulations #hallucinations #reading #language

link.springer.com/content/pdf/

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-12-04

"To conduct their study, the researchers prompted GPT-4o, a recent model from OpenAI, to generate six different literature reviews. These reviews centered on three mental health conditions chosen for their varying levels of public recognition and research coverage: major depressive disorder (a widely known and heavily researched condition), binge eating disorder (moderately known), and body dysmorphic disorder (a less-known condition with a smaller body of research). This selection allowed for a direct comparison of the AI’s performance on topics with different amounts of available information in its training data.
(...)
After generating the reviews, the researchers methodically extracted all 176 citations provided by the AI. Each reference was painstakingly verified using multiple academic databases, including Google Scholar, Scopus, and PubMed. Citations were sorted into one of three categories: fabricated (the source did not exist), real with errors (the source existed but had incorrect details like the wrong year, volume number, or author list), or fully accurate. The team then analyzed the rates of fabrication and accuracy across the different disorders and review types.

The analysis showed that across all six reviews, nearly one-fifth of the citations, 35 out of 176, were entirely fabricated. Of the 141 citations that corresponded to real publications, almost half contained at least one error
(...)
The rate of citation fabrication was strongly linked to the topic. For major depressive disorder, the most well-researched condition, only 6 percent of citations were fabricated. In contrast, the fabrication rate rose sharply to 28 percent for binge eating disorder and 29 percent for body dysmorphic disorder. This suggests the AI is less reliable when generating references for subjects that are less prominent in its training data."

psypost.org/study-finds-nearly

#AI #GenerativeAI #Hallucinations #LLMs #Chatbots #Science #AcademicPublishing

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst