{"id":5216,"date":"2025-09-15T15:33:13","date_gmt":"2025-09-15T15:33:13","guid":{"rendered":"https:\/\/musictechohio.online\/site\/fixing-hallucinations-destroy-chatgpt\/"},"modified":"2025-09-15T15:33:13","modified_gmt":"2025-09-15T15:33:13","slug":"fixing-hallucinations-destroy-chatgpt","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/fixing-hallucinations-destroy-chatgpt\/","title":{"rendered":"Fixing Hallucinations Would Destroy ChatGPT, Expert Finds"},"content":{"rendered":"<div>\n<div><img loading=\"lazy\" width=\"1200\" height=\"630\" src=\"https:\/\/wordpress-assets.futurism.com\/2025\/09\/fixing-hallucinations-destroy-chatgpt.jpg\" class=\"attachment-full size-full wp-post-image\" alt=\"University of Sheffield lecturer Wei Xing argued that the AI industry wouldn't be economically incentivized to fix hallucinations.\" style=\"margin-bottom: 15px;\" decoding=\"async\"><\/div>\n<p>In a <a href=\"https:\/\/arxiv.org\/abs\/2509.04664\">paper<\/a> published earlier this month, OpenAI researchers <a href=\"https:\/\/futurism.com\/openai-mistake-hallucinations\">said they&#8217;d found the reason<\/a> why even the most powerful AI models still suffer from rampant &#8220;hallucinations,&#8221;\u00a0in which products like ChatGPT confidently make assertions that are factually false.<\/p>\n<p>They found that the way we evaluate the output of large language models, like the ones driving ChatGPT, means they&#8217;re &#8220;optimized to be good test-takers&#8221; and that &#8220;guessing when uncertain improves test performance.&#8221;<\/p>\n<p>In simple terms, the creators of AI incentivize them to guess rather than admit they don&#8217;t know the answer \u2014 which might be a good strategy on an exam, but is outright dangerous when giving high-stakes advice about topics like <a href=\"https:\/\/futurism.com\/neoscope\/google-healthcare-ai-makes-up-body-part\">medicine<\/a> or <a href=\"https:\/\/futurism.com\/anthropic-chatbot-legal-music\">law<\/a>.<\/p>\n<p>While OpenAI claimed in an <a href=\"https:\/\/openai.com\/index\/why-language-models-hallucinate\/\">accompanying blog post<\/a> that &#8220;there is a straightforward fix&#8221; \u2014 tweaking evaluations to &#8220;penalize confident errors more than you penalize uncertainty and give partial credit for appropriate expressions of uncertainty&#8221; \u2014 one expert is warning that the strategy could pose devastating business realities.<\/p>\n<p>In an <a href=\"https:\/\/theconversation.com\/why-openais-solution-to-ai-hallucinations-would-kill-chatgpt-tomorrow-265107\">essay for <em>The Conversation<\/em><\/a>, University of Sheffield lecturer and AI optimization expert Wei Xing argued that the AI industry wouldn&#8217;t be economically incentivized to make these changes, as doing so could dramatically increase costs.<\/p>\n<p>Worse yet, having an AI repeatedly admit it can&#8217;t answer a prompt with a sufficient degree of confidence could deter users, who love a confidently positioned answer, even if it&#8217;s ultimately incorrect.<\/p>\n<p>Even if ChatGPT admitted that it doesn&#8217;t know the answer just 30 percent of the time, users could quickly become frustrated and move on, Xing argued.<\/p>\n<p>&#8220;Users accustomed to receiving confident answers to virtually any question would likely abandon such systems rapidly,&#8221; the researcher wrote.<\/p>\n<p>While there are &#8220;established methods for quantifying uncertainty,&#8221; AI models could end up requiring &#8220;significantly more computation than today\u2019s approach,&#8221; he argued, &#8220;as they must evaluate multiple possible responses and estimate confidence levels.&#8221;<\/p>\n<p>&#8220;For a system processing millions of queries daily, this translates to dramatically higher operational costs,&#8221; Xing wrote.<\/p>\n<p>Piling up the expenses at this juncture could prove disastrous. AI companies have bet big on scale, doubling down on expanding infrastructure to run increasingly power-hungry models. But try as they might, a return on investment appears to be <a href=\"https:\/\/www.axios.com\/2025\/08\/21\/ai-wall-street-big-tech\">many years<\/a>, if not decades, out. So far, tens of billions of dollars worth of capital expenditures have <a href=\"https:\/\/www.wheresyoured.at\/why-everybody-is-losing-money-on-ai\/\">eclipsed<\/a> relatively modest revenues.<\/p>\n<p>In other words, increasing already sky-high operational costs \u2014 while alienating users \u2014 could be yet another major thorn in the side of firms like OpenAI as they race to reassure investors that there&#8217;s a feasible business model in the long term.<\/p>\n<p>Xing argued that the company&#8217;s proposed fixes for hallucinations may work for &#8220;AI systems managing critical business operations or economic infrastructure&#8221; as &#8220;the cost of hallucinations far exceeds the expense of getting models to decide whether they\u2019re too uncertain.&#8221;<\/p>\n<p>&#8220;However, consumer applications still dominate AI development priorities,&#8221; he added. &#8220;Users want systems that provide confident answers to any question.&#8221;<\/p>\n<p>Arriving at a more uncertain answer faster is inherently cheaper for companies, which could disincentivize\u00a0a more careful and confident approach involving fewer hallucinations.<\/p>\n<p>How all of this will play out in the long term is anybody&#8217;s guess, especially as market forces continue to shift and companies find more efficient ways to run their AI models.<\/p>\n<p>But one thing is unlikely to change: guessing will always remain a far more economical and affordable option.<\/p>\n<p>&#8220;In short, the OpenAI paper inadvertently highlights an uncomfortable truth,&#8221; Xing concluded. &#8220;The business incentives driving consumer AI development remain fundamentally misaligned with reducing hallucinations.&#8221;<\/p>\n<p>&#8220;Until these incentives change, hallucinations will persist,&#8221; he added.<\/p>\n<p><strong>More on hallucinations:<\/strong> <em><a href=\"https:\/\/futurism.com\/openai-mistake-hallucinations\">OpenAI Realizes It Made a Terrible Mistake<\/a><\/em><\/p>\n<p>The post <a href=\"https:\/\/futurism.com\/fixing-hallucinations-destroy-chatgpt\">Fixing Hallucinations Would Destroy ChatGPT, Expert Finds<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>In a paper published earlier this month, OpenAI researchers said they&#8217;d found the reason why even the most powerful AI models still suffer from rampant &#8220;hallucinations,&#8221;\u00a0in which products like ChatGPT&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[177,183,3733,179],"tags":[],"class_list":["post-5216","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-generative-ai","category-hallucinations","category-openai"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/5216","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=5216"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/5216\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=5216"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=5216"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=5216"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}