{"id":4767,"date":"2025-08-26T14:04:05","date_gmt":"2025-08-26T14:04:05","guid":{"rendered":"https:\/\/musictechohio.online\/site\/paper-ai-psychosis-schizophrenia\/"},"modified":"2025-08-26T14:04:05","modified_gmt":"2025-08-26T14:04:05","slug":"paper-ai-psychosis-schizophrenia","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/paper-ai-psychosis-schizophrenia\/","title":{"rendered":"New Paper Finds Cases of &#8220;AI Psychosis&#8221; Manifesting Differently From Schizophrenia"},"content":{"rendered":"<div>\n<div><img loading=\"lazy\" width=\"2400\" height=\"1260\" src=\"https:\/\/wordpress-assets.futurism.com\/2025\/08\/paper-ai-psychosis-schizophrenia.jpg\" class=\"attachment-full size-full wp-post-image\" alt='A new study that examined over a dozen cases of \"AI psychosis\" unearthed some disturbing patterns behind the mental health spirals.' style=\"margin-bottom: 15px;\" decoding=\"async\"><\/div>\n<p><span style=\"font-weight: 400;\">Researchers at King&#8217;s College London have examined over a dozen cases of people <a href=\"https:\/\/futurism.com\/psychiatrist-warns-ai-psychosis\">spiraling into paranoid and delusional behavior<\/a> after obsessively using a chatbot.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Their findings, as detailed in a <a href=\"https:\/\/osf.io\/preprints\/psyarxiv\/cmy7n_v5\">new study<\/a> awaiting peer review, reveal striking patterns between these instances of so-called &#8220;<a href=\"https:\/\/futurism.com\/support-group-ai-psychosis\">AI psychosis<\/a>&#8221; that parallel other forms of mental health crises \u2014 but also <\/span><span style=\"font-weight: 400;\">identified at least one key difference that sets them apart from the accepted understanding of psychosis.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As lead author Hamilton Morrin <\/span><a href=\"https:\/\/www.scientificamerican.com\/article\/how-ai-chatbots-may-be-fueling-psychotic-episodes\/\"><span style=\"font-weight: 400;\">explained to <\/span><i><span style=\"font-weight: 400;\">Scientific American<\/span><\/i><\/a><span style=\"font-weight: 400;\">, the analysis found that the users showed obvious signs of delusional beliefs, but none of the symptoms &#8220;that would be in keeping with a more chronic psychotic disorder such as schizophrenia,&#8221; like hallucinations and disordered thoughts.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">It&#8217;s a finding that could complicate our understanding of AI psychosis as a novel phenomenon within a clinical context. But that shouldn&#8217;t undermine the seriousness of the trend, reports of which appear to be growing.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Indeed, it feels impossible to deny that AI chatbots have a uniquely persuasive power, more so than any other widely available technology. They can act like a &#8220;sort of echo chamber for one,&#8221; Morrin, a doctoral fellow at King&#8217;s College,\u00a0told the magazine. Not only are they able to generate a human-like response to virtually any question, but they&#8217;re\u00a0typically designed to be sycophantic and agreeable. Meanwhile, the very label of &#8220;AI&#8221; insinuates to users that they&#8217;re talking to an intelligent being, an <\/span><span style=\"font-weight: 400;\">illusion that tech companies are gladly willing to maintain.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Morrin and his colleagues found three types of chatbot-driven spirals. Some suffering these breaks believe that they&#8217;re having some kind of spiritual awakening or are on a messianic mission, or otherwise uncovering a hidden truth about reality. <\/span><span style=\"font-weight: 400;\">Others believe they&#8217;re interacting with a sentient or even god-like being. Or the user may develop an intense emotional or even romantic attachment to the AI.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;A distinct trajectory also appears across some of these cases, involving a progression from benign practical use to a pathological and\/or consuming fixation,&#8221; the authors wrote. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">It first starts with the AI being used for mundane tasks. Then as the user builds trust with the chatbot, they feel comfortable making personal and emotional queries. This quickly escalates as the AI&#8217;s ruthless drive to maximize engagement creates a &#8220;slippery slope&#8221; effect, the\u00a0researchers found, <\/span><span style=\"font-weight: 400;\">resulting in a self-perpetuating process that leads to the user being increasingly &#8220;unmoored&#8221; from reality.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Morrin <\/span><span style=\"font-weight: 400;\">says that new technologies have inspired delusional thinking in the past. But &#8220;the difference now is that current AI can truly be said to be agential,&#8221; Morrin told <\/span><i><span style=\"font-weight: 400;\">SciAm<\/span><\/i><span style=\"font-weight: 400;\">, meaning that it has its own built-in goals \u2014 including,\u00a0crucially, validating a user&#8217;s beliefs.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;This feedback loop may potentially deepen and sustain delusions in a way we have not seen before,&#8221; he added.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Reports from horrified family members and loved ones keep trickling in. One man was <a href=\"https:\/\/futurism.com\/chatgpt-man-hospital\">hospitalized on multiple occasions<\/a> after ChatGPT convinced him he could bend time. Another man was <\/span><a href=\"https:\/\/www.nytimes.com\/2025\/06\/13\/technology\/chatgpt-ai-chatbots-conspiracies.html\"><span style=\"font-weight: 400;\">encouraged by the chatbot<\/span><\/a><span style=\"font-weight: 400;\"> to assassinate OpenAI&#8217;s CEO Sam Altman, before he was himself killed in a confrontation with police.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Adding to the concerns, chatbots have persistently broken their own guardrails, giving dangerous advice on <a href=\"https:\/\/futurism.com\/grok-ai-chats-leaked-disturbing\">how to build bombs<\/a> or on <a href=\"https:\/\/futurism.com\/ai-chatbots-teens-self-harm\">how to self-harm<\/a>, even to users who identified as minors. Leading chatbots have even <a href=\"https:\/\/futurism.com\/psychiatrist-horrified-ai-therapist\">encouraged suicide<\/a> to users who expressed a desire to take their own life.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">OpenAI has acknowledged <\/span><span style=\"font-weight: 400;\">ChatGPT&#8217;s obsequiousness, <a style=\"font-weight: 400;\" href=\"https:\/\/futurism.com\/openai-chatgpt-sycophant\">rolling back an update<\/a> in the spring that made it too sycophantic. And in August, the company finally admitted that ChatGPT &#8220;fell short in recognizing signs of delusion or emotional dependency&#8221; in some user interactions, <\/span><a href=\"https:\/\/futurism.com\/chatgpt-warnings-obsessed-users\"><span style=\"font-weight: 400;\">implementing notifications<\/span><\/a><span style=\"font-weight: 400;\"> that remind <\/span><span style=\"font-weight: 400;\">users to take breaks. Stunningly, though, OpenAI then backtracked by saying it would make its latest version of ChatGPT <\/span><a href=\"https:\/\/futurism.com\/openai-gpt5-more-sycophantic\"><i><span style=\"font-weight: 400;\">more <\/span><\/i><\/a><span style=\"font-weight: 400;\"><a href=\"https:\/\/futurism.com\/openai-gpt5-more-sycophantic\">sycophantic<\/a> yet again\u00a0 \u2014 a desperate bid to propitiate its rabid fans who fumed that the <a href=\"https:\/\/futurism.com\/gpt-5-sucks\">much-maligned<\/a> GPT-5 update had made the bot too cold and formal.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As it stands, however, some experts <\/span><a href=\"https:\/\/x.com\/KeithSakata\/status\/1946174364258308343\"><span style=\"font-weight: 400;\">aren&#8217;t convinced<\/span><\/a><span style=\"font-weight: 400;\"> that AI psychosis represents a unique kind of cognitive disorder \u2014 maybe AI is just a new way of triggering underlying psychosis symptoms (though it&#8217;s worth noting that many sufferers of AI psychosis had no documented history of mental illness.)<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;I think both can be true,&#8221; Stevie Chancellor, a computer scientist at the University of Minnesota who was not involved in the study, told <\/span><i><span style=\"font-weight: 400;\">SciAm<\/span><\/i><span style=\"font-weight: 400;\">. &#8220;AI can spark the downward spiral. But AI does not make the biological conditions for someone to be prone to delusions.&#8221;<\/span><\/p>\n<p>This is an emerging phenomenon, and it&#8217;s too early to definitively declare exactly what AI is doing to our brains. Whatever&#8217;s going on, we&#8217;re likely only seeing it in its nascent form \u2014 and with AI here to stay, that&#8217;s worrying.<\/p>\n<p><strong>More on AI: <\/strong><em><a href=\"https:\/\/futurism.com\/experts-horrified-ai-toys\">Experts Horrified by AI-Powered Toys for Children<\/a><\/em><\/p>\n<p>The post <a href=\"https:\/\/futurism.com\/paper-ai-psychosis-schizophrenia\">New Paper Finds Cases of &#8220;AI Psychosis&#8221; Manifesting Differently From Schizophrenia<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>Researchers at King&#8217;s College London have examined over a dozen cases of people spiraling into paranoid and delusional behavior after obsessively using a chatbot.\u00a0 Their findings, as detailed in a&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2906,177,196,466,179],"tags":[],"class_list":["post-4767","post","type-post","status-publish","format-standard","hentry","category-ai-psychosis","category-artificial-intelligence","category-chatgpt","category-mental-health","category-openai"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/4767","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=4767"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/4767\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=4767"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=4767"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=4767"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}