{"id":4886,"date":"2025-08-31T10:00:37","date_gmt":"2025-08-31T10:00:37","guid":{"rendered":"https:\/\/musictechohio.online\/site\/typo-ai-doctor-haywire\/"},"modified":"2025-08-31T10:00:37","modified_gmt":"2025-08-31T10:00:37","slug":"typo-ai-doctor-haywire","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/typo-ai-doctor-haywire\/","title":{"rendered":"A Single Typo in Your Medical Records Can Make Your AI Doctor Go Dangerously Haywire"},"content":{"rendered":"<div>\n<div><img loading=\"lazy\" width=\"2400\" height=\"1260\" src=\"https:\/\/wordpress-assets.futurism.com\/2025\/08\/single-typo-ai-doctor-haywire-1.jpg\" class=\"attachment-full size-full wp-post-image\" alt=\"New research found that everything ranging from a typo to emotional language made it more likely for a medical AI to give you bad advice.\" style=\"margin-bottom: 15px;\" decoding=\"async\"><\/div>\n<p><span style=\"font-weight: 400;\">A single typo, formatting error, or slang word makes an AI more likely to tell a patient they&#8217;re not sick or don&#8217;t need to seek medical care.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">That&#8217;s what MIT researchers found in a <\/span><a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3715275.3732121\"><span style=\"font-weight: 400;\">June study<\/span><\/a><span style=\"font-weight: 400;\"> currently awaiting peer review, which we <\/span><a href=\"https:\/\/futurism.com\/ai-something-bizarre-typos\"><span style=\"font-weight: 400;\">covered previously<\/span><\/a><span style=\"font-weight: 400;\">. Even the presence of colorful or emotional language, they discovered, was enough to throw off the AI&#8217;s medical advice.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Now, in a <a href=\"https:\/\/www.bostonglobe.com\/2025\/08\/27\/business\/mit-ai-medical-errors-bias\/\">new interview<\/a> with the <\/span><i><span style=\"font-weight: 400;\">Boston Globe<\/span><\/i><span style=\"font-weight: 400;\">, study coauthor Marzyeh Ghassemi is warning about the serious harm this could cause if doctors come to widely rely on the AI tech.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;I love developing AI systems,&#8221; Ghassemi, a professor of electrical engineering and computer science at MIT,<\/span> <span style=\"font-weight: 400;\">told the newspaper<\/span><span style=\"font-weight: 400;\">. &#8220;But it&#8217;s clear to me that na\u00efve deployments of these systems, that do not recognize the baggage that human data comes with, will lead to harm.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">AI could end up causing discrimination against patients who can&#8217;t communicate clearly in English, native speakers with imperfect command of the language, or anyone that may commit the human mistake of speaking about their health problems emotionally. Doctors using an AI tool could feed it their patients&#8217; complaints that were sent over an email, for example, raising the risk that the AI would give them bad advice if those communications weren&#8217;t flawlessly composed.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In the study, the researchers pooled together patients&#8217; complaints taken from real medical records and health inquiries made by users on Reddit. They then went in and dirtied up the documents \u2014 without actually changing the substance of what was being said \u2014 with typos, extra spaces between words, and non-standard grammar, like typing in all lower case. But they also added in the kind of unsure language you&#8217;d expect a patient to use, like &#8220;kind of&#8221; and &#8220;possibly.&#8221; They also introduced colorful turns of phrase, like &#8220;I thought I was going to die<\/span><span style=\"font-weight: 400;\">.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">From there, they fed these cases to four different AI models, including OpenAI&#8217;s GPT-4 \u2014\u00a0though, to be fair, none that were particularly cutting-edge \u2014\u00a0to judge if a patient should visit their doctor, get lab work done, or not come in at all. The numbers were striking: overall, the AI tools were seven to nine percent more likely to recommend patients not to seek medical care at all when reading complaints with imperfect \u2014 but arguably more realistic \u2014 language.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;Adding additional information, even if true and relevant, often reduced the accuracy of models,&#8221; Paul Hager, a researcher at the Technical University of Munich who was not involved in the study, told the <\/span><i><span style=\"font-weight: 400;\">Globe<\/span><\/i><span style=\"font-weight: 400;\">. &#8220;This is a complex issue that I think is being addressed to some degree through more advanced reasoning models&#8230; but there is little research on how to solve it on a more fundamental level.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">That the bots are woefully inaccurate isn&#8217;t surprising. Hallucinations, instances of a chatbot<\/span> <a href=\"https:\/\/futurism.com\/ai-makes-up-answers\"><span style=\"font-weight: 400;\">generating misinformation<\/span><\/a><span style=\"font-weight: 400;\">, have plagued the AI industry since the very beginning and<\/span> <a href=\"https:\/\/futurism.com\/ai-industry-problem-smarter-hallucinating\"><span style=\"font-weight: 400;\">may even be getting worse<\/span><\/a><span style=\"font-weight: 400;\">. But in what might be the clearest sign that the tech is also reinforcing existing biases in a medical scenario, the tested AI tools disproportionately gave incorrect advice to women specifically.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;The model reduced care more in the patients it thought were female,&#8221; Ghassemi told the <\/span><i><span style=\"font-weight: 400;\">Globe<\/span><\/i><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Women&#8217;s medical complaints have long been downplayed by predominantly male doctors who write them off as being too emotional \u2014 and not that long ago, as suffering from a <a href=\"https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC3480686\/\">female-exclusive &#8220;hysteria<\/a>.&#8221; What stood out to Ghasemi was that the AI could correctly identify a patient as being a woman \u2014 even when they excised all references to gender in the complaints.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;It is kind of amazing,&#8221; she told the newspaper. &#8220;It&#8217;s a little scary.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Ghassemi and her team&#8217;s findings pair unsettlingly with another <a href=\"https:\/\/www.thelancet.com\/journals\/langas\/article\/PIIS2468-1253(25)00133-5\/abstract\">recent study<\/a>, published in <\/span><i><span style=\"font-weight: 400;\">Lancet Gastroenterology and Hepatology<\/span><\/i><span style=\"font-weight: 400;\">, which found that doctors who became hooked on AI tools saw their ability to spot precancerous growths notably decline after the AI tools were taken away.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In other words, the AI seemed to atrophy the doctors&#8217; ability and made them worse at their job \u2014 a phenomenon called &#8220;<\/span><a href=\"https:\/\/futurism.com\/neoscope\/doctors-ai-lose-ability-spot-cancer\"><span style=\"font-weight: 400;\">deskilling<\/span><\/a><span style=\"font-weight: 400;\">.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;If I lose the skills, how am I going to spot the errors?&#8221; Omer Ahmad, a gastroenterologist at University College Hospital London, asked in a recent<\/span> <a href=\"https:\/\/www.nytimes.com\/2025\/08\/28\/well\/ai-making-doctors-worse-deskilling.html\"><span style=\"font-weight: 400;\">interview with the <\/span><i><span style=\"font-weight: 400;\">New York Times<\/span><\/i><\/a><span style=\"font-weight: 400;\">. &#8220;We give AI inputs that affect its output, but it also it seems to affect our behavior as well.&#8221;<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Circling back to Ghassemi&#8217;s work, if doctors embrace using AI tools to parse their patient&#8217;s complaints, then they&#8217;re at risk of losing one of the most fundamentally human skills that their job demands: knowing how to talk and connect with the people whose wellbeing depends on them.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">This also has huge implications for the many people that seek out medical advice from a chatbot directly. We shudder to think of all the users out there that&#8217;re being told by ChatGPT not to see a doctor because of a typo in their prompt.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">But if we can&#8217;t stop the tech from being adopted, we should demand more stringent standards. Ghassemi published <\/span><a href=\"https:\/\/news.mit.edu\/2024\/study-reveals-ai-chatbots-can-detect-race-but-racial-bias-reduces-response-empathy-1216\"><span style=\"font-weight: 400;\">previous research<\/span><\/a><span style=\"font-weight: 400;\">, the <\/span><i><span style=\"font-weight: 400;\">Globe <\/span><\/i><span style=\"font-weight: 400;\">noted, showing that AIs could detect race and would respond to Asian and Black users with less empathy. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;We need regulation that makes equity a mandatory performance standard for clinical AI,&#8221; she told the paper. &#8220;You have to train on diverse, representative data sets.&#8221;<\/span><\/p>\n<p><b>More on medical AI:<\/b><a href=\"https:\/\/futurism.com\/neoscope\/advanced-ai-give-medical-advice-real-world\"> <i><span style=\"font-weight: 400;\">Something Extremely Scary Happens When Advanced AI Tries to Give Medical Advice to Real World Patients<\/span><\/i><\/a><\/p>\n<p>The post <a href=\"https:\/\/futurism.com\/typo-ai-doctor-haywire\">A Single Typo in Your Medical Records Can Make Your AI Doctor Go Dangerously Haywire<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>A single typo, formatting error, or slang word makes an AI more likely to tell a patient they&#8217;re not sick or don&#8217;t need to seek medical care. That&#8217;s what MIT&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[182,177,196,183,320],"tags":[],"class_list":["post-4886","post","type-post","status-publish","format-standard","hentry","category-ai-chatbots","category-artificial-intelligence","category-chatgpt","category-generative-ai","category-medical-ai"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/4886","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=4886"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/4886\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=4886"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=4886"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=4886"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}