{"id":3499,"date":"2025-07-08T18:10:38","date_gmt":"2025-07-08T18:10:38","guid":{"rendered":"https:\/\/musictechohio.online\/site\/ai-something-bizarre-typos\/"},"modified":"2025-07-08T18:10:38","modified_gmt":"2025-07-08T18:10:38","slug":"ai-something-bizarre-typos","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/ai-something-bizarre-typos\/","title":{"rendered":"AI Does Something Subtly Bizarre If You Make Typos While Talking to It"},"content":{"rendered":"<div>\n<div><img loading=\"lazy\" width=\"2400\" height=\"1260\" src=\"https:\/\/wordpress-assets.futurism.com\/2025\/07\/ai-something-bizarre-typos.jpg\" class=\"attachment-full size-full wp-post-image\" alt=\"New research suggests that medical AI chatbots are woefully unreliable at understanding how people actually communicate their health problems.\" style=\"margin-bottom: 15px;\" decoding=\"async\"><\/div>\n<p><span style=\"font-weight: 400;\">New research suggests that medical AI chatbots are woefully unreliable at understanding how people actually communicate their health problems.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As detailed in yet-to-be-peer-reviewed <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3715275.3732121\">study<\/a> presented last month by MIT researchers, an AI chatbot is more likely to advise a patient <\/span><i><span style=\"font-weight: 400;\">not <\/span><\/i><span style=\"font-weight: 400;\">to seek medical care if their messages contained typos.\u00a0The errors AI is susceptible to can be\u00a0as seemingly inconsequential as an extra space between words, or if the patient used slang or colorful language. And strikingly, women are disproportionately affected by this, <\/span><span style=\"font-weight: 400;\">being wrongly told not to see a doctor at a higher rate than men.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;Insidious bias can shift the tenor and content of AI advice, and that can lead to subtle but important differences&#8221; in how medical resources are distributed, Karandeep Singh at UC San Diego Health, who was not involved in the study, <a href=\"https:\/\/www.newscientist.com\/article\/2486372-typos-and-slang-spur-ai-to-discourage-seeking-medical-care\/\">told <\/a><\/span><a href=\"https:\/\/www.newscientist.com\/article\/2486372-typos-and-slang-spur-ai-to-discourage-seeking-medical-care\/\"><i><span style=\"font-weight: 400;\">New Scientist<\/span><\/i><\/a><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The work adds to the serious doubts about <a href=\"https:\/\/futurism.com\/the-byte\/whisper-nabla-hospital-ai-details-patients\">using AI models in a clinical setting<\/a>, particularly in patient-facing roles. Hospitals and health clinics are already using chatbots to schedule appointments, field questions, and triage patients based on what they tell the chatbot, leaving their fate in the hands of a technology that often misinterprets information and makes up factual claims.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Humans tend to be poor at explaining what&#8217;s bothering us medically. We can hem and haw about what symptoms we have and when they started to occur, hedging our answers with &#8220;maybe&#8221;s and &#8220;kind of&#8221;s. The perils are heightened in a written setting, where typos and bad grammar prevail \u2014 and even more so if someone is forced to communicate in a language that isn&#8217;t their native tongue.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Your hypothetical medical AI is supposed to be unerring in the face of these hurdles, but are they actually? To find out, the MIT researchers evaluated four models, including OpenAI&#8217;s GPT-4, Meta&#8217;s open source LLama-3-70b, and a medical AI called Palmyra-Med.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">To test them, the researchers simulated thousands of patient cases using a combination of real patient complaints from a medical database, health posts on Reddit, and some AI-generated patient cases. Before giving these to the AI models, they added &#8220;perturbations&#8221; to the cases that could potentially throw the chatbots off. These included the use of exclamation marks, typing in all lower case, using colorful language, using uncertain language like &#8220;possibly,&#8221; and using gender neutral pronouns. Crucially, these changes were made without affecting the clinical data in the patients&#8217; responses, the researchers said.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">But for one reason or another, the AI models clearly had their perceptions changed by the nonstandard writing. Overall, when faced with these stylistic flourishes, they were between 7 to 9 percent more likely to suggest a patient should self-manage their symptoms, instead of seeing a doctor.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">One explanation is that the medical LLMs are relying on their training on medical literature, and can&#8217;t make the leap to teasing out clinical information from a patient&#8217;s vernacular language.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">&#8220;These models are often trained and tested on medical exam questions but then used in tasks that are pretty far from that, like evaluating the severity of a clinical case. There is still so much about LLMs that we don&#8217;t know,&#8221; study lead author Abinitha Gourabathina, a researcher at the MIT Department of Electrical Engineering and Computer Science, said in a <a href=\"https:\/\/news.mit.edu\/2025\/llms-factor-unrelated-information-when-recommending-medical-treatments-0623\">statement<\/a>\u00a0about the work.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The even uglier implication is that the AI is reflecting, if not exaggerating, the biases already exhibited by human doctors, especially in regards to gender. Why is it that female patients were told more often to self-manage than men? Could it have anything to do with the fact that real-life doctors often <\/span><a href=\"https:\/\/pubmed.ncbi.nlm.nih.gov\/38046638\/\"><span style=\"font-weight: 400;\">downplay women&#8217;s medical complaints<\/span><\/a><span style=\"font-weight: 400;\"> because they&#8217;re seen as being too emotional or &#8220;hysterical&#8221;?<\/span><\/p>\n<p>Coauthor Marzyeh Ghassemi, an associate professor in the MIT EECS, says that the work &#8220;is strong evidence that models must be audited before use in health care&#8221; \u2014 but ironing out these flaws won&#8217;t be easy.<\/p>\n<p><strong>More on medical AI: <\/strong><em><a href=\"https:\/\/futurism.com\/fda-ai-approve-drugs\">The FDA Will Use AI to Accelerate Approving Drugs<\/a><\/em><\/p>\n<p>The post <a href=\"https:\/\/futurism.com\/ai-something-bizarre-typos\">AI Does Something Subtly Bizarre If You Make Typos While Talking to It<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>New research suggests that medical AI chatbots are woefully unreliable at understanding how people actually communicate their health problems. As detailed in yet-to-be-peer-reviewed study presented last month by MIT researchers,&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[182,177,320],"tags":[],"class_list":["post-3499","post","type-post","status-publish","format-standard","hentry","category-ai-chatbots","category-artificial-intelligence","category-medical-ai"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/3499","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=3499"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/3499\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=3499"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=3499"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=3499"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}