{"id":7952,"date":"2026-01-09T17:17:22","date_gmt":"2026-01-09T17:17:22","guid":{"rendered":"https:\/\/musictechohio.online\/site\/chatgpt-health-medical-records-diagnosis\/"},"modified":"2026-01-09T17:17:22","modified_gmt":"2026-01-09T17:17:22","slug":"chatgpt-health-medical-records-diagnosis","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/chatgpt-health-medical-records-diagnosis\/","title":{"rendered":"OpenAI Launches ChatGPT Health, Which Ingests Your Entire Medical Records, But Warns Not to Use It for \u201cDiagnosis or Treatment\u201d"},"content":{"rendered":"<div>\n<p class=\"article-paragraph skip\">AI chatbots may be explosively popular, but they\u2019re known to dispense some seriously wacky\u2014 and potentially dangerous \u2014 health advice, in a flood of easily accessible misinformation that has alarmed experts.<\/p>\n<p class=\"article-paragraph skip\">Their advent has turned countless users into armchair experts, who often end up relying on obsolete, misattributed, or completely made-up advice.<\/p>\n<p class=\"article-paragraph skip\">A <a href=\"https:\/\/www.theguardian.com\/technology\/2026\/jan\/02\/google-ai-overviews-risk-harm-misleading-health-information\" rel=\"nofollow\">recent investigation<\/a> by <em>The Guardian<\/em>, for instance, found that Google\u2019s AI Overviews, which accompany most search results pages, <a href=\"https:\/\/futurism.com\/artificial-intelligence\/google-ai-overviews-dangerous-health-advice\">doled out plenty of inaccurate health information<\/a> that could lead to grave health risks if followed.<\/p>\n<p class=\"article-paragraph skip\">But seemingly unperturbed by experts\u2019 repeated warnings that AI\u2019s health advice shouldn\u2019t be trusted, OpenAI is doubling down by <a href=\"https:\/\/openai.com\/index\/introducing-chatgpt-health\/\" rel=\"nofollow\">launching a new feature<\/a> called ChatGPT Health, which will ingest your medical records to generate responses \u201cmore relevant and useful to you.\u201d<\/p>\n<p class=\"article-paragraph skip\">Yet despite being \u201cdesigned in close collaboration with physicians\u201d and built on \u201cstrong privacy, security, and data controls,\u201d the feature is \u201cdesigned to support, not replace, medical care.\u201d In fact, it\u2019s shipping with a ludicrously self-defeating caveat: that the bespoke health feature is \u201cnot intended for diagnosis or treatment.\u201d<\/p>\n<p class=\"article-paragraph skip\">\u201cChatGPT Health helps people take a more active role in understanding and managing their health and wellness \u2014 while supporting, not replacing, care from clinicians,\u201d the company\u2019s website reads.<\/p>\n<p class=\"article-paragraph skip\">In reality, users are certain to use it for exactly the type of health advice that OpenAI is warning against in the fine print, which is likely to bring fresh new embarrassments for the company. <\/p>\n<p class=\"article-paragraph skip\">It\u2019ll only be heightening existing problems for the company. As <a href=\"https:\/\/www.businessinsider.com\/chatgpt-new-webmd-doctors-lawyers-medical-advice-2026-1\" rel=\"nofollow\"><em>Business Insider<\/em> reports<\/a>, ChatGPT is \u201cmaking amateur lawyers and doctors out of everyone,\u201d to the dismay of legal and medical professionals.<\/p>\n<p class=\"article-paragraph skip\">Miami-based medical malpractice attorney Jonathan Freidin told the publication that people will use chatbots like ChatGPT to fill out his firm\u2019s client contact sheet.<\/p>\n<p class=\"article-paragraph skip\">\u201cWe\u2019re seeing a lot more callers who feel like they have a case because ChatGPT or Gemini told them that the doctors or nurses fell below the standard of care in multiple different ways,\u201d he said. \u201cWhile that may be true, it doesn\u2019t necessarily translate into a viable case.\u201d<\/p>\n<p class=\"article-paragraph skip\">Then there\u2019s the fact that users are willing to surrender medical histories, including highly sensitive and personal information \u2014 a decision that OpenAI is now encouraging with ChatGPT Health \u2014 despite federal law, like HIPAA, not applying to consumer AI products. <\/p>\n<p class=\"article-paragraph skip\">Case in point, billionaire Elon Musk <a href=\"https:\/\/fortune.com\/2024\/11\/20\/elon-musk-x-user-medical-information-privacy-ai-grok\/\" rel=\"nofollow\">encouraged<\/a> people last year to upload their medical data to his ChatGPT competitor Grok, leading to a <a href=\"https:\/\/finance.yahoo.com\/news\/elon-musk-asked-people-upload-200527137.html\" rel=\"nofollow\">flood of confusion<\/a> as users received hallucinated diagnoses after sharing their X-rays and PET scans.<\/p>\n<p class=\"article-paragraph skip\">Given the AI industry\u2019s track record when it comes to privacy protection and struggles with <a href=\"https:\/\/newsroom.ibm.com\/2025-07-30-ibm-report-13-of-organizations-reported-breaches-of-ai-models-or-applications,-97-of-which-reported-lacking-proper-ai-access-controls\" rel=\"nofollow\">significant data leaks<\/a>, all these risks are as pertinent as ever.<\/p>\n<p class=\"article-paragraph skip\">\u201cNew AI health tools offer the promise of empowering patients and promoting better health outcomes, but health data is some of the most sensitive information people can share and it must be protected,\u201d Center for Democracy and Technology senior counsel Andrew Crawford <a href=\"https:\/\/ca.news.yahoo.com\/openai-launches-chatgpt-health-review-171436341.html\" rel=\"nofollow\">told the <em>BBC<\/em><\/a>.<\/p>\n<p class=\"article-paragraph skip\">\u201cEspecially as OpenAI moves to explore advertising as a business model, it\u2019s crucial that separation between this sort of health data and memories that ChatGPT captures from other conversations is airtight,\u201d he added. \u201cSince it\u2019s up to each company to set the rules for how health data is collected, used, shared, and stored, inadequate data protections and policies can put sensitive health information in real danger.\u201d<\/p>\n<p class=\"article-paragraph skip\">\u201cChatGPT is only bound by its own disclosures and promises, so without any meaningful limitation on that, like regulation or a law, ChatGPT can change the terms of its service at any time,\u201d Electronic Privacy Information Center senior counsel Sara Geoghegan <a href=\"https:\/\/therecord.media\/chatgpt-health-draws-concern-privacy-critics\" rel=\"nofollow\">told <em>The Record<\/em><\/a>.<\/p>\n<p class=\"article-paragraph skip\">Then there are concerns over highly sensitive data, like reproductive health information, being passed on to the police against the user\u2019s wishes.<\/p>\n<p class=\"article-paragraph skip\">\u201cHow does OpenAI handle [law enforcement] requests?\u201d Crawford told <em>The Record<\/em>. \u201cDo they just turn over the information? Is the user in any way informed?\u201d<\/p>\n<p class=\"article-paragraph skip\">\u201cThere\u2019s lots of questions there that I still don\u2019t have great answers to,\u201d he added.<\/p>\n<p class=\"article-paragraph skip\"><strong>More on AI and health advice:<\/strong> <a href=\"https:\/\/futurism.com\/artificial-intelligence\/google-ai-overviews-dangerous-health-advice\"><em>Google\u2019s AI Overviews Caught Giving Dangerous \u201cHealth\u201d Advice<\/em><\/a><\/p>\n<p class=\"article-paragraph skip\">\n<p>The post <a href=\"https:\/\/futurism.com\/artificial-intelligence\/chatgpt-health-medical-records-diagnosis\">OpenAI Launches ChatGPT Health, Which Ingests Your Entire Medical Records, But Warns Not to Use It for \u201cDiagnosis or Treatment\u201d<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>AI chatbots may be explosively popular, but they\u2019re known to dispense some seriously wacky\u2014 and potentially dangerous \u2014 health advice, in a flood of easily accessible misinformation that has alarmed&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[177,3841,3844,3955],"tags":[],"class_list":["post-7952","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-ethics","category-health-medicine","category-medical"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/7952","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=7952"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/7952\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=7952"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=7952"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=7952"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}