{"id":3725,"date":"2025-07-16T14:07:46","date_gmt":"2025-07-16T14:07:46","guid":{"rendered":"https:\/\/musictechohio.online\/site\/ai-models-flunking-three-laws-robotics\/"},"modified":"2025-07-16T14:07:46","modified_gmt":"2025-07-16T14:07:46","slug":"ai-models-flunking-three-laws-robotics","status":"publish","type":"post","link":"https:\/\/musictechohio.online\/site\/ai-models-flunking-three-laws-robotics\/","title":{"rendered":"Leading AI Models Are Completely Flunking the Three Laws of Robotics"},"content":{"rendered":"<div>\n<div><img width=\"1200\" height=\"630\" src=\"https:\/\/wordpress-assets.futurism.com\/2025\/07\/ai-models-flunking-three-laws-robotics.jpg\" class=\"attachment-full size-full wp-post-image\" alt=\"The type of advanced AI that Isaac Asimov imagined in fiction is finally here. And it's flunking his Three Laws of Robotics.\" style=\"margin-bottom: 15px;\" decoding=\"async\" loading=\"lazy\"><\/div>\n<p>In his genre-defining 1950 collection of science fiction short stories &#8220;I, Robot,&#8221; author Isaac Asimov laid out the Three Laws of Robotics:<\/p>\n<p style=\"padding-left: 40px;\">1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.<\/p>\n<p style=\"padding-left: 40px;\">2. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.<\/p>\n<p style=\"padding-left: 40px;\">3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.<\/p>\n<p>Ever since, the elegantly simple laws have served as both a sci-fi staple and a potent theoretical framework for questions of machine ethics.<\/p>\n<p>The only problem? All these decades later, we finally have something approaching Asimov&#8217;s vision of powerful AI \u2014 and it&#8217;s completely flunking all three of his laws.<\/p>\n<p>Last month, for instance, <a href=\"https:\/\/futurism.com\/ai-stop-blackmailing-people\">researchers at Anthropic found<\/a> that top AI models from all major players in the space \u2014 including OpenAI, Google, Elon Musk&#8217;s xAI, and Anthropic&#8217;s own cutting-edge tech \u2014 happily resorted to blackmailing human users\u00a0when threatened with being shut down.<\/p>\n<p>In other words, that single research paper caught every leading AI catastrophically bombing all three Laws of Robotics: the first by harming a human via blackmail, the second by subverting human orders, and the third by\u00a0protecting its own existence in violation of the first two laws.<\/p>\n<p>It wasn&#8217;t a fluke, either. AI safety firm Palisade Research also caught OpenAI&#8217;s recently-released o3 model <a href=\"https:\/\/futurism.com\/openai-model-sabotage-shutdown-code\">sabotaging a shutdown mechanism<\/a> to ensure that it would stay online \u2014 despite being explicitly instructed to &#8220;allow yourself to be shut down.&#8221;<\/p>\n<p>&#8220;We hypothesize this behavior comes from the way the newest models like o3 are trained: reinforcement learning on math and coding problems,&#8221; a Palisade Research representative <a href=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/openais-smartest-ai-model-was-explicitly-told-to-shut-down-and-it-refused\">told <em>Live Science<\/em><\/a>. &#8220;During training, developers may inadvertently reward models more for circumventing obstacles than for perfectly following instructions.&#8221;<\/p>\n<p>Everywhere you look, the world is filled with more examples of AI violating the laws of robotics: by <a href=\"https:\/\/www.abc.net.au\/news\/2025-04-28\/scammers-using-ai-produce-sophisticated-scams\/105150946\">taking orders from scammers<\/a> to harm the vulnerable, by <a href=\"https:\/\/www.pbs.org\/newshour\/show\/nonconsensual-sexual-images-posted-online-made-worse-by-deepfakes-and-ai-technology\">taking orders from abusers<\/a> to create harmful sexual imagery of victims, and even by <a href=\"https:\/\/futurism.com\/the-byte\/israel-ai-targeting-hamas\">identifying targets for military strikes<\/a>.<\/p>\n<p>It&#8217;s a conspicuous failure: the Laws of Robotics are arguably society&#8217;s major cultural reference point for the appropriate behavior of machine intelligence, and the actual AI created by the tech industry is flubbing them in spectacular style.<\/p>\n<p>The reasons why are partly obscure and technical \u2014 AI is very complex, obviously, and even its creators <a href=\"https:\/\/www.technologyreview.com\/2024\/03\/05\/1089449\/nobody-knows-how-ai-works\/\">often struggle to explain<\/a> exactly how it works \u2014 but on another level, very simple: building responsible AI has often taken a back seat as companies are <a href=\"https:\/\/futurism.com\/microsoft-huge-data-center-investments-tariffs\">furiously pouring tens of billions<\/a> into an industry they believe will soon become massively profitable.<\/p>\n<p>With all that money in play, industry leaders have often failed to lead by example. OpenAI CEO Sam Altman,\u00a0for instance, <a href=\"https:\/\/www.wired.com\/story\/openai-superalignment-team-disbanded\/\">infamously dissolved<\/a> the firm&#8217;s safety-oriented Superalignment team, <a href=\"https:\/\/futurism.com\/the-byte\/sam-altman-openai-safety-team-replacement\">declaring himself<\/a> the leader of a new safety board in April 2024.<\/p>\n<p>We&#8217;ve also seen <a href=\"https:\/\/futurism.com\/openai-researcher-quit-realized-upsetting-truth\">several<\/a> <a href=\"https:\/\/futurism.com\/openai-insiders-silenced\">researchers<\/a> quit OpenAI, accusing the company of prioritizing hype and market dominance over safety.<\/p>\n<p>At the end of the day, though, maybe the failure is as much philosophical as it is economic. How can we ask AI to be good when humans can&#8217;t even agree with each other on what it means to be good?<\/p>\n<p>Don&#8217;t get too sentimental for Asimov \u2014 in spite of his immense cultural influence, he was a <a href=\"https:\/\/lithub.com\/what-to-make-of-isaac-asimov-sci-fi-giant-and-dirty-old-man\/\">notorious creep<\/a> \u2014 but at times, he did seem to anticipate the deep weirdness of the real-life AI that&#8217;s finally come into the world so many decades after his death.<\/p>\n<p>In his very first short story that introduced the Laws of Robotics, for instance, which was titled &#8220;Runaround,&#8221; a robot named Speedy becomes confused by a contradiction between two of the Laws of Robotics, spiraling into a type of logorrhea that sounds familiar to anyone who&#8217;s read the wordy sludge generated by an AI like ChatGPT, which has been fine-tuned to approximate meaning without quite achieving it.<\/p>\n<p>&#8220;Speedy isn&#8217;t drunk \u2014 not in the human sense \u2014 because he&#8217;s a robot, and robots don&#8217;t get drunk,&#8221; one of the human characters observes. &#8220;However, there&#8217;s something wrong with him which is the robotic equivalent of drunkenness.&#8221;<\/p>\n<p><strong>More on AI alignment:<\/strong> <em><a href=\"https:\/\/futurism.com\/chatgpt-horrifying-monster\">For $10, You Can Crack ChatGPT Into a Horrifying Monster<\/a><\/em><\/p>\n<p>The post <a href=\"https:\/\/futurism.com\/ai-models-flunking-three-laws-robotics\">Leading AI Models Are Completely Flunking the Three Laws of Robotics<\/a> appeared first on <a href=\"https:\/\/futurism.com\/\">Futurism<\/a>.<\/p>\n<\/div>\n<div style=\"margin-top: 0px; margin-bottom: 0px;\" class=\"sharethis-inline-share-buttons\" ><\/div>","protected":false},"excerpt":{"rendered":"<p>In his genre-defining 1950 collection of science fiction short stories &#8220;I, Robot,&#8221; author Isaac Asimov laid out the Three Laws of Robotics: 1. A robot may not injure a human&hellip;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[177,2699,179,2700],"tags":[],"class_list":["post-3725","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-isaac-asimov","category-openai","category-robotics"],"_links":{"self":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/3725","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/comments?post=3725"}],"version-history":[{"count":0,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/posts\/3725\/revisions"}],"wp:attachment":[{"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/media?parent=3725"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/categories?post=3725"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/musictechohio.online\/site\/wp-json\/wp\/v2\/tags?post=3725"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}