{"id":27176,"date":"2025-05-05T09:01:39","date_gmt":"2025-05-05T09:01:39","guid":{"rendered":"https:\/\/medexperts.pro\/?p=27176"},"modified":"2025-05-05T09:29:25","modified_gmt":"2025-05-05T09:29:25","slug":"a-i-hallucinations-are-getting-worse-even-as-new-systems-become-more-powerful","status":"publish","type":"post","link":"https:\/\/medexperts.pro\/?p=27176","title":{"rendered":"A.I. Hallucinations Are Getting Worse, Even as New Systems Become More Powerful"},"content":{"rendered":"<div><\/div>\n<div class=\"css-s99gbd StoryBodyCompanionColumn\" data-testid=\"companionColumn-0\">\n<div class=\"css-53u6y8\">\n<p class=\"css-at9mc1 evys1bk0\">Last month, an A.I. bot that handles tech support for Cursor, <a class=\"css-yywogo\" href=\"https:\/\/www.cursor.com\/\" title rel=\"noopener noreferrer\" target=\"_blank\">an up-and-coming tool for computer programmers<\/a>, alerted several customers about a change in company policy. It said they were no longer allowed to use Cursor on more than just one computer.<\/p>\n<p class=\"css-at9mc1 evys1bk0\">In angry posts to <a class=\"css-yywogo\" href=\"https:\/\/news.ycombinator.com\/item?id=43683012\" title rel=\"noopener noreferrer\" target=\"_blank\">internet message boards<\/a>, the customers complained. Some canceled their Cursor accounts. And some got even angrier when they realized what had happened: The A.I. bot had announced a policy change that did not exist.<\/p>\n<p class=\"css-at9mc1 evys1bk0\">\u201cWe have no such policy. You\u2019re of course free to use Cursor on multiple machines,\u201d the company\u2019s chief executive and co-founder, Michael Truell, <a class=\"css-yywogo\" href=\"https:\/\/old.reddit.com\/r\/cursor\/comments\/1jyy5am\/psa_cursor_now_restricts_logins_to_a_single\/\" title rel=\"noopener noreferrer\" target=\"_blank\">wrote<\/a> in a Reddit post. \u201cUnfortunately, this is an incorrect response from a front-line A.I. support bot.\u201d<\/p>\n<p class=\"css-at9mc1 evys1bk0\">More than two years after <a class=\"css-yywogo\" href=\"https:\/\/www.nytimes.com\/2022\/12\/10\/technology\/ai-chat-bot-chatgpt.html\" title>the arrival of ChatGPT<\/a>, tech companies, office workers and everyday consumers are using A.I. bots for an increasingly wide array of tasks. But there is still <a class=\"css-yywogo\" href=\"https:\/\/www.nytimes.com\/2023\/05\/01\/business\/ai-chatbots-hallucination.html\" title>no way of ensuring that these systems produce accurate information<\/a>.<\/p>\n<\/div>\n<\/div>\n<div data-testid=\"Dropzone-1\"><\/div>\n<div class=\"css-s99gbd StoryBodyCompanionColumn\" data-testid=\"companionColumn-1\">\n<div class=\"css-53u6y8\">\n<p class=\"css-at9mc1 evys1bk0\">The newest and most powerful technologies \u2014 so-called <a class=\"css-yywogo\" href=\"https:\/\/www.nytimes.com\/2025\/03\/26\/technology\/ai-reasoning-chatgpt-deepseek.html\" title>reasoning systems<\/a> from companies like OpenAI, Google and the Chinese start-up DeepSeek \u2014 are generating more errors, not fewer. As their math skills have notably improved, their handle on facts has gotten shakier. It is not entirely clear why.<\/p>\n<p class=\"css-at9mc1 evys1bk0\">Today\u2019s A.I. bots are based on <a class=\"css-yywogo\" href=\"https:\/\/www.nytimes.com\/2018\/03\/06\/technology\/google-artificial-intelligence.html\" title>complex mathematical systems<\/a> that learn their skills by analyzing enormous amounts of digital data. They do not \u2014 and cannot \u2014 decide what is true and what is false. Sometimes, they just make stuff up, a phenomenon some A.I. researchers call hallucinations. On one test, the hallucination rates of newer A.I. systems were as high as 79 percent.<\/p>\n<div class=\"css-kbghgg\">\n<div class=\"css-121kum4\">\n<div class=\"css-171quhb\"><\/div>\n<div class=\"css-asuuk5\">\n<div class=\"css-7axq9l\" data-testid=\"optimistic-truncator-noscript\">\n<div data-testid=\"optimistic-truncator-noscript-message\" class=\"css-6yo1no\">\n<p class=\"css-3kpklk\">We are having trouble retrieving the article content.<\/p>\n<p class=\"css-3kpklk\">Please enable JavaScript in your browser settings.<\/p>\n<\/div>\n<\/div>\n<div class=\"css-1dv1kvn\" id=\"optimistic-truncator-a11y\">\n<hr \/>\n<p>Thank you for your patience while we verify access. If you are in Reader mode please exit and\u00a0<a href=\"https:\/\/myaccount.nytimes.com\/auth\/login?response_type=cookie&amp;client_id=vi&amp;redirect_uri=https%3A%2F%2Fwww.nytimes.com%2F2025%2F05%2F05%2Ftechnology%2Fai-hallucinations-chatgpt-google.html&amp;asset=opttrunc\">log into<\/a>\u00a0your Times account, or\u00a0<a href=\"https:\/\/www.nytimes.com\/subscription?campaignId=89WYR&amp;redirect_uri=https%3A%2F%2Fwww.nytimes.com%2F2025%2F05%2F05%2Ftechnology%2Fai-hallucinations-chatgpt-google.html\">subscribe<\/a>\u00a0for all of The Times.<\/p>\n<hr \/>\n<\/div>\n<div class=\"css-1g71tqy\">\n<div data-testid=\"optimistic-truncator-message\" class=\"css-6yo1no\">\n<p class=\"css-3kpklk\">Thank you for your patience while we verify access.<\/p>\n<p class=\"css-3kpklk\">Already a subscriber?\u00a0<a data-testid=\"log-in-link\" class=\"css-z5ryv4\" href=\"https:\/\/myaccount.nytimes.com\/auth\/login?response_type=cookie&amp;client_id=vi&amp;redirect_uri=https%3A%2F%2Fwww.nytimes.com%2F2025%2F05%2F05%2Ftechnology%2Fai-hallucinations-chatgpt-google.html&amp;asset=opttrunc\">Log in<\/a>.<\/p>\n<p class=\"css-3kpklk\">Want all of The Times?\u00a0<a data-testid=\"subscribe-link\" class=\"css-z5ryv4\" href=\"https:\/\/www.nytimes.com\/subscription?campaignId=89WYR&amp;redirect_uri=https%3A%2F%2Fwww.nytimes.com%2F2025%2F05%2F05%2Ftechnology%2Fai-hallucinations-chatgpt-google.html\">Subscribe<\/a>.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>Last month, an A.I. bot that handles tech support for Cursor, an up-and-coming tool for computer programmers, alerted several customers about a change in company policy. It said they were no longer allowed to use Cursor on more than just one computer.In angry posts to internet message boards, the customers complained. Some canceled their Cursor accounts. And some got even angrier when they realized what had happened: The A.I. bot had announced a policy change that did not exist.\u201cWe have no such policy. You\u2019re of course free to use Cursor on multiple machines,\u201d the company\u2019s chief executive and co-founder, Michael Truell, wrote in a Reddit post. \u201cUnfortunately, this is an incorrect response from a front-line A.I. support bot.\u201dMore than two years after the arrival of ChatGPT, tech companies, office workers and everyday consumers are using A.I. bots for an increasingly wide array of tasks. But there is still no way of ensuring that these systems produce accurate information.The newest and most powerful technologies \u2014 so-called reasoning systems from companies like OpenAI, Google and the Chinese start-up DeepSeek \u2014 are generating more errors, not fewer. As their math skills have notably improved, their handle on facts has gotten shakier. It is not entirely clear why.Today\u2019s A.I. bots are based on complex mathematical systems that learn their skills by analyzing enormous amounts of digital data. They do not \u2014 and cannot \u2014 decide what is true and what is false. Sometimes, they just make stuff up, a phenomenon some A.I. researchers call hallucinations. On one test, the hallucination rates of newer A.I. systems were as high as 79 percent.We are having trouble retrieving the article content.Please enable JavaScript in your browser settings.Thank you for your patience while we verify access. If you are in Reader mode please exit and\u00a0log into\u00a0your Times account, or\u00a0subscribe\u00a0for all of The Times.Thank you for your patience while we verify access.Already a subscriber?\u00a0Log in.Want all of The Times?\u00a0Subscribe.<\/p>\n","protected":false},"author":1,"featured_media":27178,"comment_status":"close","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6],"tags":[],"class_list":["post-27176","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-technology"],"_links":{"self":[{"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/posts\/27176","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/medexperts.pro\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=27176"}],"version-history":[{"count":2,"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/posts\/27176\/revisions"}],"predecessor-version":[{"id":27179,"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/posts\/27176\/revisions\/27179"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/medexperts.pro\/index.php?rest_route=\/wp\/v2\/media\/27178"}],"wp:attachment":[{"href":"https:\/\/medexperts.pro\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=27176"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/medexperts.pro\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=27176"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/medexperts.pro\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=27176"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}