{"id":275,"date":"2026-04-22T19:37:24","date_gmt":"2026-04-22T19:37:24","guid":{"rendered":"https:\/\/blog.positionhire.com\/index.php\/2026\/04\/22\/mit-researchers-develop-ai-models-capable-of-expressing-uncertainty\/"},"modified":"2026-04-22T19:37:24","modified_gmt":"2026-04-22T19:37:24","slug":"mit-researchers-develop-ai-models-capable-of-expressing-uncertainty","status":"publish","type":"post","link":"https:\/\/blog.positionhire.com\/index.php\/2026\/04\/22\/mit-researchers-develop-ai-models-capable-of-expressing-uncertainty\/","title":{"rendered":"MIT Researchers Develop AI Models Capable of Expressing Uncertainty"},"content":{"rendered":"<p>Confidence can be compelling, but in AI systems, it can also mislead. The most advanced reasoning models today often express unwavering certainty in their answers, regardless of accuracy. Researchers at MIT&#8217;s CSAIL have identified a flaw in the training of these models that leads to overconfidence and devised a solution that retains accuracy. The new method, RLCR (Reinforcement Learning with Calibration Rewards), enables models to generate calibrated confidence estimates with their responses. This approach involves the model evaluating its uncertainty and providing a confidence score alongside its answer.<\/p>\n<p>Experiments demonstrated that RLCR reduced calibration error by up to 90 percent while maintaining or improving accuracy, even on tasks the model had not been specifically trained for. The findings will be presented at the International Conference on Learning Representations. The issue arises from the reinforcement learning methods that reward models solely for correct answers without considering the process or certainty of arriving at those answers. This can lead models to confidently provide answers even when unsure, which is problematic in fields like medicine or finance where decisions rely on AI outputs.<\/p>\n<p>According to Mehul Damani, an MIT PhD student and co-lead author, the standard training method lacks incentives for models to express uncertainty. RLCR addresses this by incorporating a Brier score into the reward function, penalizing discrepancies between stated confidence and actual accuracy. During training, models learn to assess both the problem and their uncertainty, resulting in simultaneous generation of answers and confidence estimates. Incorrect confident answers and overly uncertain correct ones are both penalized.<\/p>\n<p>The math confirms that this reward structure ensures models are accurate and well-calibrated. Tests on a 7-billion-parameter model across various benchmarks, including unfamiliar datasets, showed that RLCR significantly improved calibration without sacrificing accuracy. Unlike standard RL training, which worsens calibration, RLCR reversed this effect and outperformed post-hoc methods that assign confidence scores after the fact. MIT PhD student and co-lead author Isha Puri noted that typical RL training not only fails to aid calibration but also worsens it.<\/p>\n<p>RLCR\u2019s confidence estimates proved beneficial during inference, enhancing accuracy and calibration by selecting answers based on self-reported confidence. The researchers also discovered that models reasoning about uncertainty improved classifier performance, especially in smaller models. This self-reflective reasoning offers valuable information. The research team included Stewart Slocum, Idan Shenfeld, Leshem Choshen, and senior authors Jacob Andreas and Yoon Kim, along with Damani and Puri.<\/p>\n<p class=\"ainap-source\"><strong>Original Source:<\/strong> <a href=\"https:\/\/news.mit.edu\/2026\/teaching-ai-models-to-say-im-not-sure-0422\" target=\"_blank\" rel=\"noopener noreferrer\">news.mit.edu<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Confidence can be compelling, but in AI systems, it can also mislead. The most advanced reasoning models today often express unwavering certainty in their answers, regardless of accuracy. Researchers at MIT&#8217;s CSAIL have identified a flaw in the training of these models that leads to overconfidence and devised a solution that retains accuracy. The new&#8230;<\/p>\n","protected":false},"author":1,"featured_media":276,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[4],"tags":[],"class_list":["post-275","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-general-posts"],"_links":{"self":[{"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/posts\/275","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/comments?post=275"}],"version-history":[{"count":0,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/posts\/275\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/media\/276"}],"wp:attachment":[{"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/media?parent=275"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/categories?post=275"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.positionhire.com\/index.php\/wp-json\/wp\/v2\/tags?post=275"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}