Close Menu
geekfence.comgeekfence.com
    What's Hot

    White House says it has info indicating “foreign entities, principally based in China” are engaged in “industrial scale distillation” of American AI tech (Financial Times)

    April 23, 2026

    Cisco researchers take quantum leap for networks

    April 23, 2026

    Teaching AI models to say “I’m not sure” | MIT News

    April 23, 2026
    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    Facebook Instagram
    geekfence.comgeekfence.com
    • Home
    • UK Tech News
    • AI
    • Big Data
    • Cyber Security
      • Cloud Computing
      • iOS Development
    • IoT
    • Mobile
    • Software
      • Software Development
      • Software Engineering
    • Technology
      • Green Technology
      • Nanotechnology
    • Telecom
    geekfence.comgeekfence.com
    Home»Artificial Intelligence»Teaching AI models to say “I’m not sure” | MIT News
    Artificial Intelligence

    Teaching AI models to say “I’m not sure” | MIT News

    AdminBy AdminApril 23, 2026No Comments4 Mins Read0 Views
    Facebook Twitter Pinterest LinkedIn Telegram Tumblr Email
    Teaching AI models to say “I’m not sure” | MIT News
    Share
    Facebook Twitter LinkedIn Pinterest Email



    Confidence is persuasive. In artificial intelligence systems, it is often misleading.

    Today’s most capable reasoning models share a trait with the loudest voice in the room: They deliver every answer with the same unshakable certainty, whether they’re right or guessing. Researchers at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have now traced that overconfidence to a specific flaw in how these models are trained, and developed a method that fixes it without giving up any accuracy.

    The technique, called RLCR (Reinforcement Learning with Calibration Rewards), trains language models to produce calibrated confidence estimates alongside their answers. In addition to coming up with an answer, the model thinks about its uncertainty in that answer, and outputs a confidence score. In experiments across multiple benchmarks, RLCR reduced calibration error by up to 90 percent while maintaining or improving accuracy, both on the tasks the model was trained on and on entirely new ones it had never seen. The work will be presented at the International Conference on Learning Representations later this month.

    The problem traces to a surprisingly simple source. The reinforcement learning (RL) methods behind recent breakthroughs in AI reasoning, including the training approach used in systems like OpenAI’s o1, reward models for getting the right answer, and penalize them for getting it wrong. Nothing in between. A model that arrives at the correct answer through careful reasoning receives the same reward as one that guesses correctly by chance. Over time, this trains models to confidently answer every question they are asked, whether they have strong evidence or are effectively flipping a coin.

    That overconfidence has consequences. When models are deployed in medicine, law, finance, or any setting where users make decisions based on AI outputs, a system that expresses high confidence regardless of its actual certainty becomes unreliable in ways that are difficult to detect from the outside. A model that says “I’m 95 percent sure” when it is right only half the time is more dangerous than one that simply gets the answer wrong, because users have no signal to seek a second opinion.

    “The standard training approach is simple and powerful, but it gives the model no incentive to express uncertainty or say I don’t know,” says Mehul Damani, an MIT PhD student and co-lead author on the paper. “So the model naturally learns to guess when it is unsure.” 

    RLCR addresses this by adding a single term to the reward function: a Brier score, a well-established measure that penalizes the gap between a model’s stated confidence and its actual accuracy. During training, models learn to reason about both the problem and their own uncertainty, producing an answer and a confidence estimate together. Confidently wrong answers are penalized. So are unnecessarily uncertain correct ones.

    The math backs it up: the team proved formally that this type of reward structure guarantees models that are both accurate and well-calibrated. They then tested the approach on a 7-billion-parameter model across a range of question-answering and math benchmarks, including six datasets the model had never been trained on.

    The results showed a consistent pattern. Standard RL training actively degraded calibration compared to the base model, making models worse at estimating their own uncertainty. RLCR reversed that effect, substantially improving calibration with no loss in accuracy. The method also outperformed post-hoc approaches, in which a separate classifier is trained to assign confidence scores after the fact. “What’s striking is that ordinary RL training doesn’t just fail to help calibration. It actively hurts it,” says Isha Puri, an MIT PhD student and co-lead author. “The models become more capable and more overconfident at the same time.”

    The team also demonstrated that the confidence estimates produced by RLCR are practically useful at inference time. When models generate multiple candidate answers, selecting the one with the highest self-reported confidence, or weighting votes by confidence in a majority-voting scheme, improves both accuracy and calibration as compute scales.

    An additional finding suggests that the act of reasoning about uncertainty itself has value. The researchers trained classifiers on model outputs and found that including the model’s explicit uncertainty reasoning in the input improved the classifier’s performance, particularly for smaller models. The model’s self-reflective reasoning about what it does and doesn’t know contains real information, not just decoration.

    In addition to Damani and Puri, other authors on the paper are Stewart Slocum, Idan Shenfeld, Leshem Choshen, and senior authors Jacob Andreas and Yoon Kim.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

    Related Posts

    Train, Serve, and Deploy a Scikit-learn Model with FastAPI

    April 22, 2026

    AI swarms could hijack democracy without anyone noticing

    April 21, 2026

    Agentic AI costs more than you budgeted. Here’s why.

    April 20, 2026

    How Much Coding Is Required To Work in AI and LLM-related Jobs?

    April 19, 2026

    Posit AI Blog: Implementing rotation equivariance: Group-equivariant CNN from scratch

    April 18, 2026

    The Download: bad news for inner Neanderthals, and AI warfare’s human illusion

    April 17, 2026
    Top Posts

    Understanding U-Net Architecture in Deep Learning

    November 25, 202531 Views

    Redefining AI efficiency with extreme compression

    March 25, 202625 Views

    Hard-braking events as indicators of road segment crash risk

    January 14, 202625 Views
    Don't Miss

    White House says it has info indicating “foreign entities, principally based in China” are engaged in “industrial scale distillation” of American AI tech (Financial Times)

    April 23, 2026

    Featured Podcasts Access: Silicon Valley finally has its “Succession” A show about the tech industry’s…

    Cisco researchers take quantum leap for networks

    April 23, 2026

    Teaching AI models to say “I’m not sure” | MIT News

    April 23, 2026

    Using Apache Sedona with AWS Glue to process billions of daily points from a geospatial dataset

    April 23, 2026
    Stay In Touch
    • Facebook
    • Instagram
    About Us

    At GeekFence, we are a team of tech-enthusiasts, industry watchers and content creators who believe that technology isn’t just about gadgets—it’s about how innovation transforms our lives, work and society. We’ve come together to build a place where readers, thinkers and industry insiders can converge to explore what’s next in tech.

    Our Picks

    White House says it has info indicating “foreign entities, principally based in China” are engaged in “industrial scale distillation” of American AI tech (Financial Times)

    April 23, 2026

    Cisco researchers take quantum leap for networks

    April 23, 2026

    Subscribe to Updates

    Please enable JavaScript in your browser to complete this form.
    Loading
    • About Us
    • Contact Us
    • Disclaimer
    • Privacy Policy
    • Terms and Conditions
    © 2026 Geekfence.All Rigt Reserved.

    Type above and press Enter to search. Press Esc to cancel.