[MIT] RLCR: Teaching AI models to say "I'm not sure"
Mirrored from r/LocalLLaMA for archival readability. Support the source by reading on the original site.
Confidence is persuasive. In AI systems, it is often misleading.
Today's most capable reasoning models share a trait with the loudest voice in the room: They deliver every answer with the same unshakable certainty, whether they're right or guessing. Researchers at MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL) have now traced that overconfidence to a specific flaw in how these models are trained, and developed a method that fixes it without giving up any accuracy.
[link] [comments]
More from r/LocalLLaMA
-
inclusionAI/Ring-2.6-1T · Hugging Face
May 14
-
Got local Qwen 3.5/3.6 generating meeting summaries entirely offline on an M4 Max. Demo with Wi-Fi off. This is the future.
May 14
-
NVFP4 Kimi2.6 and Kimi 2.5 released by Nvidia
May 14
-
Dropping learning rate fixed my Qlora fine-tune more than anything else i tried
May 14
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.