Skip to content
Longterm Wiki
Back

MIT News - LLM Shortcoming Makes Them Less Reliable

web

MIT News coverage of research identifying a core reliability limitation in LLMs related to self-knowledge and calibration, relevant to AI safety discussions around hallucination, deployment risk, and the gap between capability and trustworthiness.

Metadata

Importance: 58/100news articlenews

Summary

MIT researchers identified a fundamental shortcoming in large language models where they struggle to reliably distinguish between what they know and what they don't know, leading to overconfident or inconsistent outputs. This limitation undermines trust in LLM-generated information, particularly in high-stakes applications. The findings highlight a key gap between apparent LLM capability and actual reliability.

Key Points

  • LLMs exhibit a systematic inability to accurately assess their own knowledge boundaries, contributing to hallucinations and unreliable outputs.
  • The shortcoming manifests as inconsistent confidence calibration, where models may appear certain about incorrect information.
  • This reliability gap poses particular risks in high-stakes domains like medicine, law, and scientific research where accuracy is critical.
  • The research suggests current LLM architectures have intrinsic limitations in self-knowledge that may require novel approaches to address.
  • Findings have implications for AI deployment strategies, emphasizing the need for external verification and human oversight mechanisms.

Cited by 1 page

PageTypeQuality
Bridgewater AIA LabsOrganization66.0

Cached Content Preview

HTTP 200Fetched Apr 9, 202614 KB
Researchers discover a shortcoming that makes LLMs less reliable | MIT News | Massachusetts Institute of Technology 
 
 
 
 
 
 
 

 
 
 
 
 
 
 Skip to content ↓ 
 
 
 

 
 
 
 
 
 
 

 
 
 
 Massachusetts Institute of Technology 
 
 
 
 
 
 
 
 Search websites, locations, and people 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 See More Results 
 
 
 Suggestions or feedback? 

 
 
 

 
 
 
 
 Enter keywords to search for news articles: 
 
 Submit 
 
 

 
 
 
 
 

 
 
 

 
 
 
 Browse By

 
 
 
 
 Topics

 View All → 
 
 Explore: 
 
 Machine learning 

 Sustainability 

 Startups 

 Black holes 

 Classes and programs 

 
 
 
 
 
 
 
 Departments

 View All → 
 
 Explore: 
 
 Aeronautics and Astronautics 

 Brain and Cognitive Sciences 

 Architecture 

 Political Science 

 Mechanical Engineering 

 
 
 
 
 
 
 
 Centers, Labs, & Programs

 View All → 
 
 Explore: 
 
 Abdul Latif Jameel Poverty Action Lab (J-PAL) 

 Picower Institute for Learning and Memory 

 Media Lab 

 Lincoln Laboratory 

 
 
 
 
 
 
 
 Schools

 
 
 School of Architecture + Planning 

 School of Engineering 

 School of Humanities, Arts, and Social Sciences 

 Sloan School of Management 

 School of Science 

 MIT Schwarzman College of Computing 

 
 
 
 
 
 View all news coverage of MIT in the media → 
 
 
 Listen to audio content from MIT News → 
 
 
 Subscribe to MIT newsletter → 
 
 
 
 
 
 
 
 Close 
 
 
 

 

 

 
 

 
 
 
 

 
 
 

 
 
 
 
 
 
 Breadcrumb

 
 
 MIT News 
 

 
 Researchers discover a shortcoming that makes LLMs less reliable
 

 

 
 
 

 
 
 
 
 
 
 Researchers discover a shortcoming that makes LLMs less reliable 

 
 
 
 
 
 
 

 
 

 
 
 
 
 Large language models can learn to mistakenly link certain sentence patterns with specific topics — and may then repeat these patterns instead of reasoning. 
 

 
 
 

 
 Adam Zewe 
 | 
 MIT News 
 
 

 
 

 Publication Date : 
 
 November 26, 2025 
 
 

 
 
 
 
 Press Inquiries 
 
 

 Press Contact : 
 

 
 
 
 
 
 
 
 MIT Media Relations 

 

 
 
 Email:
 expertrequests@mit.edu 
 

 
 
 Phone:
 617-253-2700 
 

 
 
 Media Download

 
 
 
 

 
 

 
 
 ↓ Download Image 
 
 
 
 
 

 Caption : 
 
 An LLM might learn that a question like “Where is Paris located?” is structured as adverb/verb/proper noun/verb. If the model is given a new question with the same grammatical structure but nonsense words, like “Quickly sit Paris clouded?” it might answer “France” even though that answer makes no sense. 

 Credits : 
 
 Image: MIT News; iStock 
 

 

 
 
 
 *Terms of Use:

 
 Images for download on the MIT News office website are made available to non-commercial entities, press and the general public under a 
 Creative Commons Attribution Non-Commercial No Derivatives license .
 You may not alter the images provided, other than to crop them to size. A credit line must be used when reproducing images; if one is not provided 
 below, credit the images to "MIT." 
 

 
 
 
 
 
 
 
 Close 
 
 
 

 
 
 
 
 
 
 
 
 

 

 

 

 
 
 
 


... (truncated, 14 KB total)
Resource ID: 8682257c635430e7 | Stable ID: sid_Pwd629aXZw