Trained on Tokens, Calibrated on Concepts: The Emergence of Semantic Calibration in LLMs
Large Language Models (LLMs) often lack meaningful confidence estimates for their outputs. While base LLMs are known to exhibit next-token calibration, it remains unclear whether they can assess confidence in the actual meaning of their responses beyond the token level. We find that, when using a certain sampling-based notion of semantic calibration, base LLMs are …
Read more “Trained on Tokens, Calibrated on Concepts: The Emergence of Semantic Calibration in LLMs”