Semantic Entropy as a Regularizer for LLM Calibration
Posted on Sat 03 January 2026 in Machine Learning • Tagged with rlhf, uncertainty
This post explores using semantic entropy as a training signal for calibrating confidence in language models.
Here is what I found:
- Training on semantic entropy alone does not converge and leads to unstable behavior
- In a data abundant/compute scarce regime, standard Brier score supervision achieves strong calibration on its …
Continue reading