Discussion about this post

User's avatar
Rising Moon Astrology's avatar

I’m thinking of every “crisis” movie ever, and the one analyst in the meetings saying, “I think we need more data here” who gets sent to the basement room with the tiny desk until the 11th hour, doom approaching, when someone says, “get down to the basement now! We need them!” Power loves confirmation bias. We all do, at least a little.

Expand full comment
Jim's avatar

Great article, thanks! I hadn’t seen the effect of post training on calibration before so thanks for pointing this out. A few studies looking at optimal human-AI decision making are pointing to the need for both ‘agents’ to be aware of confidence levels eg. https://escholarship.org/content/qt0fg7g94k/qt0fg7g94k.pdf (which makes sense!) so reliable confidence signals feel very important for progress. It does feel a bit unfair to ask an LLM for its confidence on something after RLHF because it doesn’t know the confidence of the people that provided feedback?! At least, before then it can weigh up how many bits of data are pointing to an answer and have a stab.

Expand full comment
4 more comments...

No posts