logoalt Hacker News

drclaulast Friday at 11:44 AM2 repliesview on HN

How do you know the confidence scores are not hallucinated as well?


Replies

kiliankoelast Friday at 12:02 PM

They are, the model has no inherent knowledge about its confidence levels, it just adds plausible-sounding numbers. Obviously they _can_ be plausible, but trusting these is just another level up from trusting the original output.

I read a comment here a few weeks back that LLMs always hallucinate, but we sometimes get lucky when the hallucinations match up with reality. I've been thinking about that a lot lately.

show 2 replies
dfsegoatlast Friday at 12:01 PM

they 100% are unless you provide a RUBRIC / basically make it ordinal.

"Return a score of 0.0 if ...., Return a score of 0.5 if .... , Return a score of 1.0 if ..."