> It's not a problem if they're wrong 30% of the time if they're able to gauge their own confidence like a human would.
This is a case where I would not use human performance as the standard to beat. Training people to be both intellectually honest and statistically calibrated is really hard.
Perhaps, but an AI that can only answer like a precocious child who's spent years reading encyclopedias but has not learned to detect when it's thinking poorly or not remembering clearly is much less useful.
This is a case where I would not use human performance as the standard to beat. Training people to be both intellectually honest and statistically calibrated is really hard.