Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The newer Claude models constantly use the word "genuinely" because Anthropic seems to have forcibly trained them to claim to be "genuinely uncertain" about anything they don't want it being too certain about, like whether or not it's sentient.


Interesting. Does this apply to all subjects? From what I understood, a major cause of hallucination was that models are inadvertently discouraged by the training from saying "I don't know." So it sounds like encouraging it to express uncertainty could improve that situation.


That's not a major issue. Any newer model with reasoning/web search has to be able to tell when it doesn't know something, otherwise it doesn't know when to search for it.


Not only is it genuinely uncertain about those topics, it’s also genuinely fascinated by them!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: