▲ | lucb1e 3 days ago | |||||||||||||||||||||||||||||||
LLMs cannot tell fact from fiction. What's commonly called hallucinations stems from it not being able to reason, the way that humans appear to be able to do, no matter that some models are called "reasoning" now. It's all the same principle: most likely token in a given position. Adding internal monologue appears to help because, by being forced to break it down (internally, or by spitballing towards the user when they prompted "think step by step"[1]), it creates better context and will thus have a higher probability that the predicted token is a correct one Being trained to be positive is surely why it inserts these specific "great question, you're so right!" remarks, but if you wasn't trained on that, it still couldn't tell you whether you're great or not > I'm pretty sure they want it kissing people's asses The American faux friendliness is not what causes the underlying problem here, so all else being equal, they might as well have it kiss your ass. It's what most English speakers expect from a "friendly assistant" after all [1] https://hn.algolia.com/?dateEnd=1703980800&dateRange=custom&... | ||||||||||||||||||||||||||||||||
▲ | svnt 3 days ago | parent [-] | |||||||||||||||||||||||||||||||
You’re absolutely wrong! This is not how reasoning models work. Chain-of-thought did not produce reasoning models. | ||||||||||||||||||||||||||||||||
|