| ▲ | gck1 4 hours ago | |||||||
Why does this comment appear every time someone complains about CoT becoming more and more inaccessible with Claude? I have entire processes built on top of summaries of CoT. They provide tremendous value and no, I don't care if "model still did the correct thing". Thinking blocks show me if model is confused, they show me what alternative paths existed. Besides, "correct thing" has a lot of meanings and decision by the model may be correct relative to the context it's in but completely wrong relative to what I intended. The proof that thinking tokens are indeed useful is that anthropic tries to hide them. If they were useless, why would they even try all of this? Starting to feel PsyOp'd here. | ||||||||
| ▲ | dgb23 3 hours ago | parent | next [-] | |||||||
Didn't you notice that the stream is not coherent or noisy? Sometimes it goes from thought A to thought B then action C, but A was entirely unnecessary noise that had nothing to do with B and C. I also sometimes had signals in the thinking output that were red flags, or as you said it got confused, but then it didn't matter at all. Now I just never look at the thinking tokens anymore, because I got bamboozled too often. Perhaps when you summarize it, then you might miss some of these or you're doing things differently otherwise. | ||||||||
| ||||||||
| ▲ | quadruple 3 hours ago | parent | prev [-] | |||||||
I agree. Ever since the release of R1, it's like every single American AI company has realized that they actually do not want to show CoT, and then separately that they cannot actually run CoT models profitably. Ever since then, we've seen everyone implement a very bad dynamic-reasoning system that makes you feel like an ass for even daring to ask the model for more than 12 tokens of thought. | ||||||||