▲ | cayley_graph a day ago | |||||||
Wow, my first ever video on AI! I'm rather disappointed. That was devoid of meaningful content save for the two minutes where they went over the Anthropic blog post on how LLMs (don't) do addition. Importantly, they didn't remotely approach what those other papers are about, or why thinking tokens aren't important for chain-of-thought. Is all AI content this kind of slop? Sorry, no offense to the above comment, it was just a total waste of 10 minutes that I'm not used to. So, to anyone more knowledgeable than the proprietor of that channel: can you outline why it's possible to replace thinking tokens with garbage without a decline in output quality? edit: Section J of the first paper seems to offer some succint explanations. | ||||||||
▲ | layoric 20 hours ago | parent [-] | |||||||
The video is just an entertaining overview, as indicated, I'm not the author of the video, it wasn't meant to be a deep dive. I linked the three related papers directly in there. I don't know how much more you are expecting from a HN comment, but this was a point in the right direction, not the definitive guide on the matter. This is a you problem. | ||||||||
|