| ▲ | integralid 8 hours ago | |||||||||||||||||||||||||||||||||||||||||||
I also instinctively reacted to that fragment, but at this point I think this is overreacting to a single expression. It's not just a normal thing to say in English, it's something people have been saying for a long time before LLMs existed. | ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | nvme0n1p1 7 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||
There are tells all over the page: > Redefining AI efficiency with extreme compression "Redefine" is a favorite word of AI. Honestly no need to read further. > the key-value cache, a high-speed "digital cheat sheet" that stores frequently used information under simple labels No competent engineer would describe a cache as a "cheat sheet". Cheat sheets are static, but caches dynamically update during execution. Students don't rewrite their cheat sheets during the test, do they? LLMs love their inaccurate metaphors. > QJL: The zero-overhead, 1-bit trick > It reduces each resulting vector number to a single sign bit (+1 or -1). This algorithm essentially creates a high-speed shorthand that requires zero memory overhead. Why does it keep emphasizing zero overhead? Why is storing a single bit a "trick?" Either there's currently an epidemic of algorithms that use more than one bit to store a bit, or the AI is shoving in extra plausible-sounding words to pad things out. You decide which is more likely. It's 1:30am and I can't sleep, and I still regret wasting my time on this slop. | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | g-mork 5 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||
Another instinctual reaction here. This specific formulation pops out of AI all the time, there might as well have been an emdash in the title | ||||||||||||||||||||||||||||||||||||||||||||