| ▲ | borsch_not_soup 9 hours ago | ||||||||||||||||
Interesting, I’ve always thought neural network progress was primarily bottlenecked by compute. If it turns out that LLM-like models can produce genuinely useful outputs on something as constrained as a Commodore 64—or even more convincingly, if someone manages to train a capable model within the limits of hardware from that era—it would suggest we may have left a lot of progress on the table. Not just in terms of efficiency, but in how we framed the problem space for decades. | |||||||||||||||||
| ▲ | dpe82 9 hours ago | parent | next [-] | ||||||||||||||||
60s per token for that doesn't strike me as genuinely useful.Very, very cool project though! | |||||||||||||||||
| |||||||||||||||||
| ▲ | numpad0 5 hours ago | parent | prev [-] | ||||||||||||||||
Next-word prediction features always existed for flip phones... | |||||||||||||||||