| ▲ | adityashankar 6 hours ago | |
This depends on how much better the models will get from now in, if Claude Opus 4.6 was transformed into one of these chips and ran at a hypothetical 17k tokens/second, I'm sure that would be astounding, this depends on how much better claude Opus 5 would be compared to the current generation | ||
| ▲ | aurareturn 6 hours ago | parent | next [-] | |
I’m pretty sure they’d need a small data center to run a model the size of Opus. | ||
| ▲ | empath75 4 hours ago | parent | prev [-] | |
Even an O3 quality model at that speed would be incredible for a great many tasks. Not everything needs to be claude code. Imagine Apple fine tuning a mid tier reasoning model on personal assistant/MacOs/IOS sorts of tasks and burning a chip onto the mac studio motherboard. Could you run claude code on it? Probably not, would it be 1000x better than Siri? absolutely. | ||