| ▲ | sofixa 11 hours ago | |
> I mean, the most cutting edge of iPhones, iPads and MacBook Pros _today_ are quite capable of running in realtime today’s high-end local LLMs Definitely not the high end local LLMs. The small ones, yes, absolutely. > If you project out that hardware just a couple of years One of the biggest bottlenecks for LLMs is memory capacity and bandwidth. With the current glut for memory, it's unlikely we'll see lots of advancements in terms of average memory available or its bandwidth on regular (not super high end devices) in the coming years. Alternatively, it's possible we get dedicated SMLs for e.g. phone specific use cases, that are optimised and run well. | ||