| ▲ | falense 3 hours ago | |
I disagree. There is in fact a non-zero chance that we will get good enough models that are MOE optimized for desktop size hardware that can do a lot of the same things as the SOTA models. Im certainly crossing my fingers that the open-weights models continue improving. Engram from Deepseek for instance seems very interesting for a compute to memory offloading perspective. https://www.reddit.com/r/LocalLLaMA/comments/1s0czc4/round_2... | ||