| ▲ | Galanwe 12 hours ago | ||||||||||||||||||||||
I would love for local inference to be possible, but from my experience, Kimi 2.6 is the only model that would be worth it, and its a $10k (M3 Ultra max spec'd - 30s TTFT so kind of slowish) to $30k (RTX6000/700GB+ DDR5) upfront, noise / power consumption aside. | |||||||||||||||||||||||
| ▲ | mft_ 12 hours ago | parent [-] | ||||||||||||||||||||||
You're maybe missing the article's point, which is to use local models appropriately: > “But Local Models Aren’t As Smart” > Correct. > But also so what? > Most app features don’t need a model that can write Shakespeare, explain quantum mechanics, and pass the bar exam. They need a model that can do one of these reliably: summarize, classify, extract, rewrite, or normalize. > And for those tasks, local models can be truly excellent. | |||||||||||||||||||||||
| |||||||||||||||||||||||