| ▲ | mft_ 12 hours ago | |||||||
You're maybe missing the article's point, which is to use local models appropriately: > “But Local Models Aren’t As Smart” > Correct. > But also so what? > Most app features don’t need a model that can write Shakespeare, explain quantum mechanics, and pass the bar exam. They need a model that can do one of these reliably: summarize, classify, extract, rewrite, or normalize. > And for those tasks, local models can be truly excellent. | ||||||||
| ▲ | Galanwe 11 hours ago | parent | next [-] | |||||||
This is a bit naive IMHO... I have tried quite a bunch of local models, and the reality is that it's not just a matter of of "it's a small model that should be hostable easily". Its also a matter of whats your acceptable prefill TTFT and decode t/s. All the local models I used, on a _consumer grade_ server (32GB DDR5, AMD Ryzen) have been mostly unusable interactively (no use as coding agent decently possible), and even for things like classification, context size is immediatly an issue. I say that with 6m experience running various local models for classifying and summarizing my RSS feeds. Just offline summarizing ans tagging HN articles published on the front page barely make the queue sustainable and not growing continuously. | ||||||||
| ||||||||
| ▲ | mikrl 11 hours ago | parent | prev [-] | |||||||
One of my hobbyist workflows involved transcribing ETF prospecti into yaml for an optimizer to optimize over. Used to take me maybe 10-20 minutes per sheet. Then I got codex to whip up a script that sends each sheet to a fairly low parameter locally running LLM and I have the yaml in a couple seconds. My dream is to bootstrap myself to local productivity with providers… I know I’ll never get there because hedonic treadmill etc, but I do feel there’s lots more juice to squeeze. I just need to invest more time into AI engineering… | ||||||||