| ▲ | Show HN: Timber – Ollama for classical ML models, 336x faster than Python(github.com) | |||||||
| 50 points by kossisoroyce 3 hours ago | 7 comments | ||||||||
| ▲ | tl2do 31 minutes ago | parent | next [-] | |||||||
Since generative AI exploded, it's all anyone talks about. But traditional ML still covers a vast space in real-world production systems. I don't need this tool right now, but glad to see work in this area. | ||||||||
| ▲ | rudhdb773b 11 minutes ago | parent | prev | next [-] | |||||||
If the focus is performance, why use a separate process and have to deal with data serialization overhead? Why not a typical shared library that can be loaded in python, R, Julia, etc., and run on large data sets without even a memory copy? | ||||||||
| ||||||||
| ▲ | brokensegue 21 minutes ago | parent | prev | next [-] | |||||||
"classical ML" models typically have a more narrow range of applicability. in my mind the value of ollama is that you can easily download and swap-out different models with the same API. many of the models will be roughly interchangeable with tradeoffs you can compute. if you're working on a fraud problem an open-source fraud model will probably be useless (if it even could exist). and if you own the entire training to inference pipeline i'm not sure what this offers? i guess you can easily swap the backends? maybe for ensembling? | ||||||||
| ▲ | mehdibl 44 minutes ago | parent | prev | next [-] | |||||||
Ollama is quite a bad example here. Despite popular, it's a simple wrapper and more and more pushed by the app it wraps llama.cpp. Don't understand here the parallel. | ||||||||
| ▲ | Dansvidania an hour ago | parent | prev | next [-] | |||||||
Can’t check it out yet, but the concept alone sounds great. Thank you for sharing. | ||||||||
| ▲ | jnstrdm05 2 hours ago | parent | prev [-] | |||||||
I have been waiting for this! Nice | ||||||||