| ▲ | james2doyle 3 days ago | |||||||||||||||||||||||||
I was playing with the new IBM Granite models. They are quick/small and they do seem accurate. You can even try them online in the browser because they are small enough to be loaded via the filesystem: https://huggingface.co/spaces/ibm-granite/Granite-4.0-Nano-W... Not only are they a lot more recent than gemma, they seem really good at tool calling, so probably good for coding tools. I haven’t personally tried it myself for that. The actual page is here: https://huggingface.co/ibm-granite/granite-4.0-h-1b  | ||||||||||||||||||||||||||
| ▲ | firefax 3 days ago | parent | next [-] | |||||||||||||||||||||||||
Interesting. Is there a way to load this into Ollama? Doing things in browser is a cool flex, but my interest is specifically in privacy respecting LLMs -- my goal is to run the most powerful one I can on my personal machine, with the end goal being those little queries I used to send to "the cloud" can be done offline, privately.  | ||||||||||||||||||||||||||
  | ||||||||||||||||||||||||||
| ▲ | brendoelfrendo 3 days ago | parent | prev [-] | |||||||||||||||||||||||||
Not the person you replied to, but thanks for this recommendation. These look neat! I'm definitely going to give them a try.  | ||||||||||||||||||||||||||