| ▲ | fultonn 17 hours ago | |
> Edit: Hey which of the models on that page were you referring to? I was referring to the smaller ones -- `granite4:micro`, `granite4-latest`, `granite4:350m`. > I'm grabbing one now that's apparently double digit GB? You are probably downloading one of these two ids: `granite4:small-h` or `granite4:32b-a9b-h`. The "small" model _is_ small in relative terms, but is also the largest of the currently released granite models! At 32B parameters (19GB download) it's runnable locally but not in the same "run on your laptop with acceptable performance" category of the nano/micro models. > Also my "dev enviornment" is vi -- I come from infosec (so basically a glorified sysadmin) so I'm mostly making little bash and python scripts, so I'm learning a lot of new things about software engineering as I explore this space :-) Shameless plug: if you're writing Python scripts to automate things using small locally hosted models, consider trying out https://github.com/generative-computing/mellea Mellea tries to nudge toward good software engineering practices -- breaking down big tasks into smaller parts, checking outputs after nondeterministic steps, thinking in terms of data structures and invariants rather than flow charts, etc. We built it with "actual fully automated robust workflows" in mind. You can use it with big models or small models, but it really shines when used with small models.  | ||