Remix.run Logo
vunderba 2 days ago

Strongly agree. Gemma3:27b and Qwen3-vl:30b-a3b are among my favorite local LLMs and handle the vast majority of translation, classification, and categorization work that I throw at them.

curioussquirrel 2 hours ago | parent | next [-]

Give Gemma 31B a shot for translation, it does a very good job at that given its size.

misiti3780 2 days ago | parent | prev [-]

what HW are you running them on ? are you using OLLAMA ?

vunderba 2 days ago | parent [-]

I'm using the default llama-server that is part of Gerganov's LLM inference system running on a headless machine with an nVidia 16GB GPU, but Ollama's a bit easier to ease into since they have a preset model library.

https://github.com/ggml-org/llama.cpp