| ▲ | solarkraft 15 hours ago | |
Looks like it: https://ollama.com/library/qwen3-vl:30b-a3b | ||
| ▲ | thot_experiment 5 hours ago | parent [-] | |
fwiw on my machine it is 1.5x faster to inference in llama.cpp, these the settings i use for inference for the qwen i just keep in vram permanently | ||