Remix.run Logo
te0006 2 hours ago

Interesting - do you need to take any special measures to get OSS genAI models to work on this architecture? Can you use inference engines like Ollama and vLLM off-the-shelf (as Docker containers) there, with just the Radeon 8060S GPU? What token rates do you achieve?

(edit: corrected mistake w.r.t. the system's GPU)

buyucu 30 minutes ago | parent [-]

I just use llama.cpp. It worked out of the box.