| ▲ | nzeid 2 hours ago | |
I'm using llama.cpp with Gemma and tool calling is mission critical. It's perfectly fine on my end. There are definitely differences in the eagerness to tool-call that you'll need to manage. And for all local models I've ever used, I've had to micromanage the tools provided by servers to eliminate any possibility that they reach for something wonky or confusing. | ||