| ▲ | huydotnet 2 hours ago | |||||||
I was hoping for the /v1/messages endpoint to use with Claude Code without any extra proxies :( | ||||||||
| ▲ | anonym29 2 hours ago | parent [-] | |||||||
This is a breeze to do with llama.cpp, which has had Anthropic responses API support for over a month now. On your inference machine:
Obviously, feel free to change your port, context size, flash attention, other params, etc.Then, on the system you're running Claude Code on:
Note that the auth token can be whatever value you want, but it does need to be set, otherwise a fresh CC install will still prompt you to login / auth with Anthropic or Vertex/Azure/whatever. | ||||||||
| ||||||||