Remix.run Logo
its_down_again 3 days ago

Could you explain more on how to do this? e.g if I am using the Claude API in my service, how would you suggest I go about setting up and controlling my own inference endpoint?

handfuloflight 3 days ago | parent | next [-]

You can't. He means by using the open source models.

datavirtue 3 days ago | parent | prev [-]

Runa local LLM tuned for coding on LM Studio. It has a server and provides endpoints.