▲ | fforflo 5 days ago | |||||||
If you want to use Ollama to run local models, here’s a simple example: from ollama import chat, ChatResponse def call_llm(prompt, use_cache: bool = True, model="phi4") -> str: response: ChatResponse = chat( model=model, messages=[{ 'role': 'user', 'content': prompt, }] ) return response.message.content | ||||||||
▲ | mooreds 5 days ago | parent [-] | |||||||
Is the output as good? I'd love the ability to run the LLM locally, as that would make it easier to run on non public code. | ||||||||
|