Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

If you want to use Ollama to run local models, here’s a simple example:

from ollama import chat, ChatResponse

def call_llm(prompt, use_cache: bool = True, model="phi4") -> str: response: ChatResponse = chat( model=model, messages=[{ 'role': 'user', 'content': prompt, }] ) return response.message.content



Is the output as good?

I'd love the ability to run the LLM locally, as that would make it easier to run on non public code.


It's decent enough. But you'd probably have to use a model like llama2, which may set your GPU on fire.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: