Remix.run Logo
jacekm 7 days ago

What could be the benefit of paying $20 to Ollama to run inferior models instead of paying the same amount of money to e.g. OpenAI for access to sota models?

daft_pink 7 days ago | parent | next [-]

I feel the primary benefit of this Ollama Turbo is that you can quickly test and run different models in the cloud that you could run locally if you had the correct hardware.

This allows you to try out some open models and better assess if you could buy a dgx box or Mac Studio with a lot of unified memory and build out what you want to do locally without actually investing in very expensive hardware.

Certain applications require good privacy control and on-prem and local are something certain financial/medical/law developers want. This allows you to build something and test it on non-private data and then drop in real local hardware later in the process.

jerieljan 7 days ago | parent | next [-]

> quickly test and run different models in the cloud that you could run locally if you had the correct hardware.

I feel like they're competing against Hugging Face or even Colaboratory then if this is the case.

And for cases that require strict privacy control, I don't think I'd run it on emergent models or if I really have to, I would prefer doing so on an existing cloud setup already that has the necessary trust / compliance barriers addressed. (does Ollama Turbo even have their Trust center up?)

I can see its potential once it gets rolling, since there's a lot of ollama installations out there.

fluidcruft 7 days ago | parent | prev | next [-]

Me at home: $20/mo while I wait for a card that can run this or dgx box? Decisions, decisions.

dawnerd 7 days ago | parent | prev [-]

Quickly test… the two models they support? This is just another subscription to quantized models.

daft_pink 6 days ago | parent [-]

it looks like the plan is to support way more models though. gotta start somewhere.

rapind 7 days ago | parent | prev | next [-]

I'm not sure the major models will remain at $20. Regardless, I support any and all efforts to keep the space crowded and competitive.

adrr 7 days ago | parent | prev | next [-]

Running models without a filter on it. OpenAI has an overzealous filter and won’t even tell you what you violated. So you have to do a dance with prompts to see if it’s copyright, trademark or whatever. Recently it just refused to answer my questions and said it wasn’t true that a civil servant would get fired for releasing a report per their job duties. Another dance sending it links to stories that it was true so it could answer my question. I want a LLMs without training wheels.

michelsedgh 7 days ago | parent | prev | next [-]

I think its the data privacy is the main point and probably more usage before you hit limits? But mainly data privacy i guess

7 days ago | parent | prev | next [-]
[deleted]
ibejoeb 7 days ago | parent | prev | next [-]

I run a lot of mundane jobs that work fine with less capable models, so I can see the potential benefit. It all depends on the limits though.

_--__--__ 7 days ago | parent | prev | next [-]

Groq seems to do okay with a similar service but I think their pricing is probably better.

woadwarrior01 6 days ago | parent | next [-]

Groq's moat is speed, using their custom hardware.

Geezus_42 7 days ago | parent | prev [-]

Yeah, the NAZI sex not will be great for business!

fredoliveira 7 days ago | parent | next [-]

Groq (the inference service) != Grok (xAI's model)

gabagool 7 days ago | parent | prev [-]

You are thinking of Elon Grok, not Groq

janalsncm 7 days ago | parent [-]

When Grok originally came out I thought it was unlucky on Groq’s part. Now that Grok has certain connotations, it’s even more true.

owebmaster 7 days ago | parent [-]

"There's no such thing as bad publicity." PT Barnum

AndroTux 7 days ago | parent | prev | next [-]

Privacy, I guess. But at this point it’s just believing that they won’t log your data.

vanillax 7 days ago | parent | prev [-]

nothing lmao. this is just ollama trying to make money.