Remix.run Logo
__mharrison__ 5 hours ago

For the past month, I've been claiming that $20/mo codex is the best deal in AI.

Now I'm going to have to find the new best deal.

scosman 4 hours ago | parent | next [-]

Check out z.ai coder plan. The $27/mo plan is roughly the same usage as the 20x $200 Claude plan. I have both and Claude is a little better, but GLM 5.1 is much better value.

rustyhancock 4 hours ago | parent | next [-]

Agreed, I use Z.ai and the usage is fantastic the only temper that recommendation that it's often unreliable. Perhaps a few times per week it's unresponsive. Maybe more often it seems to become flakey.

It's very variable though recently I'm noticing it's more reliable but there was a patch where it was nearly unusable some days.

I guess I won't complain for the price and YMMV.

scosman 3 hours ago | parent [-]

Agreed. They had a rough patch around the 4.7 to 5 upgrade. New architecture required hardware migration. The 5 to 5.1 upgrade was much smoother (same architecture new weights). As you say, little rough around edges, but still great value. Trick I learned is that it's max 2 parallel requests per user. You can put a billion tokens a month through it, but need to manage your parallelism.

mickeyp 3 hours ago | parent | prev [-]

If you're ok with a model provider that goes down all the time and has such a poor inference engine setup that once you get past 50k tokens you're going to get stuck in endless reasoning loops.

muyuu an hour ago | parent | prev | next [-]

What has actually changed? It's unclear how much can you do right now, unless they've already switched you to the new plan and you're speaking from experience.

piyh 4 hours ago | parent | prev | next [-]

Already paying for Google photo storage, AI pro for an extra $7 is a steal with anti-gravity.

cmrdporcupine 3 hours ago | parent | next [-]

I bought one of the google AI packages that came with a pile of drive storage and Gemini access.

Unfortunately gemini as a coding agent is a steaming useless pile. They have no right selling it, cheap open weight Chinese models are better at this point.

It's not stupid it just is incompetent at tool use and makes bad mistakes. It constantly gets itself into weird dysfunctional loops when doing basic things like editing files.

I'm not sure what GOOG employees are using internally, but I hope they're not being saddled with Gemini 3.1. It's miles behind.

qingcharles 2 hours ago | parent | next [-]

Gemini 3.1 is a good coding agent. We've been totally spoiled now. Also, if you use Antigravity you can burn up Opus 4.6 credits off your Goog account instead, before you have to switch to Gem 3.1.

surajrmal 2 hours ago | parent | prev [-]

Are you using gemini CLI or antigravity? The former is not really comparable to the latter in terms of quality. I wouldn't say antigravity is as good as the competition but it's pretty close. Miles behind is overstating it.

cmrdporcupine 2 hours ago | parent [-]

Gemini CLI but also used the Gemini models via opencode. They're terrible at CLI tool use. Like I said, just editing text files, they fall over rapidly, constantly making mistakes and then mistakes fixing their mistakes.

Antigravity wants me to switch IDEs, and I'm not going to do that.

matt_heimer 4 hours ago | parent | prev | next [-]

That's only good for the web based UI. If you want Gemini API access which is what this article is about then you must go the AIStudio route and pricing is API usage based. It does have a free usage tier and new signups can get $300 in free credits for the paid tier so it's I think it's still a good deal, just not as good as using the subscriptions would be.

spijdar 4 hours ago | parent [-]

No? Isn't the article about Codex, which is roughly equivalent to "Gemini CLI" and Google's Antigravity? Google's subscriptions include quotas for both of those, albeit the $20 monthly "Pro" plan has had its "Pro" model quota slashed in the last few weeks. You still get a large number of "Gemini 3 Flash" queries, which has been good enough for the projects I've toyed with in Antigravity.

matt_heimer 3 hours ago | parent | next [-]

I guess that's true but I find Google's models better than their public tooling. The Pro subscription includes "Gemini Code Assist and Gemini CLI" but the Gemini Code Assist plugin for IntelliJ which is my daily driver is broken most of the time to the degree that it's completely unusable. Sometimes you can't even type in the input box.

The only way I can do serious development with Gemini models is with other tooling (Cline, etc) that requires API based access which isn't available as part of the subscription.

bethekind 2 hours ago | parent [-]

I agree. Gemini models are held back by their segmentation of usage between multiple products, combined with their awful harnesses and tooling. Gemini cli, antigravity, Gemini code assist, Jules.... The list goes on. Each of these products has only a small limit and they must share usage.

It gets worse than that though. Most harnesses that are made to handle codex and Claude cannot handle Gemini 3.1 correctly. Google has trained Gemini 3.1 to return different json keys than most harnesses expect resulting in awful results and failure. (Based on me perusing multiple harness GitHub issues after Gemini 3.1 came out)

operatingthetan 3 hours ago | parent | prev [-]

Google is by far the best deal for AI, they give you so many 'buckets' of usage for a variety of products, and they seem to keep adding them.

kingstnap 3 hours ago | parent [-]

If you aggressively use all buckets Google is incredibly generous. In theory for one AI pro subscription you can get what is a ridiculous return in investment in a family plan.

You could probably be charging google literally thousands if all 6 members were spamming video and image generation and antigravity.

operatingthetan 3 hours ago | parent [-]

The family sharing is the real hack lol. I don't think any other provider does that.

purrcat259 4 hours ago | parent | prev [-]

Good luck sticking within limits, I have been burning up my baseline limits insanely fast within a few prompts, a marked change from a few weeks ago.

There's a few complaints online about the same happening to multiple users.

Otherwise anti-gravity has been great.

lelanthran 2 hours ago | parent [-]

I use the free Chat AIs all the time; Claude, ChatGPT, Gemini, Grok, Mistral.

In the last month they have all clamped down quite heavily. I use to be able to deep-dive into a subject, or fix a small Python project, multiple times per day on the free Web UIs.

Claude, this morning, modified a small Python project for me and that single act exhausted all my free usage for the day. In the past I could do multiple projects per day without issue.

Same with ChatGPT. Gemini at least doesn't go full on "You can use this again at 1100AM", but it does fallback to a model that works very poorly.

Grok and Mistral I don't really use that much, but Grok's coding isn't that bad. The problem is that it is not such a good application for deep-diving a topic, because it will perform a web search before answering anything, making it take long.

Mistral tends to run out of steam very quickly in a conversation. Never tried code on it though.

aulin 3 hours ago | parent | prev | next [-]

GH Copilot is still the best deal, while it lasts

hokkos 7 minutes ago | parent | next [-]

I feel they will go token base at some point, currently if you only use it with precise prompts and not random suggestions, switch between models 5.4 and 5.4 mini depending on the work, it is the best deal.

__mharrison__ an hour ago | parent | prev [-]

Yeah, it's really good. Probably going to be the next best deal until they cut back.

I need to try the command line version.

verdverm 4 hours ago | parent | prev [-]

We are exiting a hype cycle, well into the adoption curve. Subscriptions were never going to last.

My next step is going to be evaluating open and local models to see if they are sufficiently close to par with frontier models.

My hope is that the end of seat based pricing comes with this tech cycle. I was looking for document signing provider that doesn't charge a monthly, I only need a few docs a year.

alifeinbinary 4 hours ago | parent | next [-]

I'm developing software in this area right now, so I try a lot of the new models. They're not even close for coding tasks. It basically comes down to 26b parameters vs 1T parameters / quantisation / smaller context sizs, there's no comparison. However, for agentic work, tool calling, text summarisation, local LLMs can be quite capable. Workloads that run as background tasks where you're not concerned about TTFB, cold starts, tok/s etc., this is where local AI is useful.

If you have an M processor then I would recommend that you ditch Ollama because it performs slowly. We get double or triple tok/s using omlx or vmlx, respectively, but vmlx doesn't have extensive support for some models like gpt-oss.

AstroBen 4 hours ago | parent | next [-]

Kimi K2.5 (as an example) is an open model with 1T params. I don't see a reason it has to be local for most use cases- the fact that it's open is what's important.

verdverm 2 hours ago | parent | prev [-]

first session with gemma4:31b looks pretty good, like it may actually be up to coding tasks like gemini-3-flash levels

you can tell gemma4 comes from gemini-3

__mharrison__ 4 hours ago | parent | prev [-]

I recently experimented creating a Python library from scratch with Codex. After I was done, I took the PRD and Task list that was generated and fed them to opencode with Qwen 3.5 running locally.

Opencode was able to create the library as well. It just took about 2x longer.

selectodude 4 hours ago | parent [-]

Which version of Qwen 3.5 did you use?

verdverm 4 hours ago | parent [-]

which quant as well

__mharrison__ an hour ago | parent [-]

Not at my computer now, either 27 or 35b not quantized.

Next week I will be trying qwopus 27b.