Remix.run Logo
MrCharismatist 8 hours ago

Like six months ago we got a presentation from an AWS guy on the AI tooling available and how it fit with our particular use cases.

At one point seemingly out of nowhere he pointed out on his screen share "Look at how many tokens I've used this month. I run so much Opus." It was a number that was offensively large.

I remember thinking "That's a really odd flex, this crap is so expensive the fact that you use so much should be a red flag"

He demonstrated a number of Claude Code use cases he had to manage and tweak AWS infrastructure that made me, the old greybeard sysadmin older than the internet think "You've used AI to do something that was a single command."

So this story makes sense. They were being encouraged to just blast away at it six plus months ago.

funimpoded 7 hours ago | parent | next [-]

I notice a lot of Cursor's suggestions are just stuff a linter should auto-fix.

But if you hit "tab" it'll claim that as an AI-edited line, LOL.

(A lot of the rest of it is stuff I could already have been doing just as fast if I'd ever bothered to learn to use multiple cursors, learned vim navigation, or set up some macros—I never did because my getting-code-on-the-screen speed without those has never been slow enough to hold anything up, in practice)

jjice 7 hours ago | parent | next [-]

Cursor absolutely tries to maximize what they claim is "AI-edited" and it's nonsense a lot of the time. If it writes a function and then I got in and edit that function, it claims my edits _and_ any net-new lines I add above or below the function.

giancarlostoro 2 hours ago | parent [-]

So their diff mechanism is poorly labeled (or purposely?) then.

pwillia7 an hour ago | parent | prev [-]

you don't use vim/emacs for the productivity. It's a lifestyle decision

mrbungie 6 hours ago | parent | prev | next [-]

I still don't know how to reconcile these reports with what other people say about GenAI-agentic assisted engineering being the only way of working nowadays, especially in startups.

Probably there is no dichotomy going on and it depends on multiple factors, but it seems so weird to see reports that are so different between each other.

binary0010 4 hours ago | parent | next [-]

It's not required for startups. But if you are building trashy, brittle products and your main metric is speed to market, and have the expectation of high failure chances (e.g. most yc startup batches) - then yes you have to do agentic eng.

If you are making extremely specific, high quality products over a long time window and your founders are deeply experienced in that field of engineering, then no, you don't need agentic engineering and probably want very little llm code in general (outside of some boilerplate, internal toolings, etc).

themafia an hour ago | parent | prev | next [-]

> I still don't know how to reconcile these reports

This is work related. So you can't expect everyone to have the same input demands or output expectations.

> Probably there is no dichotomy

It's literally staring you in the face.

SpicyLemonZest 4 hours ago | parent | prev | next [-]

I think GenAI-agentic assisted engineering is the only way of working nowadays, and it's the only way I personally have worked for months. I still think that an outright majority of presentations on AI tooling I've seen have been in the nonsensical "Look how many tokens I can burn" genre. Had to sit through one guy recently who explained why you need a complex agentic team with 6 different roles in order to ask Claude to investigate a bug, which you most definitely do not.

dranudin 3 hours ago | parent [-]

A guy at my company (very old company, we need to maintain our software for 30+ years) gave a presentation about how they used opus 4.6 to onboard people, like giving new team members access rights etc. Then another guy (even higher in management) proposed using a team of agents for that.. It's getting pretty wild

wahnfrieden 5 hours ago | parent | prev [-]

Wage workers are evaluated on behaviors, founders are evaluated on growth and revenue. Of course usage patterns and outcomes will be different

throwway120385 7 hours ago | parent | prev | next [-]

I think you'll find that a lot of big investment companies are buried to the hilt in a lot of tech companies and also OpenAI and Anthropic. So you can do the math on where the directive is coming from and why it's not particularly careful or measured.

Henchman21 4 hours ago | parent | prev | next [-]

> "You've used AI to do something that was a single command."

As time passes and the layers of abstraction pile up, later generations won't understand the underlying layers of the abstraction. This is a huge weakness in our systems development -- and a huge potential attack surface for adversaries.

cindyllm 4 hours ago | parent [-]

[dead]

Our_Benefactors 7 hours ago | parent | prev [-]

> You've used AI to do something that was a single command

Yes, and that’s a good thing! This is in fact where a lot of AI value lies. You dont need to know that command anymore - knowing the functional contract is now sufficient to perform the requisite work duties. This is huge!

funimpoded 6 hours ago | parent | next [-]

Not even joking that the main benefit I've seen from "AI" for editing code is that it lets me quickly do all the things I could already have been doing just as quickly if I'd ever bothered to learn to use my tools.

Of course I lose about as much time as I save to its fuck-ups, so I'd still have been better off learning to actually use a text editor properly. Though (as I mentioned in a another post) part of why I've never done that in 25ish years of writing code for pay is that my code-writing speed has never been too slow for any of the businesses I've worked in, i.e. other things move slowly enough it never mattered.

mrbungie 6 hours ago | parent | prev | next [-]

Once I learn a command that is both repeatable and useful, I prefer to either keep it in my mind or in my aliases. Thank you.

Our_Benefactors 4 hours ago | parent | next [-]

You can still do this! And AI will teach you that command far far faster than synthesizing it yourself.

mrbungie 4 hours ago | parent [-]

Yeah, I know AI is useful for that, that's why I said after I learn. Hopefully once.

mikojan 4 hours ago | parent | prev [-]

That's what Skills are for

:^)

mrbungie 4 hours ago | parent [-]

0 tokens per command >>> Hundreds of tokens per command

malfist 7 hours ago | parent | prev | next [-]

Is it? If the LLMs change broke something do you know enough to fix it?

Our_Benefactors 4 hours ago | parent [-]

The same question can be applied to work without AI, so this isn’t a meaningful criticism

eikenberry 3 hours ago | parent [-]

In one case you are using tools you understand, in the other you aren't. Seem different to me.

bluefirebrand 7 hours ago | parent | prev | next [-]

> You dont need to know that command anymore

I find it hard to read "You can do things without knowing things" as a positive improvement in work, society, life, anywhere

Our_Benefactors 4 hours ago | parent [-]

You are the worst kind of gatekeeper, then! A true reactionary who believes they are righteous for impeding others!

bluefirebrand 2 hours ago | parent | next [-]

I'm pretty sure you're being sarcastic. I hope so.

It's hard to tell anymore because I have encountered people who genuinely do seem to think that disliking AI is gatekeeping somehow

nukedindia 4 hours ago | parent | prev [-]

[dead]

MrCharismatist 5 hours ago | parent | prev | next [-]

Look, I feel for junior admins, I was one 35 years ago and the only reason I'm where I am today was because I had to learn the hard way, repeatedly and often.

I use the shit out of opencode to do things as a force multiplier, not as a way to keep me from knowing what its doing.

The point at which we're optimizing for "we don't need to know that anymore" is the point at which everything blows up, because agentic work is not fully deterministic, models hallucinate even simple things.

Blindly relying on your agent weapon of choice to just do the right thing because you didn't take the time to understand how the lego fits together is an actual problem.

thereisnospork 4 hours ago | parent [-]

Replace agent with 'direct report' and you've just described middle management. For better or worse, companies have always run on non deterministic tasks doled out by persons who barely understand the work.

paulhebert 3 hours ago | parent [-]

Honestly human employees feel closer to deterministic than LLMs.

I have a pretty good sense of the quality of work my coworkers output, where they tend to struggle, where they're talented, what level of review is required, what I should double check, etc.

By contrast LLMs are more like picking a contractor out of a hat. Even with good guardrails the quality and types of issues vary wildly prompt to prompt.

perrygeo 3 hours ago | parent | prev | next [-]

I watched people ask LLMs for linting/refactoring help, burning easily 5 minutes for something that could be completed deterministically, locally, in ms using any modern editor.

Quite frankly it was embrassing. We've had tools for static analysis for ages. Use them.

Someone with better knowledge could work 100x faster using 100x fewer resources. They did it the slow, expensive way but at least didn't have to think? Odd flex.

miyoji 7 hours ago | parent | prev | next [-]

It's also several hundred times more expensive.

Our_Benefactors 4 hours ago | parent [-]

False! Labor is the most expensive input in creating software, not joules of energy. Using AI is far far cheaper than expecting workers to synthesize all knowledge themselves.

miyoji 3 hours ago | parent [-]

Both workflows involve typing a very small number of characters and should take under 30 seconds. There's no difference in labor cost. However, the compute and energy costs of the tokens to solve the problem vs the tool call will be orders of magnitude in difference, even for trivial stuff like grepping. It gets worse as the problem gets more complicated and the tools more specialized.

quikoa 6 hours ago | parent | prev [-]

I can't tell if this comment is sarcasm or not. If you let AI run commands you don't understand (especially in production) you may end up with some nasty surprises.

xp84 4 hours ago | parent | next [-]

With a comment like that, it's no wonder you're dramatically below our minimum guidance for tokens consumed.

If AI breaks production this way, you just tell AI to fix it! And look, now you've consumed tokens twice. Think on that and I'll see you at the end-of-year performance review.

Our_Benefactors 4 hours ago | parent | prev [-]

It’s not sarcastic at all. Using AI to accomplish things and fill in your knowledge gaps is literally the whole point of it. People are downvoting and salty because they thought the value in the job was in memorizing esoteric APIs (it never was)