Remix.run Logo
Sohcahtoa82 5 hours ago

> AI is fine. The hype is annoying.

I'm finding the detractors worse than the hype, because it seems like a certain subset of detractors [0] formed their opinion on AI in late 2022/early 2023 when ChatGPT came out (REALLY!? Over 3 years ago!?) and then never updated their opinions since then. They'll say things like "why would I want to consume X amount of energy and Y amount of water just to get a wrong answer?"

In other words, the people who think generative AI is an absolutely worthless and useless product are more annoying than the ones that think it's going to solve all the world's problems. They have no idea how much AI has improved since it reached center stage 3 years ago. Hallucinations are exceptionally rare now, since they now rely on searching for answers rather than what was in its training data.

We got Claude Desktop at work and it's been a godsend. It works so much better to find information from Confluence and present it to me in a digestible format than having to search by hand and combing through a dozen irrelevant results to find the one bit of information I need.

[0] For the purpose of this comment, this subset is meant to be detraction based on the quality of the product, not the other criticisms like copyright/content theft concerns, water/energy usage, whether or not Sam Altman is a good person, etc.

onemoresoop 4 hours ago | parent | next [-]

Follow closely on what the detractors say. Most of them are using AI themselves and are just pushing back on the hype or other ludicrous claims and that's a good thing. Is the current crop of Gen AI anything near AGI? Is it worth the current valuation? Can a company fire most staff and run on gen AI? We may see the economy completely crash and not because AI takes over but because of bad investments, hype and greed.

beej71 4 hours ago | parent | prev | next [-]

I don't think it's worthless. It can greatly speed up coding. And learning foreign languages. And many other things.

But I do think humanity is worse off because of it. So I'm a detractor in that way. :)

ben_w 2 hours ago | parent | prev | next [-]

> Hallucinations are exceptionally rare now, since they now rely on searching for answers rather than what was in its training data.

Well, I wouldn't go that far, but the hallucinations have moved up to being about more complicated things than they used to be.

Also, I've seen a few recent ones that "think" (for lack of a better word) that they know enough about politics to "know" they don't need to search for current events to, for example, answer a question about the consequences of the White House threatening military action to take Greenland. (The AI replied with something like "It is completely inconceivable that the US would ever do this").

doug_durham 4 hours ago | parent | prev | next [-]

On reddit there are two sub-Reddits that are mirrors, /accelerate and /betteroffline. The people in the subs go there for dopamine hits. One for how AI is going to transform their lives and lead to a work-free future. The other how AI is worthless and how everyone (except them) is being fooled. They are the same people with opposite views. The people in either sub don't recognize this.

jarjoura 4 hours ago | parent | prev | next [-]

You do realize though that using Claude Desktop to "search" through confluence is like paying a world class architect on the hour just to give you some tips on how to layout your small loft to maximize sunlight.

This is such a perfect example of the mania behind this rollout.

There's no way you can make the financials work here compared to JetBrains spending the same millions spent on AI infrastructure and instead building better search in Confluence. Confluence search SUCKS, but that's just a lack of focus (or resources) on building a more complex, more robust solution. It's a wiki.

Either way, making a more robust search is a one time cost that benefits everyone. Instead, you're running a piece of software that goes directly to Anthropic's bank account, and to the data centers and to hyper scalers. Every single query must be re-run from scratch, costing your company a fortune, that if not managed properly will come out of spending that money elsewhere.

xboxnolifes 4 hours ago | parent | next [-]

> You do realize though that using Claude Desktop to "search" through confluence is like paying a world class architect on the hour just to give you some tips on how to layout your small loft to maximize sunlight.

If I could pay a world class architect $1.50 to give me tips on how to maximize sunlight in my loft I would.

Would it be nice if confluence just had a robust search that had a one time cost and then benefited everyone thereafter? Sure, but that's not the current reality, and I do not have control over their actions. I can only control mine.

lukevp 4 hours ago | parent | prev [-]

And what is using Confluence in the first place? Your MacBook Pro is faster than a supercomputer from 20 years ago. As we make compute cheaper, we find ways to use it that are less efficient in an absolute sense but more efficient for the end user. A graphical docs portal like Confluence is a hell of a lot easier to use than EMacs and SSH to edit plain text files on an 80 character terminal. But it uses thousands of times more compute.

It seems ridiculous right now because we don’t have hardware to accelerate the LLMs, but in 5 years this will be trivial to run.

jarjoura an hour ago | parent [-]

I'm confused by your analogy. A wiki run server is extremely efficient to run, and can be hosted from a tiny little raspberry pie. A search engine can be optimized to provide results near O(1). You can even pull up and read results on a very old computer. All of the concerns around cost and resource efficiency can be addressed as all of this is a solved problem.

Even with an LLM agent getting cheaper to run in the future, it's still fundamentally non-deterministic so the ongoing cost for a single exploration query run can never get anywhere near as cheap as running a wiki with a proper search engine.

arcxi 4 hours ago | parent | prev | next [-]

This very comment is measurably more harmful than any AI criticism that annoys you - someone will read this and assume it's appropriate to accept whatever bullshit Claude generates at face value, with terrible consequences.

In contrast, what harm do those detractors cause? They don't generate as much code per hour?

xvector 4 hours ago | parent [-]

By that logic we should all live in air-filtered bubbles. Anyone denying this is causing harm. After all, people might die if you let them out of their air-filtered bubble!

The "harm" (if you can call it that) is clear, detractors slow the pace of progress with meaningless and incorrect hand-wringing. A lack of progress harms everyone (as evidenced our amazing QoL today compared to any historical lens.)

arcxi 4 hours ago | parent | next [-]

> detractors slow the pace of progress

Considering our climate, political and economic situation, I'd say not only is slowing the pace of progress not harmful, it's actually imperative for our long-term survival.

dijit 4 hours ago | parent | prev | next [-]

that’s a stretch and taking a measured approach to change is valid

slopinthebag 4 hours ago | parent | prev [-]

That's a pretty poor straw man - the issue is the amount of harm caused, not that there is a potential for some minuscule amount.

Also we need detractors because if we race into any technological advance too quickly we may cause unnecessary harm. Not all progress is without harms, and we need to be responsible about implementing it as risk-free as possible.

heavyset_go 4 hours ago | parent | prev | next [-]

> certain subset of detractors [0] formed their opinion on AI in late 2022/early 2023 when ChatGPT came out (REALLY!? Over 3 years ago!?)

I mean, you can get mad at people you made up in your head, that's a thing people do, but this caricature falls in the same comforting bucket as "anyone who doesn't like <thing I like> is just ignorant/stupid" and "if you don't like me you're just jealous".

Maybe non-straw people have criticisms that aren't all butterflies and rainbows for good reasons, but you won't get to engage with them honestly and critically if you're telling yourself they're just ignorant from the start.

For example, I will bet that non-straw people will take issue with this, and for good reasons:

> Hallucinations are exceptionally rare now

jackie293746 4 hours ago | parent | prev | next [-]

Claude Opus 4.6 regularly makes up shit and hallucinates. I'm not a detractor by any means but "exceptionally rare" is fantasyland.

thrawa8387336 4 hours ago | parent | next [-]

Can vouch for this, plus, when it does work, stuff can take forever. Then, if I let it unsupervised, higher risk of doing the wrong thing. If I supervise it, then I become agent nanny.

surgical_fire 4 hours ago | parent | prev [-]

I have been experiencing it too.

I honestly am finding Codex considerably better, as much as I despise OpenAI.

lovasoa 4 hours ago | parent | prev | next [-]

I use the latest codex with gpt5.4 and Claude opus every day. they hallucinate every day. If you think they don't, you are probably being gaslighted by the models.

Forgeties79 4 hours ago | parent | prev | next [-]

This is going to sound flippant, but truly, I imagine most people find the group that disagrees with their take annoying as well.

SpicyLemonZest 4 hours ago | parent | prev | next [-]

I personally believe that LLMs have advanced immeasurably since ChatGPT came out, which was itself a world-historical event. I use AI daily in ways that enhance my productivity.

I say all of that to establish that I'm not a reflexive critic when I tell you, hallucinations are absolutely not exceptionally rare now. On multiple occasions this week (and it's only Tuesday!) I've had to disprove a LLM hallucination at work. They're just not as fun to talk about anymore, both because they're no longer new and because straightforward guardrails are effective at blocking the funny ones.

bigstrat2003 4 hours ago | parent | prev | next [-]

> a certain subset of detractors [0] formed their opinion on AI in late 2022/early 2023 when ChatGPT came out (REALLY!? Over 3 years ago!?) and then never updated their opinions since then.

On the contrary. I update my opinion all the time, but every time I try the latest LLM it still sucks just as much. That is why it sounds like my opinion hasn't changed.

surgical_fire 4 hours ago | parent | prev | next [-]

The detractors are a lot less numerous and certainly a lot less preachy than the ones on the hype train.

AI is alright. It's moderately useful, in certain contexts it speeds me up a lot, in other contexts not so much.

I also think that the economics of it make no sense and that it is, generally, a destructive technology. But it's not up to me to fix anything, I just try to keep on top of best practices while I need to pay bills.

The economics bit is not my problem though. If all AI companies go bust and AI services disappear I can 100% manage without it.

heavyset_go 4 hours ago | parent | next [-]

> The economics bit is not my problem though. If all AI companies go bust and AI services disappear I can 100% manage without it.

We're in "too big to fail" territory, if we handled the recession we were heading towards/in years ago, instead of letting AI hype distract and redirect massive amounts of investment, attention and labor from elsewhere, we might have been in a better position.

jarjoura 4 hours ago | parent | prev [-]

On the flip side, if all this slop is floating around, and AI services do become untenable, think of all the immediate jobs that will open up to fix and maintain all the slop that's being thrown around right now. The millions of dollars of contracts spent to use these LLMs will be redirected back to hiring.

Though, my cynical take is that the investor class seemed dead-set on forcing us all to weave LLMs deep into our corporate infrastructures in a way that I'm not too sure it will ever "disappear" now. It'll cost just as much to detangle it as it was to adopt it.

teaearlgraycold 4 hours ago | parent | prev [-]

> Hallucinations are exceptionally rare now

The way we talk about "hallucinations" is extremely unproductive. Everything an LLM outputs is a hallucination. Just like how human perception is hallucination. These days I pretty much only hear this word come up among people that are ignorant of how LLMs work or what they're used for.

I've been asked why LLMs hallucinate. As if omniscient computer programs are some achievable goal and we just need to hammer out a few kinks to make our current crop of english-speaking computers perfect.