Remix.run Logo
Joeri 14 hours ago

I already switched to claude a while ago. Didn’t bring along any context, just switched subscriptions, walked away from chatgpt and haven’t touched it again. Turned out to be a non-event, there really is no moat.

I switched not because I thought Claude was better at doing the things I want. I switched because I have come to believe OpenAI are a bad actor and I do not want to support them in any way. I’m pretty sure they would allow AGI to be used for truly evil purposes, and the events of this week have only convinced me further.

kdheiwns 12 hours ago | parent | next [-]

Yesterday was my first time trying it. One thing that felt a bit strange to me was that I asked it something and the response was just one paragraph. Which isn't bad or anything but it felt... strange? Like I always need to preface ChatGPT/gemini/whatever question with "Briefly, what is..." or it gives me enough fluff to fill a 5 page high school essay. But I didn't need to do that and just got an answer that was to the point and without loads of shit that's barely related.

And the weirdest thing that I noticed: instead of skimming the response to try finding what was relevant, I just straight up read it. Kind of felt like I got a slight amount of focus ability back.

Accuracy is something I can't really compare yet (all chatbots feel generally the same for non-pro level queries), but so far, I'm fairly satisfied.

HarHarVeryFunny 4 hours ago | parent | next [-]

I use Gemini all the time, but I have to say it's got verbal diarrhea and an EXTREMELY annoying trait of wanting to lead the conversation rather than just responding to what YOU want to do. At the end of every response Gemini will always suggest a "next step", in effect trying to 2nd guess where you want the conversation to go. I'd much rather have an AI that just did what it was asked, and let me decide what to ask next (often nothing - maybe it was just a standalone question!).

Apparently this annoying "next step" behavior is driven by the system prompt, since the other day I was running Gemini 3 Thinking, and it was displaying it's thoughts which included a reminder to itself to check that it was maintaining a consistent persona, and to make sure that it had suggested a next step. I'd love to know the thought process of whoever at Google thought that this would make for a natural or useful conversation flow! Could you imagine trying to have a conversation with a human who insisted on doing this?!

edoceo 4 hours ago | parent | next [-]

Yes. That is a salesperson. The next-step is to drive engagement. I know you're not interested in $THiNG now but can I follow up in 3 months?

In persona I think the AIs are that Claude is the engineer and Gemini is the sales-person and GPT is the eager and loud journeyman.

vrosas 4 hours ago | parent | prev | next [-]

I don’t know, I’ve found the follow ups nice sometimes. You can just ignore them if they’re not useful. The computer won’t get mad…

HarHarVeryFunny 3 hours ago | parent [-]

I find they are extremely rarely useful - they just break my own chain of thought by having to constantly read and ignore this stuff. The same goes for the excessively verbose responses too - a human has a limited "context window" and a shorter response is therefore much more useful/valuable than a long token-maxxed one.

Sure the computer won't get mad, and that is all I do - just ignore what Gemini is suggesting and pretend it never said it - but it certainly makes me mad. The main reason I stick with Gemini is because of the generous free usage limits, but I know this annoying "next step" behavior (which was a relatively recent change) is going to push me back to Claude, even if I need to pay for it.

debo_ 3 hours ago | parent | prev [-]

Yes. Gemini is mimicking its creators: this is exactly how I experienced speaking to most Googlers.

layer8 9 hours ago | parent | prev | next [-]

One issue is that Claude’s web search abilities are more limited, for example it can’t search Reddit and Stack Overflow for relevant content.

bredren 5 hours ago | parent | next [-]

Why not just write a skill and script that calls crawl4ai or similar and do this using Claude code?

You can store the page as markdown for future sessions, mash the data w other context, you name it.

The web Claude is incredibly limited both in capability and workflow integration. Doesn’t matter if you’re dealing with bids from arbor contractors or researching solutions for a DB problem.

Barbing 4 hours ago | parent | next [-]

Want this w/o killing the free open web.

Maybe I run an old PC adjacent to the scraper to manually visit the scraped pages without an adblocker, & buy something I need from an ad periodically (while a cohesive response is being generated in the meantime)

Ya sounds dumb, wishing for a middle ground that lets us be effective but also good netizens. Maybe that Cloudflare plan to charge the bots…

layer8 4 hours ago | parent | prev [-]

See https://news.ycombinator.com/item?id=47208741.

samhclark 8 hours ago | parent | prev | next [-]

That's so frustrating with Claude. If I need to widely search the web or if I need it to read a specific URL I pasted, I always turn to ChatGPT. Claude seems to hit a lot more roadblocks while trying to navigate the web.

godelski 5 hours ago | parent | next [-]

The issue is Reddit though. They're the ones blocking. They're very aggressive.

When sites are working in one chatbot and not another, there's a good chance that the latter is respecting the website rules. As an example with Reddit, you're probably blocked when using a VPN like Mullvad

ronsor 6 hours ago | parent | prev [-]

They're playing too nice. It's time to roll out the residential proxies.

andai 6 hours ago | parent | prev | next [-]

It's not that hard to roll your own web search MCP.

I made one for Crush a while ago.

https://anduil.neocities.org/blog/?page=mcp

I'm not sure about the issues with reddit though? Do they block Claude's web fetch tool? I think Codex runs it thru some kind of cache proxy.

layer8 5 hours ago | parent [-]

Rolling your own is not the solution for the common case where you’re asking an LLM a question that may or may not be supported or supplemented by a web search. ChatGPT decides by itself when and how to consult the web, and then links the relevant sources in its result. You don’t get that functionality from Claude chat, you’d have to completely build your own chat harness and apps.

Sites like Reddit are blocking AI providers, they have to have some contract with them for access. OpenAI does seem to have that.

MrDarcy 5 hours ago | parent | prev [-]

That’s a feature not a bug.

Sharlin 9 hours ago | parent | prev | next [-]

Heh, a while ago I wondered why ChatGPT had started to reply tersely, almost laconically. Then I remembered that I had explicitly told it to be brief by default in the custom personality settings… I also noticed that there are now various sliders to control things like how many emojis or bulletpoint lists ChatGPT should use, which I though was amusing. Anyway, these tools can be customized to adopt just about any style, there's no need to always prefix questions with "Briefly" or similar.

andai 5 hours ago | parent | next [-]

Here's my prompt to make ChatGPT sound more like Claude.

It works but not as well as I'd like -- the tone and word choice still ends up being really jarring to me (even after years of using ChatGPT). Maybe that's promptable too. Open to suggestions.

---

Respond in a natural conversational style. In terms of language, match my own tone and style.

Keep responses to half a page or so max. (Use context and your judgment. e.g. for example, initial response can be a page, and then specific follow up questions can be shorter, if the question is answered clearly)

Prefer minimal formatting. Don't use headings, lists etc. Bold and italics OK but keep it tasteful.

If you're starting a paragraph like so

Item name: description..

then it makes sense to bold item name for readability purposes.

AgentOrange1234 3 hours ago | parent | prev [-]

Hah. I remember some story about a chatbot that had been trained on slack conversations. You would ask it for an essay on whatever, and it would say "will do" or "I'll have it for you tomorrow." :)

lkbm 5 hours ago | parent | prev | next [-]

Yeah, I've always been a little confused why people use ChatGPT so heavily. It's better than it used to be (maybe thanks to custom configuration), but it still tends to respond like it's writing a Wikipedia article.

Wikipedia articles on demand are great, but not usually what I want.

skeledrew 8 hours ago | parent | prev | next [-]

Yep the experience is quite something. Another thing I've noticed, and you likely soon will also, is that Claude only attempts a follow-up if the one is needed or the prompt is structured for it. Meanwhile ChatGPT always prompts you with a choice of next steps. It can be nice, as sometimes the options contain improvements you never thought of and would like, but in lengthy conversations with a detailed plan it does things really piecemeal, as though trained to maximize engagement instead of getting to a final solution.

zukzuk 8 hours ago | parent [-]

I find that Claude almost always ends its response with some sort of follow up question, despite my system prompt telling it not to.

I never really used ChatGPT much though so maybe Claude is just relatively less egregious?

esperent 10 hours ago | parent | prev | next [-]

> Which isn't bad or anything but it felt... strange?

On the contrary, it's great. It's fully capable of outputting a wall of text when required, so instead of feeling like I'm talking to something that has a minimum word count requirement, I get an appropriate sized response to the task at hand.

mavamaarten 11 hours ago | parent | prev [-]

In my limited experience, that's mostly since the 4.6 release. I noticed that with the same prompt, it answers much more briefly. A bit jarring indeed, but I prefer it. Less bs and filler, and less burning off electricity for nothing.

ACCount37 10 hours ago | parent | next [-]

This behavior first appeared in 4.5, mostly for specific types of questions and in "natural conversation" workflows. 4.6 might have pushed it further.

xmonkee 10 hours ago | parent | prev [-]

It’s probably an offshoot of making Claude more and more suitable for code/cowork.

KellyCriterion 14 hours ago | parent | prev | next [-]

> there really is no moat.

For ChatGPT and Gemini, yes.

But for Claude, they have a very deep & big one: Its the only model that gets production ready output on the first detailled prompt. Yesterday I used my tokens til noon, so I tried some output from Gemini & Co. I presented a working piece of code which is already in production:

1. It changed without noticing things like "Touple.First.Date.Created" and "Touple.Second.Date.Created" and it rendered the code unworking by chaning to "Touple.FirstDate" and "Touple.SecondDate"

2. There was a const list of 12 definitions for a given context, when telling to rewrite the function it just cut 6 of these 12 definitions, making the code not compiling - I asked why they were cut: "Sorry, I was just too lazy typing" ?? LOL

3. There is a list include holding some items "_allGlobalItems" - it changed the name in the function simply to "_items", code didnt compile

As said, a working version of a similar function was given upfront.

With Claude, I never have such issues.

ptnpzwqd 13 hours ago | parent | next [-]

I have used Claude (incl. Opus 4.6) fairly extensively, and Claude still spits out quality that is far below what I would call production ready - both littered with smaller issues, but also the occasional larger blunder. Particularly when doing anything non-trivial, and even when guiding it in detail (although that admittedly reduces the amount of larger structural issues).

Maybe it is tech stack dependent (I have mostly used it with C#/.NET), but I have heard people say the same for C#. The only conclusion I have been able to draw from this, is that people have very different definitions of production ready, but I would really like to see some concrete evidence where Claude one-shots a larger/complex C# feature or the like (with or without detailed guidance).

KellyCriterion 11 hours ago | parent | next [-]

> C#/.NET

same here :)

> one-shots a larger/complex C# feature

I can show you a timeseries data-renderer which was created with 1 initial very large prompt and then 3 following "change this and that" prompts. The file is around 5000 lines and everything works fine & exactly as specified.

allajfjwbwkwja 5 hours ago | parent | next [-]

> The file is around 5000 lines

Yep, this is another case of different standards for "production ready."

KellyCriterion 2 hours ago | parent [-]

Caught, good one! :-))

++1

ptnpzwqd 11 hours ago | parent | prev [-]

Feel free to share it, would be very curious - ideally alongside the prompts.

KellyCriterion 5 hours ago | parent [-]

Do you have an email address?

ptnpzwqd 2 hours ago | parent [-]

You can use this: hnthrowaway.outboard407@passmail.net

skeledrew 8 hours ago | parent | prev | next [-]

I don't get it though. Why do you expect perfect responses? Humans continually make mistakes, and AI is trained on human data. Yet there seems to be this higher bar of expectation for the latter. Somehow people expect this thing that's been around for a few weeks/months, and cannot learn anything more beyond its training cutoff date, to always do a better job than a human who's been around for 20+ years and is able to learn on their own until death.

ptnpzwqd 8 hours ago | parent [-]

I don't expect that - am merely responding to the parent comments claim that Claude consistently one-shots production ready code (which does not at all match my observations).

peteforde 12 hours ago | parent | prev | next [-]

I see this over and over again. I don't dispute your experience. My experience with ESP32 development has been unreasonably positive. My codebase is sitting around 600k LoC and is the product of several hundred Opus 4.x Plan -> Agent -> Debug loops. I review everything that goes through, but I'm reviewing the business logic and domain gotchas, not dumb crap like what you and so many others describe.

What is so strange to me is that surely there is more C# out there than ESP-IDF code? I don't have a good explanation beyond saying that my codebase is extensively tested and used; I would know very quickly if it suddenly started shitting the bed in the way you explain.

whaleidk 7 hours ago | parent | next [-]

600k lines of code for anything on the ESP32 sounds like the absolute polar opposite of “good”

ivan_gammel 12 hours ago | parent | prev | next [-]

The more code is out there, the worse is the average in the training dataset. There will be legacy approaches and APIs, poor design choices, popular use cases irrelevant for your context etc that increase the chances of output not matching your expectations. In Java world this is exactly how it works. I need 3-5 iterations with Claude to get things done the way I expect, sometimes jumping straight to manual refactoring and then returning the result to Claude for review and learning. My CLAUDE.md (multiple of them) are growing big with all patterns and anti-patterns identified this way. To overcome this problem model needs specialized training, that I don‘t think the industry knows how to approach (it has to beat the effort put in the education system for humans).

mjdiloreto 9 hours ago | parent | next [-]

I also believe this must be true. Try asking Claude to program in Forth, I find the results to be unreasonably good. That's probably because most of the available Forth to train on is high quality.

re-thc 10 hours ago | parent | prev [-]

> To overcome this problem model needs specialized training, that I don‘t think the industry knows how to approach

We already have coding tuned models i.e. Codex. We should just have language / technology specific models with a focus on recent / modern usage.

Problem with something like Java is too old -- too many variants. Make a cut off like at least above Java 8 or 17.

xienze 9 hours ago | parent | prev [-]

> My experience with ESP32 development has been unreasonably positive. My codebase is sitting around 600k LoC and is the product of several hundred Opus 4.x Plan -> Agent -> Debug loops.

I feel like this is an example of people having different standards of what “good” code is and hence the differing opinions of how good these tools are. I’m not an embedded developer but 600K LOC seems like a lot in that context, doesn’t it? Again I could be way off base here but that sounds like there must be a lot of spaghetti and copy-paste all over the codebase for it to end up that large.

surajrmal 8 hours ago | parent [-]

I don't think it's that large. Keep in mind embedded projects take few if any dependencies. The standard library in most languages is far bigger than 600k loc.

whaleidk 7 hours ago | parent | next [-]

I work with ESP32 devices and 600k lines of code is insane.

the__alchemist 4 hours ago | parent | prev [-]

I'm curious: What does this device do?

je42 13 hours ago | parent | prev | next [-]

Interesting - what kind of structural issues have you encountered?

Is these more related to the existing source code or is this a bad pattern thar you would never do regardless of the existing code?

huflungdung 12 hours ago | parent | prev [-]

[dead]

AlecSchueler 13 hours ago | parent | prev | next [-]

> Its the only model that gets production ready output on the first detailled prompt. Yesterday I used my tokens til noon, so I tried some output from Gemini & Co. I presented a working piece of code which is already in production:

One does often hear that where LLMs shine is with greenfield code generation but they all start to struggle working with pre-existing code. It could be that this wasn't a like for like comparison.

That said I do personally feel Claude to produce far better results than competitors.

piva00 9 hours ago | parent | next [-]

> One does often hear that where LLMs shine is with greenfield code generation but they all start to struggle working with pre-existing code. It could be that this wasn't a like for like comparison.

In my experience working in a large codebase with a good set of standards that's not the case, I can supply examples already existing in the codebase for Claude to use as a guidance and it generates quite decent code.

I think it's because there's already a lot of decent code for it to slurp and derive from, good quality tests at the functional level (so regressions are caught quickly).

I do understand though that on codebases with a hodge podge of styles, varying quality of tests, etc. it probably doesn't work as well as in my experience but I'm quite impressed about how I can do the thinking, add relevant sections of the code to the context (including protocols, APIs, etc.), describe what I need to be done, and get a plan back that most times is correct or very close to correct, which I can then iterate over to fix gaps/mistakes it made, and get it implemented.

Of course, there are still tasks it fails and I don't like doing multiple iterations to correct course, for those I do them manually with the odd usage here and there to refactor bits and pieces.

Overall I believe if your codebase was already healthy you can have LLMs work quite well with pre-existing code.

jacquesm 13 hours ago | parent | prev | next [-]

> One does often hear that where LLMs shine is with greenfield code generation but they all start to struggle working with pre-existing code.

Don't we all?

astrange 2 hours ago | parent | next [-]

I'm better at pre-existing code, if only because empty text files give me writers block.

AlecSchueler 11 hours ago | parent | prev | next [-]

Whether we do or not it's besides the point. The comparison was between Claude, which produced competent greenfield code, and Gemini which struggled with brownfield. The comparison is stacked in Claude's favour.

seba_dos1 12 hours ago | parent | prev [-]

Nope.

ivan_gammel 12 hours ago | parent | prev [-]

Greenfield implementation is not flawless as well.

ajshahH 8 hours ago | parent [-]

The only sources of these “it works flawlessly” I know of are:

- literal Claude ads I see online

- my underperforming coworkers whose code I’ve had to cleanup and know first hand that no, it wasn’t flawless

This kind of sentiment is gaslighting CTOs everywhere though. Very annoying.

ben_w 14 hours ago | parent | prev | next [-]

That's been my experience too. I'm using the recent free trial of OpenAI Plus to vibe code, and from this I would say that if Claude Code is a junior with 1-3 years of experience, OpenAI's Codex is like a student coder.

Oreb 13 hours ago | parent [-]

Does it depend on what type of programming you do? Doing Swift/SwiftUI work, I have exactly the opposite experience. I’ve been using both recently, and I want to use Claude alone (especially after the last week’s events), but Codex is just so much faster and better.

boxedemp 3 hours ago | parent | next [-]

I find it very much matters. I find Gemini better for pretty frontends, Claude opus for planning. Gemini and opus for code reviews. Codex is great when I want the LLM do follow instructions more strictly- good if you already have a detailed design.

Definitely depends on your use.

ben_w 11 hours ago | parent | prev [-]

Swift/SwiftUI are two of the three experimental projects I'm using Codex on, the other is a physics simulation in python.

It keeps trying to re-invent the wheel, does a bad job of it.

The physics sim was supposed to be a thin wrapper around existing libraries, but instead of that it tried to write all the simulation code itself as a "fallback" (but it was broken), and never actually installed the real simulators that already did this stuff despite being told to use them in the first place. The last few dozen(!) prompts from me have been pairs of ~["Find all cases where you've re-invented the wheel, add them to the planning document", "now do them"]. And it's still not finished removing the original nonsense, so far as I can tell.

One of the two Swift experiments is just a dice roller, it took about 10 rounds of non-compiling metal shaders (I don't know metal, which is why I didn't give up and do that by hand after 4) before I managed to get that to work, and when it did work it immediately broke it again on the next four rounds. It wrote its own chart instead of using Swift Charts, and did it badly. It tried to put all the hamburger menu options into a UIAlertController. Something blocks the UI for several seconds when you change the dice font. I didn't count how many attempts it took to correctly label the D4.

The other Swift experiment was a musical instrument app, that got me to the prototype stage, eventually, but in a way that still felt like a student's project rather than a junior's project.

skeledrew 8 hours ago | parent | next [-]

> Find all cases where you've re-invented the wheel

Did you put in the original prompt the "wheels" you wanted it to use? It's a toss-up when you aren't very specific about what you want.

ben_w 7 hours ago | parent [-]

For the swift apps, at least half of the errors are of a type where I wouldn't expect to have needed to tell someone to not do it like that, and only a student could reasonably be expected to not know better.

For the python physics sim, step 1 was to generate the plan, the prompt included "I want actual plasma physics, including high-density, high-field regimes, externally applied fields, etc., so consider which FOSS libraries would suit this.", and then it proceeded itself to choose some existing libraries, and I made sure those specific named FOSS libraries actually ended up in the plan.

My first clue this wasn't going to work was that even from step 1 it was pushing for writing all the simulation code and not actually using e.g. WarpX despite that it itself had suggested WarpX. In fact, even when WarpX was in the plan, it was "integrate" rather than "just use this from the get-go".

I may well throw the whole thing out and try again with Claude when this trial expires. Most of the runs have been comically non-physical, to the extent you don't even need a physics degree to notice, or even a physics GCSE.

ben_w 9 hours ago | parent | prev [-]

(Just outside edit window, I now realise I was ambiguous in this comment, it was more like "Find all cases where you've re-invented the wheel, add their removal to the planning document")

otabdeveloper4 12 hours ago | parent | prev | next [-]

> Its the only model that gets production ready output on the first detailled prompt.

That's, just, like, your opinion, man.

KellyCriterion 11 hours ago | parent [-]

...and of a lot of colleagues in and out of my sector :)

littlestymaar 14 hours ago | parent | prev | next [-]

> But for Claude, they have a very deep & big one: Its the only model that gets production ready output on the first detailled promp

That's not a moat though. Claude itself wasn't there 6 months ago and there's no reason to think Chinese open models won't be at this level in a year at most.

To keep its current position Claude has to keep improving at the same pace as the competitor.

jccx70 8 hours ago | parent | prev [-]

[dead]

crossroadsguy 12 hours ago | parent | prev | next [-]

I wrote off ChatGPT/OpenAI because of Sam Altman and those eyeball scan things - so sort of even before all this was a rage and centre stage. Sometimes it's just the gut feeling, and while it may not always be accurate, if something doesn't "feel" right, maybe it is not right. No one else is all good either, but what I mean to say is there are some entities/people who repeatedly don't feel right, have things attached to them that never felt right, etc., and you get a combined "gut feeling". At least that's how it was for me.

Buttons840 7 hours ago | parent | prev | next [-]

I love no mote!

One day I'd like to create a server in my basement that just runs a few really really nice models, and then get some friends and CO workers to pay me $10 a month for unlimited access.

All with the understanding that if you hog the entire server I'm going to kick you off, and if you generate content that makes the feds knock on my door I'm turning over the server logs and your information. Don't be an idiot, and this can be a good thing between us friends.

It would be like running a private Minecraft server. Trust means people can usually just do what they want in an unlimited way, but "unlimited" doesn't necessarily mean you can start building an x86 processor out of redstone and lagging the whole server. And you can't make weird naked statues everywhere either.

Usually these things aren't issues among a small group. Usually the private server just means more privacy and less restriction.

afcool83 7 hours ago | parent [-]

Amazing how analogous this is to the early Internet when people started running web servers out of their basement and then eventually graduated up to being their own dial-in ISP…

jacquesm 13 hours ago | parent | prev | next [-]

> I’m pretty sure they would allow AGI to be used for truly evil purposes

It's perfectly possible that 'truly evil purposes' were the goal all along. Slogans and ethics departments are mere speed bumps on the way to generational wealth.

rustyhancock 14 hours ago | parent | prev | next [-]

I know this is necessarily a very unpopular opinion however.

I think HN in particular as a crowd are very vulnerable to the halo effect and group think when it comes to Anthropic.

Even being generous they are only very minimally a "better actor" than OpenAI.

However, we are so enthralled by their product that we tend to let the view bleed over to their ethics.

Saying we want out tools used in line with the US constitution within the US on one particular point. Is hardly a high moral bar, it's self preservation.

All Anthropic have said is:

1. No mass domestic surveillance of Americans.

2. No fully autonomous lethal weapons yet.

My goodness that's what passes for a high moral standard? Really anything that doesn't hit those very carefully worded points is not "evil"?

JauntyHatAngle 14 hours ago | parent | next [-]

Lets generalise a bit more here - every company at any time could completely heel-turn and do awful things. Even my favourite private companies (e.g. Valve) have done things that I would consider evil.

However, I would think I'm not alone in that I'm generally wanting to do good while also wanting convenience, I know that really every bit of consumption I do is probably negative in some ways, and there is no real "apolitical" action anyone can take.

But can't I at least get annoyed and take my money somewhere else for the short amount of time another company is doing it better?

Yes, if openAI suddenly leaps forwards with codex and pounds anthropic into the dust, I'll likely switch back despite my moral grievances, but in a situation where I can get mildly motivated to jump over for something that - to me - seems like a better morality without much punishment to me, I'll do it.

bluGill 11 hours ago | parent [-]

There are no universial morals. Anything - everything you think is evil some culture (possibly in history) thinks is good). I can't even think of something good that I'm confident everyone would agree is good.

there are some people (companies are run by people) that are so bad I boycott them. Most bad I treat like society cannot work without accepting them anyway.

codechicago277 an hour ago | parent [-]

There’s no possibility or need for morality to be universal, and societies have improved their ethics many times throughout history. Your take is nihilistic and presupposes that moral progress isn’t possible, even though we’ve seen objective moral progress many times.

earthnail 14 hours ago | parent | prev | next [-]

Well, they did stand up to the US administration and lost a lot of money in the process. That takes courage. They clearly were being bullied into compliance, and they stood their ground.

You can see the significance of this is you look at German Nazi history. If more companies had stood up to the administration, the Nazi state would have been significantly harder to build.

In my opinion, what Anthropic did is not a small thing at all.

rustyhancock 12 hours ago | parent [-]

The comment I replied to said that they believed OpenAI would allow "AGI to be used for truly evil purposes".

By contrast Anthropic wouldn't? Yet Anthropics stance is only two narrow restrictions. As I said are those two things the only evil things possible?

If not, why is it that people on HN think Anthropic would not allow evil usage?

My hypothesis is a halo effect. We are so enthralled by Claudes performance that some struggle to rationally assess what Anthropic has actually done.

Yes it's no small thing to say no to the Trump administration but that does not mean they haven't said Yes to otherwise facilitated other evils.

In fact to me the statements from Anthropic seem to make clear they are okay with many evils.

thunky 10 hours ago | parent [-]

> Yet Anthropics stance is only two narrow restrictions.

Really I think Anthropic should have a single restriction: to not assist with illegal or unconstitutional activities. If automated killings etc is illegal then it would be covered by that one rule.

I don't think Anthropic should be in the business of deciding what is "evil".

toss1 8 hours ago | parent [-]

If each of us individually or as corporations should not be in the business of deciding what it "evil", who should be in that business?

Everyone SHOULD continuously consider, decide, and live by moral judgements and codes they internalize, and use to make choices in life.

This aspect of life should NEVER be outsourced — of course, learn from and use codes others have developed and lived by — but ALWAYS consider deeply how it works in your situation and life.

(And no, I do NOT mean use situational ethics, I mean each considering, choosing, and internalizing the codes by which they live).

So, yes, Anthropic and anyone else building products absolutely should be deciding for themselves what they will build, for what purposes it is fit to use, and telling others about those purposes. For products like AI, this absolutely includes deciding what is "evil" and preventing such uses.

If the customer finds such restrictions are not what they want, they ARE FREE to not use the product.

thunky 2 hours ago | parent [-]

> If each of us individually or as corporations should not be in the business of deciding what it "evil", who should be in that business?

This is easy imo. Two methods:

1. The law. It should not be legal for the US Govt to murder people at will. If it is legal, then of course they'll use tools to make it easier. Maybe AI, maybe Clippy. If they can't use AI then they'll fall back to using some other way of doing it like they've already been doing for several years.

2. Voting. For representatives that actually represent us and have our interest in mind rather than their own corrupt interests. And voting with our wallet against companies that do legal but morally bankrupt things.

Of course we're failing both of these hard right now. But imo the answer is not to give up and let corporations make the rules.

In other words, if it were legal for a normal citizen to murder anyone they wanted, of course they'll use Google Maps to help them do that. We don't put restrictions on how people can use Google Maps. Instead we've made murder illegal. We should be doing the same thing here.

jacquesm 13 hours ago | parent | prev | next [-]

It's not high. But it is higher.

rustyhancock 12 hours ago | parent [-]

We'll take anything we can right now. I agree.

Although we shouldn't let that mean we misjudge what we are actually getting.

jacquesm 12 hours ago | parent [-]

As a rule when there are large companies and/or billionaires involved you are in for trouble.

ekianjo 11 hours ago | parent | prev [-]

Let's not forget they also lobby to forbid models from China and pretend that distillation is stealing. but somehow just because they said no to two points the majority of HN folks think them as virtuous.

bko 7 hours ago | parent | prev | next [-]

I never understood the point of this kind of comment. It doesn't add any value or anything to the discussion. Its basically two paragraphs with some presupposition (openai bad) and how the author is virtuous by canceling his subscription. No explanation, argument, nuance. Its just virtue signaling. Actually... I guess I do know the point of this kind of comment. I just don't know why these kinds of comments get upvoted, even if you do agree openai bad

9 hours ago | parent | prev | next [-]
[deleted]
bossyTeacher 12 hours ago | parent | prev | next [-]

I tried Claude recently (after they dropped the nonsensical requirement to give them your phone number) and I was surprised to see how significantly less sycophant it was. Chatgpt, unless you are talking hard science, tends to be overly agreeable. Claude questions you a lot (you ask for x and it asks you stuff like: why are you interested in x, or based on our previous convo, x might not be suitable to you, or I see your point but based on our previous convo, y is better than x, etc). Chatgpt rarely does that.

Of course, also OpenAI being ran by openly questionable people while Dario so far doesn't seem nowhere near as bad even if none of them are angels.

samiv 11 hours ago | parent | prev | next [-]

I did the same thing and cancelled my OpenAI plan today. Besides boycotting it for their latest grifting I also found it to not really produce much value in my use cases.

Moving back to doing this archaic thing called using my own brain to do my work. Shocking.

mannanj 8 hours ago | parent | prev | next [-]

Yes they have a great marketing team and a powerful astro turfing presence though, especially with the recent "Claude beat up OpenClaw! OpenAI is supporting the community by buying it!" and that nonsense.

Though tbh I hardly feel Claude is innocent either. When their safety engineer/leader left, I didn't see any statements from the Anthropic team not one addressing the legitimate points of his for why he left. Instead we got an eager over-push in the media cycle of "Anthropic standing up to DOD! Here's why you can trust us!"

It's all sounds too similar to propaganda and astroturfing to me.

Gooblebrai 13 hours ago | parent | prev | next [-]

Claude still doesn't have image generation?

Sammi 12 hours ago | parent | next [-]

Image generation isn't what most devs spend most of their time on?

wongarsu 9 hours ago | parent | prev | next [-]

It is semi-competent at making SVGs. Which are the only kind of images I really need in dev work.

For marketing or personal stuff I do sometimes want images, but I don't really mind going somewhere else for that

nkmnz 12 hours ago | parent | prev | next [-]

Interesting. Have been using Gemini, Gpt and Claude extensively in parallel and never noticed that.

toss1 8 hours ago | parent | prev | next [-]

I'm switching over to Claude from OpenAI, and I don't care. OpenAI's image generation is terrible anyway. Just try to get it to generate something to scale, like a cabinet for a specific kitchen or bathroom space. Give it all the explicit constraints, initial sketches, etc. it wants.

The results are laughably bad.

Sure, it does get some of the tones and features, but any kind of actual real-world constraint is so far off, and the dimension indicators it includes are hilarious if they weren't so bad.

oldpersonintx 13 hours ago | parent | prev [-]

[dead]

neya 14 hours ago | parent | prev [-]

[flagged]

pants2 7 hours ago | parent | next [-]

OpenAI actually does have two excellent OSS models. Not Anthropic. Not that OpenAI is 'open' per se, but more so than Anthropic. Also see the Codex vs Claude Code extensibility.

neya 5 hours ago | parent [-]

They are far from excellent and they were open sourced due to the mounting pressure for calling themselves "Open" AI and not doing anything open. At the time, they also had Chinese competitors wiping the market value of many stocks (NVidia, etc.) after releasing true OSS models that performed as good as SOTA models and they had to retaliate. I don't know of anyone who uses those OSS models in production instead of Qwen series or DeepSeek.

Mashimo 12 hours ago | parent | prev | next [-]

What is your definition of NPC behavior?

neya 5 hours ago | parent [-]

> Random guy on the internet posts links to cancel ChatGPT subscription

> Cancels subscription

> Random guy on the internet tells you to be outraged

> Gets outraged

I'm not even a fan of OpenAI generally speaking, but, this is just silly cancelling them for no reason. If not them, some other lab would have done it. Or worse, DoW would've forced them to.

bsder 13 hours ago | parent | prev [-]

> I swear HN is just a bunch of fanboys full of NPC behavior.

Why are you assuming these are real people and not NPCs?

The amount of money flowing around AI is staggering. To believe that the AI companies aren't flooding all the social media zones with propaganda is disingenuous.

neya 13 hours ago | parent | next [-]

> To believe that the AI companies aren't flooding all the social media zones with propaganda is disingenuous.

Touché

TacticalCoder 9 hours ago | parent | prev [-]

> To believe that the AI companies aren't flooding all the social media zones with propaganda is disingenuous.

You don't use "believe" with "disingenuous": it literally makes zero sense.

If people honestly believe that, they may be naive. Or they can be "disingenuous" if they're not being sincere. But if you just say what you believe, you're sincere (and maybe naive), and hence cannot possibly be disingenuous.