Remix.run Logo
extr 4 days ago

It’s crazy how Anthropic keeps coming up with sticky “so simple it seems obvious” product innovations and OpenAI plays catch up. MCP is barely a protocol. Skills are just md files. But they seem to have a knack for framing things in a way that just makes sense.

Jimmc414 4 days ago | parent | next [-]

Skills are lazy loaded prompt engineering. They are simple, but powerful. Claude sees a one line index entry per skill. You can create hundreds. The full instructions only load when invoked.

Those instructions can reference external scripts that Claude executes without loading the source. You can package them with hooks and agents in plugins. You pay tokens for the output, not the code that calls it.

Install five MCPs and you've burned a large chunk of tokens before typing a prompt. With skills, you only pay for what you use.

You can call deterministic code (pipelines, APIs, domain logic) with a non-deterministic model, triggered by plain language, without the context bloat.

robrenaud 4 days ago | parent | prev | next [-]

They are the LLM whisperers.

In the same way Nagel knew what it was like to be a bat, Anthropic has the highest fraction of people who approximately know what it's like to be a frontier ai model.

gabaix 4 days ago | parent | next [-]

Nagel's point is that he could not know what it was like to be a bat.

4 days ago | parent [-]
[deleted]
01HNNWZ0MV43FF 4 days ago | parent | prev | next [-]

Huh https://en.wikipedia.org/wiki/What_Is_It_Like_to_Be_a_Bat%3F

uoaei 4 days ago | parent | prev [-]

It's surprising to me that Anthropic's CEO is the only one getting real recognition for their advances. The people around him seem to be as or more crucial for their mission.

ACCount37 4 days ago | parent | next [-]

Is that really true?

I can name OpenAI CEO but not Anthropic CEO off the top of my head. And I actually like Anthropic's work way more than what OpenAI is doing right now.

uoaei 3 days ago | parent [-]

Pick up the newest edition of Time.

blueblisters 4 days ago | parent | prev | next [-]

Amanda Askell, Sholto Douglas have somewhat of a fan following on twitter

adastra22 4 days ago | parent | prev [-]

That’s always the case.

altmanaltman 4 days ago | parent | prev | next [-]

> https://www.anthropic.com/news/donating-the-model-context-pr...

This is a prime example of what you're saying. Creating a "foundation" for a protocol created an year ago that's not even a protocol

Has the Gavin Belson tecthics energy

sigmoid10 4 days ago | parent | next [-]

Anthropic is in a bit of a rough spot if you look at the raw data points we have available. Their valuation is in the same order of magnitude as OpenAI, but they have orders of magnitude fewer users. And current leaderboards for famous unsolved benchmarks like ARC AGI and HLE are also dominated by Google and OpenAI. Announcements like the one you linked are the only way for Anthropic to stay in the news cycle and justify its valuation to investors. Their IPO rumours are yet another example of this. But I really wonder how long that strategy can keep working.

ramraj07 4 days ago | parent | next [-]

Those benchmarks mean nothing. Anthropic still makes the models that gets real work done in enterprise. We want to move but are unable to.

If anyone disagrees,I would like to see their long running deep research agents built on gemini or openai.

sigmoid10 4 days ago | parent | next [-]

I have built several agents based on OpenAI now that are running real life business tasks. OpenAI's tool calling integration still beats everyone else (in fact it did from the very beginning), which is what actually matters in real world business applications. And even if some small group of people prefer Anthropic for very specific tasks, the numbers are simply unfathomable. Their business strategy has zero chance of working long-term.

dotancohen 4 days ago | parent [-]

In writing code, from what I've seen, Anthropic's models are still the most widely used. I would venture that over 50% of vibe coded apps, garbage though they are, are written by Claude Code. And they capture the most market in real coding shops as well, from what I've seen.

sigmoid10 2 days ago | parent [-]

What data are you basing your assumption on? OpenRouter? That itself is only used by a tiny fraction of people. According to the latest available numbers, OpenAI has ~800x more monthly active users than OpenRouter. So even if only 0.5% of them use it for code, it will dwarf everything that Anthropic's models produce.

taylorius 3 days ago | parent | prev [-]

Just out of interest, why do you want to move? What's wrong with Claude and Anthropic in your view? (I use it, and it works really well.)

biorach 3 days ago | parent | prev | next [-]

> Their valuation is in the same order of magnitude as OpenAI, but they have orders of magnitude fewer users.

it's an open question how many of OpenAI's users are monetizable.

There's an argument to be made that your brand being what the general public identifies with AI is a medium term liability in light of the vast capital and operating costs involved.

It may well be that Anthropic focusing on an order of magnitudes smaller, but immediately monetiazable market will play out better.

bfuller 2 days ago | parent | prev | next [-]

Anthropic has less users, but I think their value per user is higher due to claude mostly producing code. I know my shop is just gonna keep paying for $200 max subscriptions until one of these open source clients with a chinese LLM can beat sonnet 4.5 (which may be now, but not worth it for me to explore until its solid enough for my uses)

robrenaud 3 days ago | parent | prev | next [-]

Low scores on HLE and ARC AGI might be a good sign. They didn't goodhart their models. ARG AGI in particular doesn't mean much, IMO. It's just some weird hard geometry induction. I don't think it correlates well with real world problem solving.

AFAICT, claude code is the biggest engineering mind share. An apple software engineer of mine says he sometimes uses $100/day of claude code tokens at work and gets sad, because that's the budget.

Also, look at costs and revenue. OpenAI is bleeding way more than Antropic.

losvedir 3 days ago | parent | prev | next [-]

Not sure how relevant it is, but I finally decided to dip my toes in last night and write my first agent. Despite paying for ChatGPT Pro, Claude Pro, etc, you still have to load up credits to use the API version of them. I started with Claude, but there was a bug on the add credit form and I couldn't submit (I'm guessing they didn't test on MacOS Safari, maybe?). So I gave up and moved on to OpenAI's developer thing.

Maybe they should do less vibe coding on their checkout flow and they might have more users.

extr 3 days ago | parent | prev | next [-]

Hard to believe you could be so misinformed. Anthropic is not far behind OAI on revenue and has a much more stable position with most of it coming from enterprise/business customers.

andy99 4 days ago | parent | prev [-]

I’d argue openAI has put their cards on the table and they don’t have anything special, while Anthropic has not.

Their valuations come from completely different calculus: Anthropic looks much more like a high potential early startup still going after PMF while OpenAI looks more like a series B flailing to monetize.

The cutting edge has largely moved past benchmarks, beyond a certain performance threshold that all these models have reached, nobody really cares about scores anymore, except people overfitting to them. They’re going for models that users like better, and Claude has a very loyal following.

TLDR, OpenAI has already peaked, Anthropic hasn’t, this the valuation difference.

DonHopkins 3 days ago | parent | prev | next [-]

I just re-binge-watched Silicon Valley in its entirety, with the benefit of a decade of hindsight, so I could get all the interconnected characters and sub-plots and cultural references together in my head better than the first time I watched it in real time at one episode per month.

It really should be required viewing for anyone in the industry, it has so much spot-on social commentary, it's just not "tecthical" not to be fully aware of it, even if it stings.

https://silicon-valley.fandom.com/wiki/Tethics

>Meanwhile, Gavin Belson (Matt Ross) comes up with a code of ethics for tech, which he lamely calls "tethics", and urges all tech CEOs to sign a pledge to abide by the tethics code. Richard refuses to sign, he considers the pledge to be unenforceable and meaningless.

>Belson invites Richard to the inauguration of the Gavin Belson Institute for Tethics. Before Belson's speech, Richard confronts the former Hooli CEO with the fact that the tethics pledge is a stream of brazenly plagiarized banalities, much like Belson's novel Cold Ice Cream & Hot Kisses.

>Once at the podium, Belson discards his planned speech and instead confesses to his misdeeds when he was CEO of Hooli. Belson urges California's attorney general to open an investigation.

>Richard mistakenly thinks that Belson is repentant for all his past bad behavior. But, as Ron LaFlamme (Ben Feldman) explains, Belson's contrite act is just another effort to sandbag Richard. If the attorney general finds that Belson acted unethically during his tenure as Hooli CEO, the current Hooli CEO would be the one who has to pay the fine. And since Pied Piper absorbed Hooli, it would be Pied Piper that has to pay the fine.

beng-nl 4 days ago | parent | prev [-]

Tethics, Denpok.

mhalle 3 days ago | parent | prev | next [-]

Skills are not just markdown files. They are markdown files combined with code and data, which only work universally when you have a general purpose cloud-based code execution environment.

Out of the box Claude skills can call python scripts that load modules from Pypi or even GitHub, potentially ones that include data like sqlite files or parquet tables.

Not just in Claude Code. Anywhere, including the mobile app.

rcarmo 3 days ago | parent [-]

They’re not alone in that.

lacy_tinpot 4 days ago | parent | prev | next [-]

Their name is Anthropic. Their entire schtick is a weird humanization of AIs.

MCP/Tool use, Skills, and I'm sure others that I can't think of.

This is might be because of some core direction that is more coherent than other labs.

JoshuaDavid 4 days ago | parent [-]

... I am pretty sure that the name "Anthropic" is as in "principle" not as in "pertaining to human beings".

kaashif 4 days ago | parent | next [-]

The anthropic principle is named as such because it is "pertaining to human beings".

This is like saying McDonald's is named after the McDonald's happy meal rather than the McDonald brothers.

lacy_tinpot a day ago | parent | prev | next [-]

Just look at their aesthetic/branding, or the way they've trained their models. Very Anthrop-ic.

GlitchInstitute 4 days ago | parent | prev | next [-]

anthropic is derived from the Greek word anthropos (human)

https://en.wikipedia.org/wiki/Anthropic_principle

yunohn 4 days ago | parent | prev [-]

Really? Anthropic is /the/ AI company known for anthropomorphizing their models, giving them ethics and “souls”, considering their existential crises, etc.

JoshuaDavid 4 days ago | parent | next [-]

Anthropic was founded by a group of 7 former OpenAI employees who left over differences in opinions about AI Safety. I do not see any public documentation that the specific difference in opinion was that that group thought that OpenAI was too focused on scaling and that there needed to be a purely safety-focused org that still scaled, though that is my impression based on conversations I've had.

But regardless anthropic reasoning was extremely in the intellectual water supply of the Anthropic founders, and they explicitly were not aiming at producing a human-like model.

simonw 3 days ago | parent [-]

They tried to fire Sam Altman and left to form their own company when that didn't work. https://simonwillison.net/2023/Nov/22/before-altmans-ouster-...

rcarmo 3 days ago | parent | prev [-]

“you are totally right!” does feel like a very human behavior in some respects…

losvedir 3 days ago | parent | prev | next [-]

MCP is a terribly designed (and I assume vibe-designed) protocol. Give me the requirements that an LLM needs to be able to load tools dynamically from another server and invoke them like an RPC, and I could give you a much simpler, better solution.

The modern HTTP Streamable version is light-years better, but took a year and was championed by outside engineers faced with the real problem of integrating it, and I imagine was designed by a human.

OpenAI was there first, but unfortunately the models weren't quite good enough yet, so their far superior approach unfortunately didn't take off.

smokel 4 days ago | parent | prev | next [-]

Also, MCP is a serious security disaster. Too simple, I'd wager.

valzam 4 days ago | parent | next [-]

Id argue that this isn't so much a fault of the MCP spec but how 95% of AI 'engineers' have no engineering background. MCP is just an OpenAPI spec. It's the same as any other API. If you are exposing sensitive data without any authz/n that's on the developer.

sam_lowry_ 4 days ago | parent | prev | next [-]

complex is synonym of insecure

3 days ago | parent | prev | next [-]
[deleted]
brazukadev 4 days ago | parent | prev [-]

MCP biggest problem is not being simple

nl 4 days ago | parent | prev | next [-]

Also `CLAUDE.md` (which is `AGENTS.md` everywhere? else now)

msy 4 days ago | parent | prev | next [-]

I get the impression the innovation drivers at OpenAI have all moved on and the people that have moved in were the ones chasing the money, the rest is history.

nrhrjrjrjtntbt 4 days ago | parent | prev | next [-]

The RSS of AI

uoaei 4 days ago | parent [-]

I like this line of analogy. The next obvious step would be IRC (or microservices?) of AI (for co-reasoning) which could offer the space for specialized LLMs rather than the current approach of monoliths.

jbgt 4 days ago | parent [-]

Oh wow conreasoning through an IRC like chat. That's a great idea.

Would be cool (sci fi) for LLMs of different users to chat and discuss approaches to what the humans are talking about etc.

exe34 4 days ago | parent [-]

omg that's how crystal society starts and then it goes downhill! highly recommended series in this space.

speakspokespok 4 days ago | parent | prev | next [-]

I noticed something like this earlier, in the android app you can have it rewrite a paragraph, and then and only then do you have the option to send that as a text message. It's just a button that pops up. Claude has an elegance to it.

ivape 4 days ago | parent | prev | next [-]

It’s the only AI company that isn’t monetize at all costs. I’m curious how deep their culture goes as it’s remarkable they even have any discernible value system in today’s business world.

rcarmo 3 days ago | parent | prev | next [-]

Well, my MCP servers only really started working when I implemented the prompt endpoints, so I’m happy I’ll never have to use MCP again if this sticks.

blitzar 4 days ago | parent | prev | next [-]

Anthropic are using Ai beyond the chat window. Without external information, context and tools the "magic" of Ai evaporates after a few minutes.

baxtr 4 days ago | parent | prev | next [-]

A good example of:

Build things and then talk about them in a way that people remember and share it with friends.

I guess some call it clever product marketing.

extr 4 days ago | parent | prev | next [-]

Oh yeah I forgot the biggest one. Claude fucking code. Lol

baby 4 days ago | parent [-]

I was very skeptical about anything not OpenAI for a while, and then discovered Claude code, Anthropic blogposts, etc. It's basically the coolest company in the field.

mh- 4 days ago | parent | next [-]

Claude Code and its ecosystem is what made me pick Anthropic over OpenAI for our engineers, when we decided to do seat licensing for everyone last week.

It's a huge asset.

joemazerino 3 days ago | parent | prev | next [-]

I appreciate Claude not training on my data by default. ChatGPT through the browser does not give you that option.

skeptic_ai 4 days ago | parent | prev [-]

Same here. Until I read more about them and actually seem sketchy too. All about “safety” reasons not to do certain things.

_pdp_ 4 days ago | parent | prev | next [-]

I hate to be that guy but skills are not an invention of sorts. It a simple mechanism that exists already in many places.

The biggest unlock was tool calling that was in invented at OpenAI.

simonw 3 days ago | parent [-]

I'd credit tool calling to the ReAct paper, which was Princeton CA and Google DeepMind: https://arxiv.org/abs/2210.03629

_pdp_ 3 days ago | parent [-]

Oh nice. I did not know. Thanks for the link.

CuriouslyC 3 days ago | parent | prev [-]

Anthropic has good marketing, but ironically their well marketed mediocre ideas retard development of better standards.