Remix.run Logo
quantumwannabe 9 hours ago

More details on the difference between the OpenAI and Anthropic contracts from one of the Under Secretaries of State:

>The axios article doesn’t have much detail and this is DoW’s decision, not mine. But if the contract defines the guardrails with reference to legal constraints (e.g. mass surveillance in contravention of specific authorities) rather than based on the purely subjective conditions included in Anthropic’s TOS, then yes. This, btw, was a compromise offered to—and rejected by—Anthropic.

https://x.com/UnderSecretaryF/status/2027566426970530135

> For the avoidance of doubt, the OpenAI - @DeptofWar contract flows from the touchstone of “all lawful use” that DoW has rightfully insisted upon & xAI agreed to. But as Sam explained, it references certain existing legal authorities and includes certain mutually agreed upon safety mechanisms. This, again, is a compromise that Anthropic was offered, and rejected.

> Even if the substantive issues are the same there is a huge difference between (1) memorializing specific safety concerns by reference to particular legal and policy authorities, which are products of our constitutional and political system, and (2) insisting upon a set of prudential constraints subject to the interpretation of a private company and CEO. As we have been saying, the question is fundamental—who decides these weighty questions? Approach (1), accepted by OAI, references laws and thus appropriately vests those questions in our democratic system. Approach (2) unacceptably vests those questions in a single unaccountable CEO who would usurp sovereign control of our most sensitive systems.

> It is a great day for both America’s national security and AI leadership that two of our leading labs, OAI and xAI have reached the patriotic and correct answer here

https://x.com/UnderSecretaryF/status/2027594072811098230

toraway 8 hours ago | parent | next [-]

  It is a great day for both America’s national security and AI leadership that two of our leading labs, OAI and xAI have reached the patriotic and correct answer here
He's an administration official openly cheerleading his team. This should be characterized as the insider perspective/spin, not a neutral analysis of the relevant facts.
MostlyStable 6 hours ago | parent | prev | next [-]

Even this most-charitable-possible (to DoW) explanation does not even come close to justifying the supply chain risk designation. It is absolutely enough (and honestly more than enough) for a contract cancellation and a switch to a competitor. DoW could have done that for any reason at all, or no reason at all. If they had issues with Anthropics terms, they 100% should have done that.

Nothing in the quoted text comes anywhere close to the realm of justifying the retaliatory actions.

ukblewis 16 minutes ago | parent | next [-]

AFAIK, the U.S. government is fully entitled to serve them under the U.S. Department of War’s terms as per the Defense Production Act. The government has yet to do this, but a company acting in a way that the Department of War perceives as benefiting enemy states could certainly be a justification for declaring a supply chain risk. Anthropic’s decision timing as the U.S. has launched a war in the Middle East to save millions of Iranian lives (tens if not hundreds of thousands of Iranians have already been killed by the Islamic Regime) definitely seems to be unjustifiable and the U.S. Department of War (so weird for me to type that instead of DOD) was smart, in my opinion, not to force Anthropic to work with them but to drop all work with them and move to providers who will meet the military’s needs while at war.

(Just in case anyone was wondering, I live in Israel)

piker 2 hours ago | parent | prev [-]

I find myself totally agreeing with the quoted text and also this sentiment. It just makes no sense to nuke Anthropic as a negotiation tactic if your interest is in preserving the republic long term.

advisedwang 8 hours ago | parent | prev | next [-]

A government promise that they'll only do lawful things is not reassuring at all:

1. We've seen government lawyers write memos explaining why such-and-such obviously illegal act is legal (see: torture memo). Until challenged, this is basically law.

2. We've seen government change the law to make whatever they want legal (see: patriot act)

3. We've seen courts just interpret laws to make things legal

A contractor doesn't realistically have the power to push back against any of these avenues if they agree to allow anything legal.

(At the risk of triggering Godwin's Law, remember that for the most part the Holocaust was entirely legal - the Nazi's established the necessary authorization. Just to illustrate that when it comes to certain government crimes, the law alone is an insufficient shield.)

Tepix 6 hours ago | parent [-]

This is it exactly.

makeramen 6 hours ago | parent | prev | next [-]

The DoW wants to only be beholden to the laws, and not to Anthropics TOS.

So the question is: do you trust the government to effectively govern its own use of AI? or do you trust Anthropic's enforcement of its TOS?

nullocator 6 hours ago | parent [-]

They DoW doesn't care about laws, that's the whole point. Anthropic did not believe the most law breaking administration in history when their drunkard incompetent leader said "lol trust us bro"

5 hours ago | parent [-]
[deleted]
ignoramous an hour ago | parent | prev | next [-]

> More details on the difference...

Does the qualifier "domestic" for mass surveillance mean that OpenAI allows the use of its models for whatever isn't "domestic"?

  ... Two of our most important safety principles are prohibitions on domestic mass surveillance and human responsibility for the use of force ...
SpicyLemonZest 9 hours ago | parent | prev [-]

You're quoting social media posts from a regime official who says he didn't participate in these negotiations and doesn't work for the relevant department.

If his characterization of the agreement is correct, which I will not believe and you should not believe until a trustworthy news outlet publishes the text, I suppose this would convince me that Hegseth does not literally plan to build a Terminator for democracy-ending purposes. There's a lot of inexcusable stuff here regardless, but perhaps merely boycotting OpenAI and the US military would be a sufficient response if this all checks out.

qwerasdf5 8 hours ago | parent [-]

> which I will not believe and you should not believe

It seems like you chose to immediately disbelieve it.

> until a trustworthy news outlet publishes the text

If you've found one of these, let me know. I'm still looking...

SpicyLemonZest 7 hours ago | parent [-]

I did choose to immediately disbelieve it. If a Trump regime official tells me something, and they could plausibly benefit from lying to me about it, I assume until proven otherwise that they're lying. They've earned this reputation through a large number of consequential and later disproven lies; my apologies to Mr. Lewin if he personally is an honest man, although I might encourage him to think about whether the good he's doing in his role is so important that it outweighs the lies he's providing cover for and the gradual erosion of his integrity.

> If you've found one of these, let me know. I'm still looking...

I do not assume, and I would recommend that you do not assume, that there is such a thing as a text of the contract. It's much easier to lie about contents of documents that don't actually exist yet. Then you can craft the text in response to public feedback, writing it down in early March and telling people that it's totally a copy of what was agreed to on February 27.

As a corollary, you should be skeptical of any purported text that is not widely published soon. If there is indeed such a contract, and it says what Altman claims, he will desperately want to ensure that his employees have read a "leak" of the text by Monday morning.