| ▲ | txrx0000 5 hours ago |
| This is why you can't gatekeep AI capabilities. It will eventually be taken from you by force. It's time to open-source everything. Papers, code, weights, financial records. Do all of your research in the open. Run 100% transparent labs so that there's nothing to take from you. Level the playing field for good and bad actors alike, otherwise the bad actors will get their hands on it while everyone else is left behind. Start a movement to make fully transparent AI labs the worldwide norm, and any org that doesn't cooperate is immediately boycotted. Stop comparing AI capabilities to nuclear weapons. A nuke cannot protect against or reverse the damage of another nuke. AI capabilities are not like nukes. General intelligence should not be in the hands of a few. Give it to everyone and the good will prevail. Build a world where millions of AGIs run on millions of gaming PCs, where each AI is aligned with an individual human, not a corporation or government (which are machiavellian out of necessity). This is humanity's best chance at survival. |
|
| ▲ | magicalist 4 hours ago | parent | next [-] |
| > This is why you can't gatekeep AI capabilities. What is why? You never actually say that part, unless it's "It will eventually be taken from you by force" which doesn't seem applicable to this situation or this site? |
| |
| ▲ | txrx0000 4 hours ago | parent [-] | | I'm referring to the current situation. How is it not applicable? I think the government wants to eventually nationalize these companies and we have to stop them. | | |
| ▲ | noisy_boy 2 hours ago | parent [-] | | Nationalisation is an option worse than the advantage of having the companies at their whim and command while keeping them around as a separate entities for blame-gaming and convenience based distancing. |
|
|
|
| ▲ | bottlepalm 4 hours ago | parent | prev | next [-] |
| What use are weights without the hardware to run them? That's the gate. Local AI right now is a toy in comparison. Nukes are actually a great example of something also gated by resources. Just having the knowledge/plans isn't good enough. |
| |
| ▲ | txrx0000 3 hours ago | parent | next [-] | | Scaling has hit a wall and will not get us to AGI. Open-source models are only a couple of months behind closed models, and the same level of capability will require smaller and smaller models in the future. This is where open research can help: make the models smaller ASAP. I think it's likely that we'll be able to get something human-level to run on a single 16GB GPU before the end of the decade. | | |
| ▲ | tbrownaw 3 hours ago | parent | next [-] | | > human-level to run on a single 16GB GPU before the end of the decade. That's apparently about 6k books' worth of data. | | |
| ▲ | txrx0000 2 hours ago | parent [-] | | For the weights and temporary state, yes. It doesn't sound like a lot until you remember that your DNA is about 600 books worth of data by the same metric. |
| |
| ▲ | drdaeman 2 hours ago | parent | prev [-] | | > Open-source models are only a couple of months behind closed models Oh, come on, surely not just a couple months. Benchmarks may boast some fancy numbers, but I just tried to save some money by trying out Qwen3-Next 80B and Qwen3.5 35B-A3B (since I've recently got a machine that can run those at a tolerable speed) to generate some documentation from a messy legacy codebase. It was nowhere close neither in the output quality nor in performance to any current models that the SaaS LLM behemoth corps offer. Just an anecdote, of course, but that's all I have. |
| |
| ▲ | fooker 4 hours ago | parent | prev | next [-] | | > hardware to run them Costs a few hundred thousand per server, it's a huge expense if you want it at your home but a rounding error for most organizations. | | |
| ▲ | bottlepalm 4 hours ago | parent [-] | | You're buying what exactly for a few hundred thousand? and running what model on it? to support how many users? at what tps? | | |
| ▲ | fooker 2 hours ago | parent [-] | | Not every use case is a cloud provider or tech giant. Newer Blackwell does 200+ tokens per second on the largest models and tens of thousands on the smaller models. Most military applications require fast smaller models, I'd imagine. Also, custom chips are reportedly approaching an order of magnitude more for the price. It's a matter of availability right now, but that will be solved at some point. |
|
| |
| ▲ | reactordev 4 hours ago | parent | prev [-] | | I run local models on Mac studios and they are more than capable. Don’t spread fud. | | |
| ▲ | bottlepalm 4 hours ago | parent [-] | | You're spreading fud. There's nothing you can run locally that's on par with the speed/intelligence of a SOTA model. | | |
| ▲ | 3836293648 4 hours ago | parent [-] | | You may be correct about the level of models you can actually run on consumer hardware, but it's not fud and you're being needlessly aggressive here. |
|
|
|
|
| ▲ | msuniverse2026 4 hours ago | parent | prev | next [-] |
| I'd prefer something akin to the Biological Weapons Treaty which prohibits development, production and transfer. If you think it isn't possible you have to tell me why the bioweapons convention was successful and why it wouldn't be in the case of AI. |
| |
| ▲ | tgma 4 hours ago | parent | next [-] | | > bioweapons convention was successful Was it successful? The jury is still out. | | |
| ▲ | xpe 4 hours ago | parent [-] | | The point I would make: there are historical examples of international cooperation that work at least for some lengths of time. This is a good thing, a good tool to strive for, albeit difficult to reach. |
| |
| ▲ | Muromec 4 hours ago | parent | prev | next [-] | | Because bioweapons suck, this is why. On the other hand AI sucks too, but it has at least some use | | |
| ▲ | jrumbut 4 hours ago | parent [-] | | There might be a small percentage of people nihilistic enough to want to unleash a truly devastating bioweapon, but basically everyone wants what AI has to offer. I think that's a key difference as well. And how would a treaty like that be enforced? Every country has legitimate uses for GPUs, to make a rendering farm or simulations or do anything else involving matrix operations. All of the technology involved, in more or less the configuration needed to make your own ChatGPT, is dual use. |
| |
| ▲ | smegger001 4 hours ago | parent | prev | next [-] | | because bio-weapons labs take more to run than a workstation pc under your desk with a good graphics card. both in equipment material and training. Its hard to outlaw use of linear algebra and matrix multiplications. | | |
| ▲ | aaronblohowiak 4 hours ago | parent [-] | | The last part of your post doesn’t necessarily follow or support your argument; the corollary is “It’s hard to outlaw rna”. |
| |
| ▲ | txrx0000 4 hours ago | parent | prev [-] | | Don't compare general intelligence to bioweapons. A bioweapon cannot defend against or reverse the effects of another bioweapon. | | |
| ▲ | drdeca 4 hours ago | parent [-] | | I don’t see why you think that AGI can reverse the effects of another AGI? | | |
|
|
|
| ▲ | medi8r 4 hours ago | parent | prev | next [-] |
| Open Source here is not enough as hardware ownership matters. In an open source world, you and I cannot run the 10 trillion param model, but the data center controllers can. |
| |
| ▲ | txrx0000 4 hours ago | parent | next [-] | | I agree. We will need hardware ownership as well eventually. But the earlier you open-source, the more you slow down the centralization because people will be more likely to buy hardware to run stuff at home and that gives hardware companies an opening to do the right thing. | |
| ▲ | layer8 4 hours ago | parent | prev [-] | | Sure, but we could have Hetzners and OVHs who just provide the compute for whatever model we want to run. | | |
| ▲ | medi8r 3 hours ago | parent [-] | | Checked the DDR5 price lately? | | |
| ▲ | layer8 3 hours ago | parent [-] | | I didn’t claim that it would be cheap. But I’d rather see the real cost of SOTA LLM use exposed. On the other hand, reportedly SOTA LLM inference is profitable nowadays, so it can’t be that expensive. |
|
|
|
|
| ▲ | jefftk 4 hours ago | parent | prev | next [-] |
| A "world where millions of AGIs run on millions of gaming PCs, where each AI is aligned with an individual human" would be a world in which people could easily create humanity-ending bioweapons. I would love to live in a less vulnerable world, and am working full time to bring about such a world, but in the meantime what you describe would likely be a disaster. |
| |
| ▲ | m4rtink 2 hours ago | parent | next [-] | | I think it is much more likely they will be (and are) generating protorealistic images of ther favourite person (real or fictional) with cat ears. Never underestimate what adding cat ears does. OK, maybe someone will build a bioweapon that does that for real. :P | |
| ▲ | txrx0000 4 hours ago | parent | prev | next [-] | | There are plenty of physical and legal barriers to creating a bioweapon and that's not going to change if everyone becomes smarter with AI. And even if we really somehow end up in a world where everyone has a lab at home and people can easily create viruses, they can also easily create vaccines and anti-virals. The advancements in medicine will outpace bioweapons by a lot because most people are afraid of bioweapons. Intelligence itself is not dangerous unless only a few orgs control it and it's aligned to those orgs' values rather than human values. The safety narrative is just "intelligence for me, but not for thee" in disguise. | | |
| ▲ | jefftk 4 hours ago | parent [-] | | There mostly aren't physical barriers. Unlike nukes, where you need specific materials and equipment that we can try to keep tabs on, bioweapons can be made entirely with materials and equipment that would not be out of place in an academic or commercial lab. The largest limitation is knowledge, and the barriers there are falling quickly. On your second point, see my response to oceanplexian below: https://news.ycombinator.com/item?id=47189385 |
| |
| ▲ | oceanplexian 4 hours ago | parent | prev [-] | | I’m tired of these bizarre hypothetical gotcha arguments. If AI can create bioweapons, it can equally create vaccines and antidotes to them. We live in a free society. AI should be democratized like any other technology. | | |
| ▲ | jefftk 4 hours ago | parent | next [-] | | Symmetry is not guaranteed. If someone creates a deadly pathogen with a long pre-symptomatic period (which we know is possible, since HIV works this way) it could infect essentially everyone before discovery. Yes, powerful AI would likely rapidly speed up the process of responding to the threat after detection, especially in designing countermeasures, but if we don't learn about the threat in time we lose. There are people today who could create such a pathogen, but not many. Widespread access to powerful AI risks lowering the bar enough that we get overlap between "people who want to kill us all" and "people able to kill us all". This is not a gotcha argument, this is what I work full time on preventing: https://naobservatory.org The world must be in a position to detect attacks early enough that they won't succeed, and we're not there yet. | | |
| ▲ | txrx0000 3 hours ago | parent | next [-] | | For every person that thinks about creating the HIV-like deadly pathogen, there will be millions more thinking about how to defend people against such pathogen, how to detect it faster before symptoms arise, how to put up barriers to creating them, and possibly even how to modify our bodies to be naturally resilient to all similar pathogens. Just like what you're doing here. I don't think we should mark knowledge or intelligence itself as the problem. If that's true then we should be making everyone dumber. | |
| ▲ | jph00 4 hours ago | parent | prev [-] | | In the alternative, asymmetry is guaranteed. When you only allow gov and big tech access to powerful AI, you create a much more dangerous and unstable world. |
| |
| ▲ | dcre 4 hours ago | parent | prev [-] | | This is just not thinking clearly. There are bad things that are asymmetric in character, dramatically easier to do than to mitigate. There’s no antidote or vaccine to nuclear weapons. | | |
| ▲ | jph00 3 hours ago | parent | next [-] | | This is exactly the thinking that has characterized responses to new sources of power through history, and has been consistently used to excuse hoarding of that power. In the end, enlightenment thinking has largely won out in the western world, and society has prospered as a result. Centralizing power is dangerous and leads to power struggles and instability. | |
| ▲ | txrx0000 4 hours ago | parent | prev [-] | | It is not easy to create weapons. Why do you think the physical and legal barriers that exist today that prevent you from acquiring equipment and creating nuclear weapons will go away when everyone becomes smarter? |
|
|
|
|
| ▲ | claudiojulio 4 hours ago | parent | prev | next [-] |
| If it's taken by force, it will stagnate. It makes no sense at all. |
| |
| ▲ | avaer 4 hours ago | parent | next [-] | | The logic used in the treats is that it's a national security risk to not use Claude, but it's also a national security risk to use Claude. We shouldn't expect these people to consider how the logic breaks down one step ahead when it never made sense in the first place. | |
| ▲ | wahnfrieden 4 hours ago | parent | prev [-] | | Is TikTok stagnating in the US? |
|
|
| ▲ | pluc 5 hours ago | parent | prev | next [-] |
| When have US corporations (or simply "the US" really) ever done the right thing for humanity? |
| |
|
| ▲ | no_wizard 4 hours ago | parent | prev | next [-] |
| This letter and all of this is meaningless. If they actually wanted to do something they wouldn’t have sat back and funded Republican political campaigns because they were pissed about the head of the ftc under Biden. But they didn’t. They gave millions to this guy and now they’re feigning ignorance or change ir wherever this is. It’s meaningless. Utterly meaningless. Get what you pay for, I suppose. |
| |
| ▲ | inkysigma 4 hours ago | parent | next [-] | | What are you talking about? Google employees and the corporation itself in particular overwhelmingly donated to the Harris campaign. https://www.opensecrets.org/orgs/alphabet-inc/recipients?id=... The corporation gave millions _after_ Trump had already won. If your criticism is that, then that does not apply to the people signing. | |
| ▲ | SpicyLemonZest 4 hours ago | parent | prev [-] | | We shouldn't be scammed by people who intend to get back on the Trump train once they've gotten what they want. But if someone's willing to openly oppose the Trump regime, even out of self-interest, I'm happy to let them feign as much ignorance as they'd like. If his power isn't broken the details of who resisted him when won't matter. |
|
|
| ▲ | 5o1ecist 4 hours ago | parent | prev | next [-] |
| They control the compute. |
|
| ▲ | xpe 4 hours ago | parent | prev [-] |
| > This is why you can't gatekeep AI capabilities. They will eventually be taken from you by force. Some form of US AI lab nationalization is possible, but it hasn't happened yet. We'll see. Nationalization can take different forms, not to mention various arrangements well short of it. I interpret the comment above as a normative claim (what should happen). It implies the nationalization threat forces the decision by the AI labs. No. I will grant it influences, in the sense that AI labs have to account for it. |