| ▲ | AI 2027(ai-2027.com) |
| 784 points by Tenoke a day ago | 517 comments |
| |
|
| ▲ | Vegenoid 4 hours ago | parent | next [-] |
| I think we've actually had capable AIs for long enough now to see that this kind of exponential advance to AGI in 2 years is extremely unlikely. The AI we have today isn't radically different from the AI we had in 2023. They are much better at the thing they are good at, and there are some new capabilities that are big, but they are still fundamentally next-token predictors. They still fail at larger scope longer term tasks in mostly the same way, and they are still much worse at learning from small amounts of data than humans. Despite their ability to write decent code, we haven't seen the signs of a runaway singularity as some thought was likely. I see people saying that these kinds of things are happening behind closed doors, but I haven't seen any convincing evidence of it, and there is enormous propensity for AI speculation to run rampant. |
| |
| ▲ | jug 2 hours ago | parent | next [-] | | > there are some new capabilities that are big, but they are still fundamentally next-token predictors Anthropic recently released research where they saw how when Claude attempted to compose poetry, it didn't simply predict token by token and "react" to when it thought it might need a rhyme and then looked at its context to think of something appropriate, but actually saw several tokens ahead and adjusted for where it'd likely end up, ahead of time. Anthropic also says this adds to evidence seen elsewhere that language models seem to sometimes "plan ahead". Please check out the section "Planning in poems" here; it's pretty interesting! https://transformer-circuits.pub/2025/attribution-graphs/bio... | | |
| ▲ | percentcer 2 hours ago | parent [-] | | Isn't this just a form of next token prediction? i.e. you'll keep your options open for a potential rhyme if you select words that have many associated rhyming pairs, and you'll further keep your options open if you focus on broad topics over niche | | |
| ▲ | DennisP an hour ago | parent | next [-] | | Assuming the task remains just generating tokens, what sort of reasoning or planning would say is the threshold, before it's no longer "just a form of next token prediction?" | |
| ▲ | throwuxiytayq 2 hours ago | parent | prev [-] | | In the same way that human brains are just predicting the next muscle contraction. | | |
| ▲ | alfalfasprout an hour ago | parent [-] | | Except that's not how it works... | | |
| ▲ | Workaccount2 6 minutes ago | parent [-] | | To be fair, we don't actually know how the human mind works. The most sure things we know is that it is a physical system, and that does feel like something to be one of these systems. |
|
|
|
| |
| ▲ | benlivengood 3 hours ago | parent | prev | next [-] | | METR [0] explicitly measures the progress on long term tasks; it's as steep a sigmoid as the other progress at the moment with no inflection yet. As others have pointed out in other threads RLHF has progressed beyond next-token prediction and modern models are modeling concepts [1]. [0] https://metr.org/blog/2025-03-19-measuring-ai-ability-to-com... [1] https://www.anthropic.com/news/tracing-thoughts-language-mod... | | |
| ▲ | Vegenoid 2 hours ago | parent | next [-] | | At the risk of coming off like a dolt and being super incorrect: I don't put much stock into these metrics when it comes to predicting AGI. Even if the trend of "length of task an AI can reliably do doubles every 7 months" continues, as they say that means we're years away from AI that can complete tasks that take humans weeks or months. I'm skeptical that the doubling trend will continue into that timescale, I think there is a qualitative difference between tasks that take weeks or months and tasks that take minutes or hours, a difference that is not reflected by simple quantity. I think many people responsible for hiring engineers are keenly aware of this distinction, because of their experience attempting to choose good engineers based on how they perform in task-driven technical interviews that last only hours. Intelligence as humans have it seems like a "know it when you see it" thing to me, and metrics that attempt to define and compare it will always be looking at only a narrow slice of the whole picture. To put it simply, the gut feeling I get based on my interactions with current AI, and how it is has developed over the past couple of years, is that AI is missing key elements of general intelligence at its core. While there's more lots more room for its current approaches to get better, I think there will be something different needed for AGI. I'm not an expert, just a human. | | |
| ▲ | Enginerrrd an hour ago | parent | next [-] | | There is definitely something qualitatively different about weeks/months long tasks. It reminds me of the difference between a fresh college graduate and an engineer with 10 years of experience. There are many really smart and talented college graduates. But, while I am struggling to articulate exactly why, I know that when I was a fresh graduate, despite my talent and ambition, I would have failed miserably at delivering some of the projects that I now routinely deliver over time periods of ~1.5 years. I think LLM's are really good at emulating the types of things I might say are the types of things that would make someone successful at this if I were to write it down in a couple paragraphs, or an article, or maybe even a book. But... knowing those things as written by others just would not quite cut it. Learning at those time scales is just very different than what we're good at training LLM's to do. A college graduate is in many ways infinitely more capable than a LLM. Yet there are a great many tasks that you just can't give an intern if you want them to be successful. There are at least half a dozen different 1000-page manuals that one must reference to do a bare bones approach at my job. And there are dozens of different constituents, and many thousands of design parameters I must adhere to. Fundamentally, all of these things often are in conflict and it is my job to sort out the conflicts and come up with the best compromise. It's... really hard to do. Knowing what to bend so that other requirements may be kept rock solid, who to negotiate with for different compromises needed, which fights to fight, and what a "good" design looks like between alternatives that all seem to mostly meet the requirements. Its a very complicated chess game where it's hopelessly impossible to brute force but you must see the patterns along the way that will point you like sign posts into a good position in the end game. The way we currently train LLM's will not get us there. Until an LLM can take things in it's context window, assess them for importance, dismiss what doesn't work or turns out to be wrong, completely dismiss everything it knows when the right new paradigm comes up, and then permanently alter its decision making by incorporating all of that information in an intelligent way, it just won't be a replacment for a human being. | |
| ▲ | benlivengood 2 hours ago | parent | prev [-] | | > I think there is a qualitative difference between tasks that take weeks or months and tasks that take minutes or hours, a difference that is not reflected by simple quantity. I'd label that difference as long-term planning plus executive function, and wherever that overlaps with or includes delegation. Most long-term projects are not done by a single human and so delegation almost always plays a big part. To delegate, tasks must be broken down in useful ways. To break down tasks a holistic model of the goal is needed where compartmentalization of components can be identified. I think a lot of those individual elements are within reach of current model architectures but they are likely out of distribution. How many gantt charts and project plans and project manager meetings are in the pretraining datasets? My guess is few; rarely published internal artifacts. Books and articles touch on the concepts but I think the models learn best from the raw data; they can probably tell you very well all of the steps of good project management because the descriptions are all over the place. The actual doing of it is farther toward the tail of the distribution. |
| |
| ▲ | Fraterkes 3 hours ago | parent | prev [-] | | The METR graph proposes a 6 year trend, based largely on 4 datapoints before 2024. I get that it is hard to do analyses since were in uncharted territory, and I personally find a lot of the AI stuff impressive, but this just doesn't strike me as great statistics. | | |
| ▲ | benlivengood 2 hours ago | parent [-] | | I agree that we don't have any good statistical models for this. If AI development were that predictable we'd likely already be past a singularity of some sort or in a very long winter just by reverse-engineering what makes the statistical model tick. |
|
| |
| ▲ | boznz an hour ago | parent | prev | next [-] | | > we haven't seen the signs of a runaway singularity as some thought was likely. The signs are not there but while we may not be on an exponential curve (which would be difficult to see), we are definitely on a steep upward one which may get steeper or may fizzle out if LLM's can only reach human level intelligence but not surpass it. Original article was a fun read though and 360,000 words shorter than my very similar fiction novel :-) | | |
| ▲ | grey-area 17 minutes ago | parent [-] | | LLMs don’t have any sort of intelligence at present, they have a large corpus of data and can produce modified copies of it. |
| |
| ▲ | ComplexSystems 2 hours ago | parent | prev | next [-] | | > They are much better at the thing they are good at, and there are some new capabilities that are big, but they are still fundamentally next-token predictors. I don't really get this. Are you saying autoregressive LLMs won't qualify as AGI, by definition? What about diffusion models, like Mercury? Does it really matter how inference is done if the result is the same? | | |
| ▲ | Vegenoid an hour ago | parent [-] | | > Are you saying autoregressive LLMs won't qualify as AGI, by definition? No, I am speculating that they will not reach capabilities that qualify them as AGI. |
| |
| ▲ | uejfiweun an hour ago | parent | prev | next [-] | | Isn't the brain kind of just a predictor as well, just a more complicated one? Instead of predicting and emitting tokens, we're predicting future outcomes and emitting muscle movements. Which is obviously different in a sense but I don't think you can write off the entire paradigm as a dead end just because the medium is different. | |
| ▲ | byearthithatius 4 hours ago | parent | prev [-] | | Disagree. We know it _can_ learn out of distribution capabilities based on similarities to other distributions. Like the TikZ Unicorn[1] (which was not in training data anywhere) or my code (which has variable names and methods/ideas probably not seen 1:1 in training). IMO this out of distribution learning is all we need to scale to AGI. Sure there are still issues, it doesn't always know which distribution to pick from. Neither do we, hence car crashes. [1]: https://arxiv.org/pdf/2303.12712 or on YT https://www.youtube.com/watch?v=qbIk7-JPB2c |
|
|
| ▲ | visarga 6 hours ago | parent | prev | next [-] |
| The story is entertaining, but it has a big fallacy - progress is not a function of compute or model size alone. This kind of mistake is almost magical thinking. What matters most is the training set. During the GPT-3 era there was plenty of organic text to scale into, and compute seemed to be the bottleneck. But we quickly exhausted it, and now we try other ideas - synthetic reasoning chains, or just plain synthetic text for example. But you can't do that fully in silico. What is necessary in order to create new and valuable text is exploration and validation. LLMs can ideate very well, so we are covered on that side. But we can only automate validation in math and code, but not in other fields. Real world validation thus becomes the bottleneck for progress. The world is jealously guarding its secrets and we need to spend exponentially more effort to pry them away, because the low hanging fruit has been picked long ago. If I am right, it has implications on the speed of progress. Exponential friction of validation is opposing exponential scaling of compute. The story also says an AI could be created in secret, which is against the validation principle - we validate faster together, nobody can secretly outvalidate humanity. It's like blockchain, we depend on everyone else. |
| |
| ▲ | tomp 4 hours ago | parent | next [-] | | Did we read the same article? They clearly mention, take into account and extrapolate this; LLM have first scaled via data, now it's test time compute, but recent developments (R1) clearly show this is not exhausted yet (i.e. RL on synthetically (in-silico) generated CoT) which implies scaling with compute. The authors then outline further potential (research) developments that could continue this dynamic, literally things that have already been discovered just not yet incorporated into edge models. Real-world data confirms their thesis - there have been a lot of sceptics about AI scaling, somewhat justified ("whoom" a.k.a. fast take-off hasn't happened - yet) but their fundamental thesis has been wrong - "real-world data has been exhausted, next algorithmic breakthroughs will be hard and unpredictable". The reality is, while data has been exhausted, incremental research efforts have resulted in better and better models (o1, r1, o3, and now Gemini 2.5 which is a huge jump! [1]). This is similar to how Moore's Law works - it's not given that CPUs get better exponentially, it still requires effort, maybe with diminishing returns, but nevertheless the law works... If we ever get to models be able to usefully contribute to research, either on the implementation side, or on research ideas side (which they CANNOT yet, at least Gemini 2.5 Pro (public SOTA), unless my prompting is REALLY bad), it's about to get super-exponential. Edit: then once you get to actual general intelligence (let alone super-intelligence) the real-world impact will quickly follow. | | |
| ▲ | Jianghong94 4 hours ago | parent [-] | | Well based on what I'm reading, the OP's intent is that, not all (hence 'fully') validation, if not most of, can be done in-silico. I think we all agree that and that's the major bottleneck making agents useful - you have to have human-in-the-loop to closely guardrail the whole process. Of course you can get a lot of mileage via synthetically generated CoT but does that lead to LLM speed up developing LLM is a big IF. | | |
| ▲ | tomp 4 hours ago | parent [-] | | No, the entire point of this article is that when you get to self-improving AI, it will become generally intelligent, then you can use that to solve robotics, medicine etc. (like a generally-intelligent baby can (eventually) solve how to move boxes, assemble cars, do experiments in labs etc. - nothing special about a human baby, it's just generally intelligent). | | |
| ▲ | Jianghong94 3 hours ago | parent | next [-] | | Not only does the article claim that when we get to self-improving ai it becomes generally intelligent, it's assuming that AI is pretty close right now: > OpenBrain focuses on AIs that can speed up AI research. They want to win the twin arms races against China (whose leading company we’ll call “DeepCent”)16 and their US competitors. The more of their research and development (R&D) cycle they can automate, the faster they can go. So when OpenBrain finishes training Agent-1, a new model under internal development, it’s good at many things but great at helping with AI research. > It’s good at this due to a combination of explicit focus to prioritize these skills, their own extensive codebases they can draw on as particularly relevant and high-quality training data, and coding being an easy domain for procedural feedback. > OpenBrain continues to deploy the iteratively improving Agent-1 internally for AI R&D. Overall, they are making algorithmic progress 50% faster than they would without AI assistants—and more importantly, faster than their competitors. > what do we mean by 50% faster algorithmic progress? We mean that OpenBrain makes as much AI research progress in 1 week with AI as they would in 1.5 weeks without AI usage. To me, claiming today's AI IS capable of such thing is too hand-wavy. And I think that's the crux of the article. | |
| ▲ | polynomial 3 hours ago | parent | prev [-] | | You had me at "nothing special about a human baby" |
|
|
| |
| ▲ | nikisil80 5 hours ago | parent | prev | next [-] | | Best reply in this entire thread, and I align with your thinking entirely. I also absolutely hate this idea amongst tech-oriented communities that because an AI can do some algebra and program an 8-bit video game quickly and without any mistakes, it's already overtaking humanity. Extrapolating from that idea to some future version of these models, they may be capable of solving grad school level physics problems and programming entire AAA video games, but again - that's not what _humanity_ is about. There is so much more to being human than fucking programming and science (and I'm saying this as an actual nuclear physicist). And so, just like you said, the AI arm's race is about getting it good at _known_ science/engineering, fields in which 'correctness' is very easy to validate. But most of human interaction exists in a grey zone. Thanks for this. | | |
| ▲ | m11a 8 minutes ago | parent | next [-] | | > that's not what _humanity_ is about I've not spent too long thinking on the following, so I'm prepared for someone to say I'm totally wrong, but: I feel like the services economy can be broadly broken down into: pleasure, progress and chores. Pleasure being poetry/literature, movies, hospitality, etc; progress being the examples you gave like science/engineering, mathematics; and chore being things humans need to coordinate or satisfy an obligation (accountants, lawyers, salesmen). In this case, if we assume AI can deal with things not in the grey zone, then it can deal with 'progress' and many 'chores', which are massive chunks of human output. There's not much grey zone to them. (Well, there is, but there are many correct solutions; equivalent pieces of code that are acceptable, multiple versions of a tax return, each claiming different deductions, that would fly by the IRS, etc) | |
| ▲ | wruza 3 hours ago | parent | prev | next [-] | | programming entire AAA video games Even this is questionable, cause we're seeing it making forms and solving leetcodes, but no llm yet created a new approach, reduced existing unnecessary complexity (which we created mountains of), made something truly new in general. All they seem to do is rehash of millions of "mainstream" works, and AAA isn't mainstream. Cranking up the parameter count or the time of beating around the bush (aka cot) doesn't magically substitute for lack of a knowledge graph with thick enough edges, so creating a next-gen AAA video game is far out of scope of llm's abilities. They are stuck in 2020 office jobs and weekend open source tech, programming-wise. | | |
| ▲ | m11a 2 minutes ago | parent [-] | | "stuck" is a bit strong of a term. 6 months ago I remember preferring to write even Python code myself because Copilot would get most things wrong. My most successful usage of Copilot was getting it to write CRUD and tests. These days, I can give Claude Sonnet in Cursor's agent mode a high-level Rust programming task (e.g. write a certain macro that would allow a user to define X) and it'll modify across my codebase, and generally the thing just works. At current rate of progress, I really do think in another 6 months they'll be pretty good at tackling technical debt and overcomplication, at least in codebases that have good unit/integration test coverage or are written in very strongly typed languages with a type-friendly structure. (Of course, those usually aren't the codebases needing significant refactoring, but I think AIs are decent at writing unit tests against existing code too.) |
| |
| ▲ | loandbehold 4 hours ago | parent | prev [-] | | OK but getting good at science/engineering is what matters because that's what gives AI and people who wield it power. Once AI is able to build chips and datacenters autonomously, that's when singularity starts. AI doesn't need to understand humans or act human-like to do those things. |
| |
| ▲ | nfc 34 minutes ago | parent | prev | next [-] | | I agree with your point about the validation bottleneck becoming dominant over raw compute and simple model scaling. However, I wonder if we're underestimating the potential headroom for sheer efficiency breakthroughs at our levels of intelligence. Von Neumann for example was incredibly brilliant, yet his brain presumably ran on roughly the same power budget as anyone else's. I mean, did he have to eat mountains of food to fuel those thoughts? ;) So it looks like massive gains in intelligence or capability might not require proportionally massive increases in fundamental inputs at least at the highest levels of intelligence a human can reach, and if that's true for the human brain why not for other architecture of intelligence. P.S. It's funny, I was talking about something along the lines of what you said with a friend just a few minutes before reading your comment so when I saw it I felt that I had to comment :) | |
| ▲ | the8472 2 hours ago | parent | prev [-] | | Many tasks are amenable to simulation training and synthetic data. Math proofs, virtual game environments, programming. And we haven't run out of all data. High-quality text data may be exhausted, but we have many many life-years worth of video. Being able to predict visual imagery means building a physical world model. Combine this passive observation with active experimentation in simulated and real environments and you get millions of hours of navigating and steering a causal world.
Deepmind has been hooking up their models to real robots to let them actively explore and generate interesting training data for a long time. There's more to DL than LLMs. |
|
|
| ▲ | stego-tech 16 hours ago | parent | prev | next [-] |
| It’s good science fiction, I’ll give it that. I think getting lost in the weeds over technicalities ignores the crux of the narrative: even if this doesn’t lead to AGI, at the very least it’s likely the final “warning shot” we’ll get before it’s suddenly and irreversibly here. The problems it raises - alignment, geopolitics, lack of societal safeguards - are all real, and happening now (just replace “AGI” with “corporations”, and voila, you have a story about the climate crisis and regulatory capture). We should be solving these problems before AGI or job-replacing AI becomes commonplace, lest we run the very real risk of societal collapse or species extinction. The point of these stories is to incite alarm, because they’re trying to provoke proactive responses while time is on our side, instead of trusting self-interested individuals in times of great crisis. |
| |
| ▲ | wruza 15 hours ago | parent | next [-] | | No one's gonna solve anything. "Our" world is based on greedy morons concentrating power through hands of just morons who are happy to hit you with a stick. This system doesn't think about what "we" should or allowed to do, and no one's here is at the reasonable side of it either. lest we run the very real risk of societal collapse or species extinction Our part is here. To be replaced with machines if this AI thing isn't just a fart advertised as mining equipment, which it likely is. We run this risk, not they. People worked on their wealth, people can go f themselves now. They are fine with all that. Money (=more power) piles in either way. No encouraging conclusion. | | |
| ▲ | Davidzheng 6 hours ago | parent | next [-] | | Don't think it's correct to blame the fact that AI acceleration is the only viable self-protecting policy on "greedy morons". | |
| ▲ | jrvarela56 13 hours ago | parent | prev [-] | | https://slatestarcodex.com/2014/07/30/meditations-on-moloch/ | | |
| ▲ | wruza 12 hours ago | parent | next [-] | | Thanks for the read. One could think that the answer is to simply stop being a part of it, but then again you're from the genus that outcompeted everyone else in staying alive. Nature is such a shitty joke by design, not sure how one is supposed to look at the hypothetical designer with warmth in their heart. | | |
| ▲ | braebo 8 hours ago | parent [-] | | Fleshy meat sacks on a space rock eating one another alive and shitting them out on a march towards inevitable doom in the form of a (likely) painful and terrifying death is a genius design, no? |
| |
| ▲ | Aeolun 10 hours ago | parent | prev [-] | | I read for such a long time, and I still couldn’t get through that, even though it never got boring. I like that it ends with a reference to Kushiel and Elua though. |
|
| |
| ▲ | fmap 9 hours ago | parent | prev | next [-] | | > even if this doesn’t lead to AGI, at the very least it’s likely the final “warning shot” we’ll get before it’s suddenly and irreversibly here. I agree that it's good science fiction, but this is still taking it too seriously. All of these "projections" are generalizing from fictional evidence - to borrow a term that's popular in communities that push these ideas. Long before we had deep learning there were people like Nick Bostrom who were pushing this intelligence explosion narrative. The arguments back then went something like this: "Machines will be able to simulate brains at higher and higher fidelity. Someday we will have a machine simulate a cat, then the village idiot, but then the difference between the village idiot and Einstein is much less than the difference between a cat and the village idiot. Therefore accelerating growth[...]" The fictional part here is the whole brain simulation part, or, for that matter, any sort of biological analogue. This isn't how LLMs work. We never got a machine as smart as a cat. We got multi-paragraph autocomplete as "smart" as the average person on the internet. Now, after some more years of work, we have multi-paragraph autocomplete that's as "smart" as a smart person on the internet. This is an imperfect analogy, but the point is that there is no indication that this process is self-improving. In fact, it's the opposite. All the scaling laws we have show that progress slows down as you add more resources. There is no evidence or argument for exponential growth. Whenever a new technology is first put into production (and receives massive investments) there is an initial period of rapid gains. That's not surprising. There are always low-hanging fruit. We got some new, genuinely useful tools over the last few years, but this narrative that AGI is just around the corner needs to die. It is science fiction and leads people to make bad decisions based on fictional evidence. I'm personally frustrated whenever this comes up, because there are exciting applications which will end up underfunded after the current AI bubble bursts... | | |
| ▲ | gwd 8 hours ago | parent | next [-] | | > Someday we will have a machine simulate a cat, then the village idiot... This isn't how LLMs work. I think you misunderstood that argument. The simulate the brain thing isn't a "start from the beginning" argument, it's an "answer a common objection" argument. Back around 2000, when Nick Bostrom was talking about this sort of thing, computers were simply nowhere near powerful enough to come even close to being smart enough to outsmart a human, except in very constrained cases like chess; we did't even have the first clue how to create a computer program to be even remotely dangerous to us. Bostrom's point was that, "We don't need to know the computer program; even if we just simulate something we know works -- a biological brain -- we can reach superintelligence in a few decades." The idea was never that people would actually simulate a cat. The idea is, if we don't think of anything more efficient, we'll at least be able to simulate a cat, and then an idiot, and then Einstein, and then something smarter. And since we almost certainly will think of something more efficient than "simulate a human brain", we should expect superintelligence to come much sooner. > There is no evidence or argument for exponential growth. Moore's law is exponential, which is where the "simulate a brain" predictions have come from. > It is science fiction and leads people to make bad decisions based on fictional evidence. The only "fictional evidence" you've actually specified so far is the fact that there's no biological analog; and that (it seems to me) is from a misunderstanding of a point someone else was making 20 years ago, not something these particular authors are making. I think the case for AI caution looks like this: A. It is possible to create a superintelligent AI B. Progress towards a superintelligent AI will be exponential C. It is possible that a superintelligent AI will want to do something we wouldn't want it to do; e.g., destroy the whole human race D. Such an AI would be likely to succeed. Your skepticism seems to rest on the fundamental belief that either A or B is false: that superintelligence is not physically possible, or at least that progress towards it will be logarithmic rather than exponential. Well, maybe that's true and maybe it's not; but how do you know? What justifies your belief that A and/or B are false so strongly, that you're willing to risk it? And not only willing to risk it, but try to stop people who are trying to think about what we'd do if they are true? What evidence would cause you to re-evaluate that belief, and consider exponential progress towards superintelligence possible? And, even if you think A or B are unlikely, doesn't it make sense to just consider the possibility that they're true, and think about how we'd know and what we could do in response, to prevent C or D? | | |
| ▲ | Vegenoid 2 hours ago | parent | next [-] | | > Moore's law is exponential, which is where the "simulate a brain" predictions have come from. To address only one thing out of your comment, Moore's law is not a law, it is a trend. It just gets called a law because it is fun. We know that there are physical limits to Moore's law. This gets into somewhat shaky territory, but it seems that current approaches to compute can't reach the density of compute power present in a human brain (or other creatures' brains). Moore's law won't get chips to be able to simulate a human brain, with the same amount of space and energy as a human brain. A new approach will be needed to go beyond simply packing more transistors onto a chip - this is analogous to my view that current AI technology is insufficient to do what human brains do, even when taken to their limit (which is significantly beyond where they're currently at). | |
| ▲ | fmap 7 hours ago | parent | prev [-] | | > The idea is, if we don't think of anything more efficient, we'll at least be able to simulate a cat, and then an idiot, and then Einstein, and then something smarter. And since we almost certainly will think of something more efficient than "simulate a human brain", we should expect superintelligence to come much sooner. The problem with this argument is that it's assuming that we're on a linear track to more and more intelligent machines. What we have with LLMs isn't this kind of general intelligence. We have multi-paragraph autocomplete that's matching existing texts more and more closely. The resulting models are great priors for any kind of language processing and have simple reasoning capabilities in so far as those are present in the source texts. Using RLHF to make the resulting models useful for specific tasks is a real achievement, but doesn't change how the training works or what the original training objective was. So let's say we continue along this trajectory and we finally have a model that can faithfully reproduce and identify every word sequence in its training data and its training data includes every word ever written up to that point. Where do we go from here? Do you want to argue that it's possible that there is a clever way to create AGI that has nothing to do with the way current models work and that we should be wary of this possibility? That's a much weaker argument than the one in the article. The article extrapolates from current capabilities - while ignoring where those capabilities come from. > And, even if you think A or B are unlikely, doesn't it make sense to just consider the possibility that they're true, and think about how we'd know and what we could do in response, to prevent C or D? This is essentially https://plato.stanford.edu/entries/pascal-wager/ It might make sense to consider, but it doesn't make sense to invest non-trivial resources. This isn't the part that bothers me at all. I know people who got grants from, e.g., Miri to work on research in logic. If anything, this is a great way to fund some academic research that isn't getting much attention otherwise. The real issue is that people are raising ridiculous amounts of money by claiming that the current advances in AI will lead to some science fiction future. When this future does not materialize it will negatively affect funding for all work in the field. And that's a problem, because there is great work going on right now and not all of it is going to be immediately useful. | | |
| ▲ | gwd 26 minutes ago | parent | next [-] | | > We have multi-paragraph autocomplete that's matching existing texts more and more closely. OK, I think I see where you're coming from. It sounds like what you're saying is: E. LLMs only do multi-paragraph autocomplete; they are and always will be incapable of actual thinking. F. Any approach capable of achieving AGI will be completely different in structure. Who knows if or when this alternate approach will even be developed; and if it is developed, we'll be starting from scratch, so we'll have plenty of time to worry about progress then. With E, again, it may or may not be true. It's worth noting that this is a theoretical argument, not an empirical one; but I think it's a reasonable assumption to start with. However, there are actually theoretical reasons to think that E may be false. The best way to predict the weather is to have an internal model which approximates weather systems; the best way to predict the outcome of a physics problem is to have an internal model which approximates the physics of the thing you're trying to predict. And the best way to predict what a human would write next is to have a model of a human mind -- including a model of what the human mind has in its model (e.g., the state of the world). There is some empirical data to support this argument, albeit in a very simplified manner: They trained a simple LLM to predict valid moves for Othello, and then probed it and discovered an internal Othello board being simulated inside the neural network: https://thegradient.pub/othello/ And my own experience with LLMs better match the "LLMs have an internal model of the world" theory than the "LLMs are simply spewing out statistical garbage" theory. So, with regard to E: Again, sure, LLMs may turn out to be a dead end. But I'd personally give the idea that LLMs are a complete dead end a less than 50% probability; and I don't think giving it an overwhelmingly high probability (like 1 in a million of being false) is really reasonable, given the theoretical arguments and empirical evidence against it. With regard to F, again, I don't think this is true. We've learned so much about optimizing and distilling neural nets, optimizing training, and so on -- not to mention all the compute power we've built up. Even if LLMs are a dead end, whenever we do find an architecture capable of achieving AGI, I think a huge amount of the work we've put into optimizing LLMs will put is way ahead in optimizing this other system. > ...that the current advances in AI will lead to some science fiction future. I mean, if you'd told me 5 years ago that I'd be able to ask a computer, "Please use this Golang API framework package to implement CRUD operations for this particular resource my system has", and that the resulting code would 1) compile out of the box, 2) exhibit an understanding of that resource and how it relates to other resources in the system based on having seen the code implementing those resources 3) make educated guesses (sometimes right, sometimes wrong, but always reasonable) about details I hadn't specified, I don't think I would have believed you. Even if LLM progress is logarithmic, we're already living in a science fiction future. EDIT: The scenario actually has very good technical "asides"; if you want to see their view of how a (potentially dangerous) personality emerges from "multi-paragraph auto-complete", look at the drop-down labelled "Alignment over time", and specifically what follows "Here’s a detailed description of how alignment progresses over time in our scenario:". https://ai-2027.com/#alignment-over-time | |
| ▲ | hannasanarion 4 hours ago | parent | prev [-] | | > So let's say we continue along this trajectory and we finally have a model that can faithfully reproduce and identify every word sequence in its training data and its training data includes every word ever written up to that point. Where do we go from here? This is a fundamental misunderstanding of the entire point of predictive models (and also of how LLMs are trained and tested). For one thing, ability to faithfully reproduce texts is not the primary scoring metric being used for the bulk of LLM training and hasn't been for years. But more importantly, you don't make a weather model so that it can inform you of last Tuesday's weather given information from last Monday, you use it to tell you tomorrow's weather given information from today. The totality of today's temperatures, winds, moistures, and shapes of broader climatic patterns, particulates, albedos, etc etc etc have never happened before, and yet the model tells us something true about the never-before-seen consequences of these never-before-seen conditions, because it has learned the ability to reason new conclusions from new data. Are today's "AI" models a glorified autocomplete? Yeah, but that's what all intelligence is. The next word I type is the result of an autoregressive process occurring in my brain that produces that next choice based on the totality of previous choices and experiences, just like the Q-learners that will kick your butt in Starcraft choose the best next click based on their history of previous clicks in the game combined with things they see on the screen, and will have pretty good guesses about which clicks are the best ones even if you're playing as Zerg and they only ever trained against Terran. A highly accurate autocomplete that is able to predict the behavior and words of a genius, when presented with never before seen evidence, will be able to make novel conclusions in exactly the same way as the human genius themselves would when shown the same new data. Autocomplete IS intelligence. New ideas don't happen because intelligences draw them out of the aether, they happen because intelligences produce new outputs in response to stimuli, and those stimuli can be self-inputs, that's what "thinking" is. If you still think that all today's AI hubbub is just vacuous hype around an overblown autocomplete, try going to Chatgpt right now. Click the "deep research" button, and ask it "what is the average height of the buildings in [your home neighborhood]"?, or "how many calories are in [a recipe that you just invented]", or some other inane question that nobody would have ever cared to write about ever before but is hypothetically answerable from information on the internet, and see if what you get is "just a reproduced word sequence from the training data". |
|
| |
| ▲ | tim333 8 hours ago | parent | prev | next [-] | | >There is no evidence or argument for exponential growth I think the growth you are thinking of, self improving AI, needs the AI to be as smart as a human developer/researcher to get going and we haven't got there yet. But we quite likely will at some point. | | |
| ▲ | maerF0x0 5 hours ago | parent [-] | | and the article specifically mentions the fictional company (clearly designed to generalize the Google/OpenAI's of the world) are supposedly (according to the article) working on building that capability. First by augmenting human researchers, later by augmenting itself. |
| |
| ▲ | whiplash451 7 hours ago | parent | prev | next [-] | | > there are exciting applications which will end up underfunded after the current AI bubble bursts Could you provide examples? I am genuinely interested. | |
| ▲ | whiplash451 7 hours ago | parent | prev | next [-] | | There is no need to simulate Einstein to transform the world with AI. A self-driving car would already be plenty. | | |
| ▲ | skydhash 4 hours ago | parent [-] | | And a self driving car is not even necessary if we’re thinking about solving transportation problems. Train and bus are better at solving road transportation at scale. |
| |
| ▲ | vonneumannstan 6 hours ago | parent | prev [-] | | >All of these "projections" are generalizing from fictional evidence - to borrow a term that's popular in communities that push these ideas. This just isn't correct. Daniel and others on the team are experienced world class forecasters. Daniel wrote another version of this in 2021 predicting the AI world in 2026 and was astonishingly accurate. This deserves credence. https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-... >he arguments back then went something like this: "Machines will be able to simulate brains at higher and higher fidelity. Complete misunderstanding of the underlying ideas. Just in not even wrong territory. >We got some new, genuinely useful tools over the last few years, but this narrative that AGI is just around the corner needs to die. It is science fiction and leads people to make bad decisions based on fictional evidence. You are likely dangerously wrong. The AI field is near universal in predicting AGI timelines under 50 years. With many under 10. This is an extremely difficult problem to deal with and ignoring it because you think it's equivalent to overpopulation on mars is incredibly foolish. https://www.metaculus.com/questions/5121/date-of-artificial-... https://wiki.aiimpacts.org/doku.php?id=ai_timelines:predicti... | | |
| ▲ | loganmhb 5 hours ago | parent | next [-] | | I respect the forecasting abilities of the people involved, but I have seen that report described as "astonishingly accurate" a few times and I'm not sure that's true. The narrative format lends itself somewhat to generous interpretation and it's directionally correct in a way that is reasonably impressive from 2021 (e.g. the diplomacy prediction, the prediction that compute costs could be dramatically reduced, some things gesturing towards reasoning/chain of thought) but many of the concrete predictions don't seem correct to me at all, and in general I'm not sure it captured the spiky nature of LLM competence. I'm also struck by the extent to which the first series from 2021-2026 feels like a linear extrapolation while the second one feels like an exponential one, and I don't see an obvious justification for this. | |
| ▲ | Workaccount2 5 hours ago | parent | prev [-] | | >2025:...Making models bigger is not what’s cool anymore. They are trillions of parameters big already. What’s cool is making them run longer, in bureaucracies of various designs, before giving their answers. Dude was spot on in 2021, hot damn. |
|
| |
| ▲ | bko 8 hours ago | parent | prev | next [-] | | > The problems it raises - alignment, geopolitics, lack of societal safeguards - are all real, and happening now (just replace “AGI” with “corporations”, and voila, you have a story about the climate crisis and regulatory capture). Can you point to the data that suggests these evil corporations are ruining the planet? Carbon emissions are down in every western country since 1990s. Not down per-capita, but down in absolute terms. And this holds even when adjusting for trade (i.e. we're not shipping our dirty work to foreign countries and trading with them). And this isn't because of some regulation or benevolence. It's a market system that says you should try to produce things at the lowest cost and carbon usage is usually associated with a cost. Get rid of costs, get rid of carbon. Other measures for Western countries suggests the water is safer and overall environmental deaths have decreased considerably. The rise in carbon emissions is due to Chine and India. Are you talking about evil Chinese and Indians corporations? https://ourworldindata.org/co2-emissions https://ourworldindata.org/consumption-based-co2 | | |
| ▲ | lordswork 3 hours ago | parent | next [-] | | Emissions are trending downward because of shift from coal to natural gas, growth in renewable energy, energy efficiencies, among other things. Major oil and gas companies in the US like Chevron and ExxonMobil have spent millions on lobbying efforts to resist stricter climate regulations and fight against the changes that led to this trend, so I'd say they are the closest to these evil corporations OP described. Additionally, the current administration refers to doing anything about climate change a "climate religion", so this downward trend will likely slow. The climate regulations are still quite weak. Without a proper carbon tax, a US company can externalize the costs of carbon emissions and get rich by maximizing their own emissions. | |
| ▲ | boh 6 hours ago | parent | prev | next [-] | | Thanks for letting us know everything is fine, just in case we get confused and think the opposite. | | |
| ▲ | bko 5 hours ago | parent [-] | | You're welcome. I know too many upper middle class educated people that don't want to have kids because they believe the earth will cease to be inhabitable in the next 10 years. It's really bizarre to see and they'll almost certainly regret it when they wake up one day alone in a nursing home, look around and realize that the world still exists. And I think the neuroticism around this topic has led young people into some really dark places (anti-depressants, neurotic anti social behavior, general nihilism). So I think it's important to fight misinformation about end of world doomsday scenarios with both facts and common sense. | | |
| ▲ | WXLCKNO 2 hours ago | parent [-] | | I think you're discrediting yourself by talking about dark places and opening your parentheses with anti-depressants. Not all brains function like they're supposed to, people getting help they need shouldn't be stigmatized. You also make no argument about your take on things being the right one, you just oppose their worldview to yours and call theirs wrong like you know it is rather than just you thinking yours is right. | | |
|
| |
| ▲ | ktusznio 6 hours ago | parent | prev | next [-] | | He must be talking about the good, benevolent Western corporations that have outsourced their carbon emissions to the evil and greedy Chinese and Indian corporations. | | | |
| ▲ | philipwhiuk 6 hours ago | parent | prev | next [-] | | > Can you point to the data that suggests these evil corporations are ruining the planet? Can you point to data that this is 'because' of corporations rather than despite them. | | |
| ▲ | om8 6 hours ago | parent [-] | | Burden of proof lies on you, since you mentioned corporations first |
| |
| ▲ | jplusequalt 6 hours ago | parent | prev [-] | | I think a healthy amount of skepticism is warranted when reading about the "reduction" of carbon emissions by companies. Why should we take them at their word when they have a vested interest in fudging the numbers? | | |
| ▲ | bko 6 hours ago | parent [-] | | Carbon emissions are monitored by dozens of independent agencies in many different ways over decades. It would be a giant scale coordination of suppression. Do you have a source that suggests carbon emissions from Western nations is rising? |
|
| |
| ▲ | torginus 9 hours ago | parent | prev | next [-] | | The most amusing thing about is the unshakable belief that any part of humanity will be able to build a single nuclear reactor by 2027 to power datacenters, let alone a network of them. | |
| ▲ | kelsey978126 8 hours ago | parent | prev | next [-] | | bingo. many don't realize superintelligence exists today already, in the form of human super intelligence. artificial super intelligence is already here too, but just as hybrid human machine workloads. Fully automated super intelligence is no different from a corporation, a nation state, a religion. When does it count as ASI? when the chief executive is an AI? Or when they use AI to make decisions? Does it need to be at the board level? We are already here, all this changes is what labor humans will do and how they do it, not the amount. | |
| ▲ | api 10 hours ago | parent | prev | next [-] | | You don’t just beat around the bush here. You actually beat the bush a few times. Large corporations, governments, institutionalized churches, political parties, and other “corporate” institutions are very much like a hypothetical AGI in many ways: they are immortal, sleepless, distributed, omnipresent, and possess beyond human levels of combined intelligence, wealth, and power. They are mechanical Turk AGIs more or less. Look at how humans cycle in, out, and through them, often without changing them much, because they have an existence and a weird kind of will independent of their members. A whole lot, perhaps all, of what we need to do to prepare for a hypothetical AGI that may or may not be aligned consists of things we should be doing to restrain and ensure alignment of the mechanical Turk variety. If we can’t do that we have no chance against something faster and smarter. What we have done over the past 50 years is the opposite: not just unchain them but drop any notion that they should be aligned. Are we sure the AI alignment discourse isn’t just “occulted” progressive political discourse? Back when they burned witches philosophers would encrypt possibly heretical ideas in the form of impenetrable nonsense, which is where what we call occultism comes from. You don’t get burned for suggesting steps to align corporate power, but a huge effort has been made to marginalize such discourse. Consider a potential future AGI. Imagine it has a cult of followers around it, which it probably would, and champions that act like present day politicians or CEOs for it, which it probably would. If it did not get humans to do these things for it, it would have analogous functions or parts of itself. Now consider a corporation or other corporate entity that has all those things but replace the AGI digital brain with a committee or shareholders. What, really, is the difference? Both can be dangerously unaligned. Other than perhaps in magnitude? The real digital AGI might be smarter and faster but that’s the only difference I see. | | |
| ▲ | brookst 9 hours ago | parent [-] | | I looked but I couldn’t find any evidence that “occultism” comes from encryption of heretical ideas. It seems to have been popularized in renaissance France to describe the study of hidden forces. I think you may be hallucinating here. | | |
| |
| ▲ | nroets 14 hours ago | parent | prev | next [-] | | I fail to see how corporations are responsible for the climate crisis: Politicians won't tax gas because they'll get voted out. We know that Trump is not captured by corporations because his trade policies are terrible. If anything, social media is the evil that's destroying the political center: Americans are no longer reading mainstream newspapers or watching mainstream TV news. The EU is saying the elections in Romania was manipulated through manipulation of TikTok accounts and media. | | |
| ▲ | baq 13 hours ago | parent | next [-] | | If you put a knife in someone’s heart, you’re the one who did it and ultimately you’re responsible. If someone told you to do it and you were just following orders… you still did it. If you say there were no rules against putting knives in other people’s hearts, you still did it and you’re still responsible. If it’s somehow different for corporations, please enlighten me how. | | |
| ▲ | nroets 13 hours ago | parent [-] | | The oil companies are saying their product is vital to the economy and they are not wrong. How else will we get food from the farms to the store ? Ambulances to the hospitals ? And many, many other things. Taxes are the best way to change behaviour (smaller cars driving less. Less flying etc). So government and the people who vote for them is to blame. | | |
| ▲ | fire_lake 12 hours ago | parent | next [-] | | What if people are manipulated by bot farms and think tanks and talking points supported by those corporations? I think this view of humans - that they look at all the available information and then make calm decisions in their own interests - is simply wrong. We are manipulated all the damn time. I struggle to go to the supermarket without buying excess sugar. The biggest corporations in the world grew fat off showing us products to impulse buy before our more rational brain functions could stop us. We are not a little pilot in a meat vessel. | | |
| ▲ | nroets 10 hours ago | parent [-] | | Corporations would prefer lower corporate tax. US corporate tax rates are actually every high. Partly due to the US having almost no consumption tax. EU members have VAT etc. |
| |
| ▲ | brookst 9 hours ago | parent | prev | next [-] | | The oil companies also knew and lied about global warming for decades. They paid and continue to pay for as science to stall action. I am completely mystified how you can find them blameless for venal politicians and a populace that largely believes their lies. | |
| ▲ | baq 12 hours ago | parent | prev | next [-] | | I agree with everything here, we've had a great run of economic expansion for basically two centuries and I like my hot showers as much as anyone - but that doesn't change the CO2 levels. | |
| ▲ | matthewdgreen 10 hours ago | parent | prev [-] | | There are politicians in multiple states trying to pass laws that slow down the deployment of renewable energy because they’re afraid if they don’t intervene it will be deployed too quickly and harm fossil fuel interests. Trump is promising to bring back coal, while he bans new wind leases. The whole “oil is the only way aw shucks people chose it” shtick is like a time capsule from 1990. That whole package of beliefs served its purpose and has been replaced with a muscular state-sponsored plan to defend fossil fuel interests even as they become economically obsolete and the rest of the world moves on. |
|
| |
| ▲ | netsharc 13 hours ago | parent | prev | next [-] | | > Politicians won't tax gas because they'll get voted out. I wonder if that's corporations' fault after all: shitty working conditions and shitty wages, so that Bezos can afford to send penises into space. What poor person would agree to higher tax on gas? And the corps are the ones backing politicians who'll propagandize that "Unions? That's communism! Do you want to be Chaina?!" (and spread by those dickheads on the corporate-owned TV and newspaper, drunk dickheads who end up becoming defense secretary) | | |
| ▲ | nroets 12 hours ago | parent [-] | | When people have more money, they tend to buy larger cars that they drive further. Flying is also a luxury. So corporations are involved in the sense that they pay people more than a living wage. |
| |
| ▲ | sofixa 12 hours ago | parent | prev [-] | | > Politicians won't tax gas because they'll get voted out. Have you seen gas tax rates in the EU? > We know that Trump is not captured by corporations because his trade policies are terrible. Unless you think it's a long con for some rich people to be able to time the market by getting him to crash it. > The EU is saying the elections in Romania was manipulated through manipulation of TikTok accounts and media. More importantly, Romanian courts say that too. And it was all out in the open, so not exactly a secret | | |
| ▲ | lucianbr 12 hours ago | parent [-] | | Romainan courts say all kinds of things, many of them patently false. It's absurd to claim that since romanian courts say something, it must be true. It's absurd in principle, because there's nothing in the concept of a court that makes it infallible, and it's absurd in this precise case, because we are corrupt as hell. I'm pretty sure the election was manipulated, but the court only said so because it benefits the incumbents, which control the courts and would lose their power. It's a struggle between local thieves and putin, that's all. The local thieves will keep us in the EU, which is much better than the alternative, but come on. "More importantly, Romanian courts say so"? Really? | | |
| ▲ | sofixa 6 hours ago | parent [-] | | > I'm pretty sure the election was manipulated, but the court only said so because it benefits the incumbents, which control the courts and would lose their power. Why do you think that's the only reason the court said so? The election law was pretty blatantly violated (he declared campaign funding of 0, yet tons of ads were bought for him and influencers paid to advertise him). |
|
|
| |
| ▲ | andrepd 8 hours ago | parent | prev | next [-] | | You said it right, science fiction. Honestly is exactly the tenor I would expect from the AI hype: this text is completely bereft of any rigour while being dressed up in scientific language. There's no evidence, nothing to support their conclusions, no explanation based on data or facts or supporting evidence. It's purely vibes based. Their promise is unironically "the CEOs of AI companies say AGI is 3 years away"! But it's somehow presented as this self important study! Laughable. But it's par on course. Write prompts for LLMs to compete? It's prompt engineering. Tell LLMs to explain their "reasoning" (lol)? It's Deep Research Chain Of Thought. Etc. | | |
| ▲ | somebodythere 5 hours ago | parent [-] | | Did you see the supplemental material that explains how they arrived at their timelines/capabilities forecasts? https://ai-2027.com/research | | |
| ▲ | A_D_E_P_T an hour ago | parent [-] | | It's not at all clear that performance rises with compute in a linear way, which is what they seem to be predicting. GPT-4.5 isn't really that much smarter than 2023's GPT-4, nor is it at all smarter than DeepSeek. There might be (strongly) diminishing returns past a certain point. Most of the growth in AI capabilities has to do with improving the interface and giving them more flexibility. For e.g., uploading PDFs. Further: OpenAI's "deep research" which can browse the web for an hour and summarize publicly-available papers and studies for you. If you ask questions about those studies, though, it's hardly smarter than GPT-4. And it makes a lot of mistakes. It's like a goofy but earnest and hard-working intern. |
|
| |
| ▲ | YetAnotherNick 10 hours ago | parent | prev | next [-] | | > very real risk of societal collapse or species extinction No, there is no risk of species extinction in the near future due to climate change and repeating the line will just further the divide and make the people not care about other people's and even real climate scientist's words. | | |
| ▲ | Aeolun 10 hours ago | parent | next [-] | | Don’t say the things people don’t want to hear and everything will be fine? That sounds like the height of folly. | | |
| ▲ | YetAnotherNick 2 hours ago | parent [-] | | Don't say false things. Especially if it is political and there isn't any way to debate it. |
| |
| ▲ | ttw44 10 hours ago | parent | prev [-] | | The risk is a quantifiable 0.0%? I find that hard to believe. I think the current trends suggest there is a risk that continued environmental destruction could annihilate society. | | |
| ▲ | brookst 9 hours ago | parent | next [-] | | Risk can never be zero, just like certainty can never be 100%. There is a non-zero chance that the ineffable quantum foam will cause a mature hippopotamus to materialize above your bed tonight, and you’ll be crushed. It is incredibly, amazingly, limits-of-math unlikely. Still a non-zero risk. Better to think of “no risk” as meaning “negligible risk”. But I’m with you that climate change is not a negligible risk; maybe way up in the 20% range IMO. And I wouldn’t be sleeping in my bed tonight if sudden hippos over beds were 20% risks. | | |
| ▲ | ttw44 9 hours ago | parent [-] | | Lol, I've always loved that about physics. Some boltzmann brain type stuff. |
| |
| ▲ | SpicyLemonZest 2 hours ago | parent | prev [-] | | It's hard to produce a quantifiable chance of human extinction in the absence of any model by which climate change would lead to it. No climate organization I'm aware of evaluates the end of humanity as even a worst-case risk; the idea simply doesn't exist outside the realm of viral Internet misinformation. |
|
| |
| ▲ | bsenftner 12 hours ago | parent | prev [-] | | Whatever the future is, it is not American, not the United States. The US's cultural individualism has been Capitalistically weaponized, and the educational foundation to take the country forward is not there. The US is kaput, and we are merely observing the ugly demise. The future is Asia, with all of western culture going down. Yes, it is not pretty, The failed experiment of American self rule. | | |
| ▲ | treis 7 hours ago | parent | next [-] | | People said the same thing about Japan but they ran into their own structural issues. It's going to happen to China as well. They've got demographic problems, rule of law problems, democracy problems, and on and on. | | |
| ▲ | nthingtohide 5 hours ago | parent [-] | | I really don't understand this : us vs them viewpoint. Here's a fictional scenario. Imagine Yellowstone erupts tomorrow and whole of America becomes inhabitable but Africa is unscathed. Now think about this, if America had "really" developed African continent, wouldn't it provide shelter to scurrying Americans. Many people forget, the real value of money is in what you can exchange it for. Having skilled people and associated RnD and subsequent products / services is what should have been encouraged by the globalists instead of just rent extraction or stealing. I don't understand the ultimate endgame for globalists. Do each of them desire to have 100km yacht with helicopter perched on it to ferry them back and forth? |
| |
| ▲ | brookst 9 hours ago | parent | prev | next [-] | | I agree but see it as less dire. All of western culture is not ending; it will be absorbed into a more Asia-dominated culture in much he was Asian culture was subsumed into western for the past couple of hundred years. And if Asian culture is better educated and more capable of progress, that’s a good thing. Certainly the US has announced loud and clear that this is the end of the line for us. | | |
| ▲ | rchaud 7 hours ago | parent | next [-] | | > it will be absorbed into a more Asia-dominated culture in much he was Asian culture was subsumed into western for the past couple of hundred years. Was Asian culture dominated by the west to any significant degree? Perhaps in countries like India where the legal and parliamentary system installed by the British remained intact for a long time post-independence. Elsewhere in East and Southeast Asia, the legal systems, education, cultural traditions, and economic philosophies have been very different from the "west", i.e. post-WWII US and Western Europe. The biggest sign of this is how they developed their own information networks, infrastructure and consumer networking devices. Europe had many of these regional champions themselves (Phillips, Nokia, Ericsson, etc) but now outside of telecom infrastructure, Europe is largely reliant on American hardware and software. | |
| ▲ | bsenftner 8 hours ago | parent | prev [-] | | Of course it will not end, western culture just will no longer lead. Despite the sky falling perspective of many, it is simply an attitude adjustment. So one group is no longer #1, and the idea that I was part of that group, ever, was an illusion of propaganda anyway. Life will go on, surprisingly the same. | | |
| ▲ | nthingtohide 5 hours ago | parent [-] | | here's an example. https://x.com/RnaudBertrand/status/1901133641746706581 I finally watched Ne Zha 2 last night with my daughters. It absolutely lives up to the hype: undoubtedly the best animated movie I've ever seen (and I see a lot, the fate of being the father of 2 young daughters ). But what I found most fascinating was the subtle yet unmistakable geopolitical symbolism in the movie. Warning if you haven't yet watched the movie: spoilers! So the story is about Ne Zha and Ao Bing, whose physical bodies were destroyed by heavenly lightning. To restore both their forms, they must journey to the Chan sect—headed by Immortal Wuliang—and pass three trials to earn an elixir that can regenerate their bodies. The Chan sect is portrayed in an interesting way: a beacon of virtue that all strive to join. The imagery unmistakably refers to the US: their headquarters is an imposingly large white structure (and Ne Zha, while visiting it, hammers the point: "how white, how white, how white") that bears a striking resemblance to the Pentagon in its layout. Upon gaining membership to the Chan sect, you receive a jade green card emblazoned with an eagle that bears an uncanny resemblance to the US bald eagle symbol. And perhaps most telling is their prized weapon, a massive cauldron marked with the dollar sign... Throughout the movie you gradually realize, in a very subtle way, that this paragon of virtue is, in fact, the true villain of the story. The Chan sect orchestrates a devastating attack on Chentang Pass—Ne Zha's hometown—while cunningly framing the Dragon King of the East Sea for the destruction. This manipulation serves their divide-and-conquer strategy, allowing them to position themselves as saviors while furthering their own power. One of the most pointed moments comes when the Dragon King of the East Sea observes that the Chan sect "claims to be a lighthouse of the world but harms all living beings." Beyond these explicit symbols, I was struck by how the film portrays the relationships between different groups. The dragons, demons, and humans initially view each other with suspicion, manipulated by the Chan sect's narrative. It's only when they recognize their common oppressor that they unite in resistance and ultimately win. The Chan sect's strategy of fostering division while presenting itself as the arbiter of morality is perhaps the key message of the movie: how power can be maintained through control of the narrative. And as the story unfolds, Wuliang's true ambition becomes clear: complete hegemony. The Chan sect doesn't merely seek to rule—it aims to establish a system where all others exist only to serve its interests, where the dragons and demons are either subjugated or transformed into immortality pills in their massive cauldron. These pills are then strategically distributed to the Chan sect's closest allies (likely a pointed reference to the G7). What makes Ne Zha 2 absolutely exceptional though is that these geopolitical allegories never overshadow the emotional core of the story, nor its other dimensions (for instance it's at times genuinely hilariously funny). This is a rare film that makes zero compromise, it's both a captivating and hilarious adventure for children and a nuanced geopolitical allegory for adults. And the fact that a Chinese film with such unmistakable anti-American symbolism has become the highest-grossing animated film of all time globally is itself a significant geopolitical milestone. Ne Zha 2 isn't just breaking box office records—it's potentially rewriting the rules about what messages can dominate global entertainment. |
|
| |
| ▲ | tim333 8 hours ago | parent | prev [-] | | Perhaps but on the AI front most of the leading research has been in the US or UK, with China being a follower. |
|
|
|
| ▲ | ivraatiems 20 hours ago | parent | prev | next [-] |
| Though I think it is probably mostly science-fiction, this is one of the more chillingly thorough descriptions of potential AGI takeoff scenarios that I've seen. I think part of the problem is that the world you get if you go with the "Slowdown"/somewhat more aligned world is still pretty rough for humans: What's the point of our existence if we have no way to meaningfully contribute to our own world? I hope we're wrong about a lot of this, and AGI turns out to either be impossible, or much less useful than we think it will be. I hope we end up in a world where humans' value increases, instead of decreasing. At a minimum, if AGI is possible, I hope we can imbue it with ethics that allow it to make decisions that value other sentient life. Do I think this will actually happen in two years, let alone five or ten or fifty? Not really. I think it is wildly optimistic to assume we can get there from here - where "here" is LLM technology, mostly. But five years ago, I thought the idea of LLMs themselves working as well as they do at speaking conversational English was essentially fiction - so really, anything is possible, or at least worth considering. "May you live in interesting times" is a curse for a reason. |
| |
| ▲ | lm28469 11 hours ago | parent | next [-] | | > Slowdown"/somewhat more aligned world is still pretty rough for humans: What's the point of our existence if we have no way to meaningfully contribute to our own world? We spend the best 40 years of our lives working 40-50 hours a week to enrich the top 0.1% while living in completely artificial cities. People should wonder what is the point of our current system instead of worrying about Terminator tier sci fi system that may or may not come sometimes in the next 5 to 200 years | | |
| ▲ | anonzzzies 10 hours ago | parent | next [-] | | A lot of people in my surroundings are not buying this life anymore; especially young people are asking why would they. Unlike in the US, they won't end up under a bridge (unless some real collapse, which can of course happen but why worry about it; it might not) so they work simple jobs (data entry or whatnot) to make enough money to eat and party and nothing more. Meaning many of them work no more than a few hours a month. They live rent free at their parents and when they have kids they stop partying but generally don't go work more (well; raising kids is hard work of course but I mean for money). Many of them will inherit the village house from their parents and have a garden so they grow stuff to eat , have some animals and make their own booze so they don't have to pay for that. In cities, people feel the same 'who would I work for the ferrari of the boss we never see', but it is much harder to not to; more expensive and no land and usually no property to inherit (as that is in the countryside or was already sold to not have to work for a year or two). Like you say, people but more our govs need to worry about what is the point at this moment, not scifi in the future; this stuff has already bad enough to worry about. Working your ass off for diminishing returns , paying into a pension pot that won't make it until you retire etc is driving people to really focus on the now and why they would do these things. If you can just have fun with 500/mo and booze from your garden, why work hard and save up etc. I noticed even people from my birth country with these sentiments while they have it extraordinarily good for the eu standards but they are wondering why would they do all of this for nothing (...) more and more and cutting hours more and more. It seems more an education and communication thing really than anything else; it is like asking why pay taxes: if you are not well informed, it might feel like theft, but when you spell it out, most people will see how they benefit. | |
| ▲ | brookst 9 hours ago | parent | prev [-] | | Well said. I keep reading these fearmongering articles and looking around wondering where all of these deep meaning and human agency is today. I’m led to believe that we see this stuff because the tiny subset of humanity that has the wealth and luxury to sit around thinking about thinking about themselves are worried that AI may disrupt the navel-gazing industry. |
| |
| ▲ | zdragnar 15 hours ago | parent | prev | next [-] | | > What's the point of our existence if we have no way to meaningfully contribute to our own world? You may find this to be insightful:
https://meltingasphalt.com/a-nihilists-guide-to-meaning/ In short, "meaning" is a contextual perception, not a discrete quality, though the author suggests it can be quantified based on the number of contextual connections to other things with meaning. The more densely connected something is, the more meaningful it is; my wedding is meaningful to me because my family and my partners family are all celebrating it with me, but it was an entirely meaningless event to you. Thus, the meaningfulness of our contributions remains unchanged, as the meaning behind them is not dependent upon the perspective of an external observer. | | |
| ▲ | lo_zamoyski 7 hours ago | parent | next [-] | | People talk about meaning, but they rarely define it. Ultimately, "meaning" is a matter of "purpose", and purpose is a matter of having an end, or telos. The end of a thing is dependent on the nature of a thing. Thus, the telos of an oak tree is different from the telos of a squirrel which is different from that of a human being. The telos or end of a thing is a marker of the thing's fulfillment or actualization as the kind of thing it is. A thing's potentiality is structured and ordered toward its end. Actualization of that potential is good, the frustration of actualization is not. As human beings, what is most essential to us is that we are rational and social animals. This is why we are miserable when we live lives that are contrary to reason, and why we need others to develop as human beings. The human drama, the human condition, is, in fact, our failure to live rationally, living beneath the dignity of a rational agent, and very often with knowledge of and assent to our irrational deeds. That is, in fact, the very definition of sin: to choose to act in a way one knows one should not. Mistakes aren't sins, even if they are per se evil, because to sin is to knowingly do what you should not (though a refusal to recognize a mistake or to pay for a recognized mistake would constitute a sin). This is why premeditated crimes are far worse than crimes of passion; the first entails a greater knowledge of what one is doing, while someone acting out of intemperance, while still intemperate and thus afflicted with vice, was acting out of impulse rather fully conscious intent. So telos provides the objective ground for the "meaning" of acts. And as you may have noticed, implicitly, it provides the objective basis for morality. To be is synonymous with good, and actualization of potential means to be more fully. | | |
| ▲ | nthingtohide 5 hours ago | parent [-] | | Meaning is a matter of context. Most of the context resides in the past and future. Ludwig's claim that word's meaning is dependent on how it is used. This applies generally. Daniel Dennett - Information & Artificial Intelligence https://www.youtube.com/watch?v=arEvPIhOLyQ Daniel Dennett bridges the gap between everyday information and Shannon-Weaver information theory by rejecting propositions as idealized meaning units. This fixation on propositions has trapped philosophers in unresolved debates for decades. Instead, Dennett proposes starting with simple biological cases—bacteria responding to gradients—and recognizing that meaning emerges from differences that affect well-being. Human linguistic meaning, while powerful, is merely a specialized case.
Neural states can have elaborate meanings without being expressible in sentences. This connects to AI evolution: "good old-fashioned AI" relied on propositional logic but hit limitations, while newer approaches like deep learning extract patterns without explicit meaning representation.
Information exists as "differences that make a difference"—physical variations that create correlations and further differences. This framework unifies information from biological responses to human consciousness without requiring translation into canonical propositions. |
| |
| ▲ | ionwake 14 hours ago | parent | prev [-] | | Please don't be offended by my opinion, I mean it in good humour to share some strong disagreements - Im going to give my take after reading your comment and the article which both seem completely OTT ( contextwise regarding my opinions ). >meaning behind them is not dependent upon the perspective of an external observer. (Yes brother like cmon) Regarding the author, I get the impression he grew up without a strong father figure? This isnt ad hominem I just get the feeling of someone who is so confused and lost in life that he is just severely depressed possibly related to his directionless life. He seems so confused he doesn't even take seriously the fact most humans find their own meaning in life and says hes not even going to consider this, finding it futile.( he states this near the top of the article ). I believe his rejection of a simple basic core idea ends up in a verbal blurb which itself is directionless. My opinion ( Which yes maybe more floored than anyones ), is to deal with Mazlows hierarchy, and then the prime directive for a living organism which after survival , which is reproduction. Only after this has been achieved can you then work towards your family community and nation. This may seem trite, but I do believe that this is natural for someone with a relatively normal childhood. My aim is not to disparage, its to give me honest opinion of why I disagree and possible reasons for it. If you disagree with anything I have said please correct me. Thanks for sharing the article though it was a good read - and I did struggle myself with meaning sometimes. | | |
| ▲ | zdragnar 9 hours ago | parent [-] | | To use a counter example, consider Catholic priests who do not marry or raise children. It would be quite the argument indeed to suggest their lives are without meaning or purpose. Aha, you might say, but they hold leadership roles! They have positions of authority! Of course they have meaning, as they wield spiritual responsibility to their community as a fine substitute for the family life they will not have. To that, I suggest looking deeper, at the nuns and monks. To a cynical non-believer, they surely are wanting for a point to their existence, but to them, what they do is a step beyond Maslow's self actualization, for they live in communion with God and the saints. Their medications and good works in the community are all expressions of that purpose, not the other way around. In short, though their "graph of contextual meaning" doesn't spread as far, it is very densely packed indeed. Two final thoughts: 1) I am both aware of and deeply amused by the use of priests and nuns and monks to defend the arguments of a nihilist's search for meaning. 2) I didn't bring this up so much to take the conversation off topic, so much as to hone in on the very heart of what troubled the person I originally responded to. The question of purpose, the point of existence, in the face of superhuman AI is in fact unchanged. The sense of meaning and purpose one finds in life is found not in the eyes of an unfeeling observer, whether the observers are robots or humans. It must come from within. |
|
| |
| ▲ | joshdavham 18 hours ago | parent | prev | next [-] | | > I hope we're wrong about a lot of this, and AGI turns out to either be impossible, or much less useful than we think it will be. For me personally, I hope that we do get AGI. I just don't want it by 2027. That feels way too fast to me. But AGI 2070 or 2100? That sounds much more preferable. | |
| ▲ | TheDong 17 hours ago | parent | prev | next [-] | | > What's the point of our existence if we have no way to meaningfully contribute to our own world? For a sizable number of humans, we're already there. The vast majority of hacker news users are spending their time trying to make advertisements tempt people into spending money on stuff they don't need. That's an active societal harm. It doesn't contribute in any positive way to the world. And yet, people are fine to do that, and get their dopamine hits off instagram or arguing online on this cursed site, or watching TV. More people will have bullshit jobs in this SF story, but a huge number of people already have bullshit jobs, and manage to find a point in their existence just fine. I, for one, would be happy to simply read books, eat, and die. | | |
| ▲ | bshacklett 7 hours ago | parent | next [-] | | I was hoping someone would bring up Bullshit Jobs. There are definitely a lot of people spending the majority of their time doing "work" that doesn't have any significant impact to the world already. I don't know that some future AI takeover would really change much, except maybe remove some vale of perception around meaningless work. At the same time, I wouldn't necessarily say that people are currently fine getting dopamine hits from social media. Coping would probably be a better description. There are a lot of social and societal problems that have been growing at a rapid rate since Facebook and Twitter began tapping into the reward centers of the brain. From a purely anecdotal perspective, I find my mood significantly affected by how productive and impactful I am with how I spend my time. I'm much happier when I'm making progress on something, whether it's work or otherwise. | |
| ▲ | john_texas 17 hours ago | parent | prev [-] | | Targeted advertising is about determining and giving people exactly what they need. If successful, this increases consumption and grows the productivity of the economy. It's an extremely meaningful job as it allows for precise, effective distribution of resources. | | |
| ▲ | the_gipsy 9 hours ago | parent [-] | | In practice you're just selling shittier or unnecessary stuff. Advertising makes society objectively worse. |
|
| |
| ▲ | abraxas 20 hours ago | parent | prev | next [-] | | I think LLM or no LLM the emergence of intelligence appears to be closely related to the number of synapses in a network whether a biological or a digital one. If my hypothesis is roughly true it means we are several orders of magnitude away from AGI. At least the kind of AGI that can be embodied in a fully functional robot with the sensory apparatus that rivals the human body.
In order to build circuits of this density it's likely to take decades. Most probably transistor based, silicon based substrate can't be pushed that far. | | |
| ▲ | joshjob42 17 hours ago | parent | next [-] | | I think generally the expectation is that there are around 100T synapses in the brain, and of course it's probably not a 1:1 correspondence with neural networks, but it doesn't seem infeasible at all to me that a dense-equivalent 100T parameter model would be able to rival the best humans if trained properly. If basically a transformer, that means it needs at inference time ~200T flops per token. The paper assumes humans "think" at ~15 tokens/second which is about 10 words, similar to the reading speed of a college graduate. So that would be ~3 petaflops of compute per second. Assuming that's fp8, an H100 could do ~4 petaflops, and the authors of AI 2027 guesstimate that purpose wafer scale inference chips circa late 2027 should be able to do ~400petaflops for inference, ~100 H100s worth, for ~$600k each for fabrication and installation into a datacenter. Rounding that basically means ~$6k would buy you the compute to "think" at 10 words/second. Generally speaking that'd probably work out to maybe $3k/yr after depreciation and electricity costs, or ~30-50¢/hr of "human thought equivalent" 10 words/second. Running an AI at 50x human speed 24/7 would cost ~$23k/yr, so 1 OpenBrain researcher's salary could give them a team of ~10-20 such AIs running flat out all the time. Even if you think the AI would need an "extra" 10 or even 100x in terms of tokens/second to match humans, that still puts you at genius level AIs in principle runnable at human speed for 0.1 to 1x the median US income. There's an open question whether training such a model is feasible in a few years, but the raw compute capability at the chip level to plausibly run a model that large at enormous speed at low cost is already existent (at the street price of B200's it'd cost ~$2-4/hr-human-equivalent). | | |
| ▲ | brookst 9 hours ago | parent [-] | | Excellent back of napkin math and it feels intuitively right. And I think training is similar — training is capital intensive therefore centralized, but if 100m people are paying $6k for their inference hardware, add on $100/year as a training tax (er, subscription) and you’ve got $10B/year for training operations. |
| |
| ▲ | nopinsight 18 hours ago | parent | prev | next [-] | | If by “several” orders of magnitude, you mean 3-5, then we might be there by 2030 or earlier. https://situational-awareness.ai/from-gpt-4-to-agi/ | |
| ▲ | ivraatiems 20 hours ago | parent | prev | next [-] | | I think there is a good chance you are roughly right. I also think that the "secret sauce" of sapience is probably not something that can be replicated easily with the technology we have now, like LLMs. They're missing contextual awareness and processing which is absolutely necessary for real reasoning. But even so, solving that problem feels much more attainable than it used to be. | | |
| ▲ | throwup238 17 hours ago | parent | next [-] | | I think the missing secret sauce is an equivalent to neuroplasticity. Human brains are constantly being rewired and optimized at every level: synapses and their channels undergo long term potentiation and depression, new connections are formed and useless ones pruned, and the whole system can sometimes remap functions to different parts of the brain when another suffers catastrophic damage. I don’t know enough about the matrix multiplication operations that power LLMs, but it’s hard to imagine how that kind of organic reorganization would be possible with GPUs matmul. It’d require some sort of advanced “self aware” profile guided optimization and not just trial and error noodling with Torch ops or CUDA kernels. I assume that thanks to the universal approximation theorem it’s theoretically possible to emulate the physical mechanism, but at what hardware and training cost? I’ve done back of the napkin math on this before [1] and the number of “parameters” in the brain is at least 2-4 orders of magnitude more than state of the art models. But that’s just the current weights, what about the history that actually enables the plasticity? Channel threshold potentials are also continuous rather than discreet and emulating them might require the full fp64 so I’m not sure how we’re even going to get to the memory requirements in the next decade, let alone whether any architecture on the horizon can emulate neuroplasticity. Then there’s the whole problem of a true physical feedback loop with which the AI can run experiments to learn against external reward functions and the core survival reward function at the core of evolution might itself be critical but that’s getting deep into the research and philosophy on the nature of intelligence. [1] https://news.ycombinator.com/item?id=40313672 | | |
| ▲ | lblume 5 hours ago | parent [-] | | Transformers already are very flexible. We know that we can basically strip blocks at will, reorder modules, transform their input in predictable ways, obstruct some features and they will after a very short period of re-training get back to basically the same capabilities they had before. Fascinating stuff. |
| |
| ▲ | narenm16 19 hours ago | parent | prev [-] | | i agree. it feels like scaling up these large models is such an inefficient route that seems to be warranting new ideas (test-time compute, etc). we'll likely reach a point where it's infeasible for deep learning to completely encompass human-level reasoning, and we'll need neuroscience discoveries to continue progress. altman seems to be hyping up "bigger is better," not just for model parameters but openai's valuation. |
| |
| ▲ | baq 13 hours ago | parent | prev | next [-] | | Exponential growth means the first order of magnitude comes slowly and the last one runs past you unexpectedly. | | |
| ▲ | Palmik 12 hours ago | parent [-] | | Exponential growth generally means that the time between each order of magnitude is roughly the same. | | |
| ▲ | brookst 9 hours ago | parent [-] | | At the risk of pedantry, is that true? Something that doubles annually sure seems like exponential growth to me, but the orders of magnitude are not at all the same rate. Orders of magnitude are a base-10 construct but IMO exponents don’t have to be 10. EDIT: holy crap I just discovered a commonly known thing about exponents and log. Leaving comment here but it is wrong, or at least naive. |
|
| |
| ▲ | UltraSane 18 hours ago | parent | prev [-] | | Why can't the compute be remote from the robot? That is a major advantage of human technology over biology. | | |
| ▲ | abraxas 17 hours ago | parent [-] | | Mostly latency. But even if a single robot could be driven by a data centre consider the energy and hardware investment requirements to make such a creature practical. | | |
| ▲ | UltraSane an hour ago | parent | next [-] | | The Figure robots use a two level control scheme with a fast LLM at 200Hz directly controlling the robot and a slow planning LLM running at 7Hz. This planning LLM could be very far away indeed and still have less than 142.8ms of latency. | |
| ▲ | Jensson 12 hours ago | parent | prev | next [-] | | 1ms latency is more than fast enough, you probably have bigger latency than that between the cpu and the gpu. | | |
| ▲ | Symmetry 9 hours ago | parent [-] | | We've got 10ms of latency between our brains and our hands along our nerve fibers and we function all right. |
| |
| ▲ | UltraSane 17 hours ago | parent | prev [-] | | Latency would be kept low be keeping the compute nearby. One 1U or 2U server per robot would be reasonable. |
|
|
| |
| ▲ | baron816 15 hours ago | parent | prev | next [-] | | My vision for an ASI future involves humans living in simulations that are optimized for human experience. That doesn’t mean we are just live in a paradise and are happy all the time. We’d experience dread and loss and fear, but it would ultimately lead to a deeply satisfying outcome. And we’d be able to choose to forget things, including whether we’re in a simulation so that it feels completely unmistakeable from base reality. You’d live indefinitely, experiencing trillions of lifespans where you get to explore the multiverse inside and out. My solution to the alignment problem is that an ASI could just stick us in tubes deep in the Earth’s crust—it just needs to hijack our nervous system to input signals from the simulation. The ASI could have the whole rest of the planet, or it could move us to some far off moon in the outer solar system—I don’t care. It just needs to do two things for it’s creators—preserve lives and optimize for long term human experience. | |
| ▲ | Davidzheng 6 hours ago | parent | prev | next [-] | | I think two years is entirely reasonable timeline. | |
| ▲ | arisAlexis 11 hours ago | parent | prev [-] | | do you really think that AGI is impossible after all that happened up to today? how is this possible? |
|
|
| ▲ | KaiserPro a day ago | parent | prev | next [-] |
| > AI has started to take jobs, but has also created new ones. Yeah nah, theres a key thing missing here, the number of jobs created needs to be more than the ones it's destroyed, and they need to be better paying and happen in time. History says that actually when this happens, an entire generation is yeeted on to the streets (see powered looms, Jacquard machine, steam powered machine tools) All of that cheap labour needed to power the new towns and cities was created by automation of agriculture and artisan jobs. Dark satanic mills were fed the decedents of once reasonably prosperous crafts people. AI as presented here will kneecap the wages of a good proportion of the decent paying jobs we have now. This will cause huge economic disparities, and probably revolution. There is a reason why the royalty of Europe all disappeared when they did... So no, the stock market will not be growing because of AI, it will be in spite of it. Plus china knows that unless they can occupy most of its population with some sort of work, they are finished. AI and decent robot automation are an existential threat to the CCP, as much as it is to what ever remains of the "west" |
| |
| ▲ | kypro a day ago | parent | next [-] | | > and probably revolution I theorise that revolution would be near-impossible in post-AGI world. If people consider where power comes from it's relatively obvious that people will likely suffer and die on mass if we ever create AGI. Historically the general public have held the vast majority of power in society. 100+ years ago this would have been physical power – the state has to keep you happy or the public will come for them with pitchforks. But in an age of modern weaponry the public today would be pose little physical threat to the state. Instead in todays democracy power comes from the publics collective labour and purchasing power. A government can't risk upsetting people too much because a government's power today is not a product of its standing army, but the product of its economic strength. A government needs workers to create businesses and produce goods and therefore the goals of government generally align with the goals of the public. But in an post-AGI world neither businesses or the state need workers or consumers. In this world if you want something you wouldn't pay anyone for it or workers to produce it for you, instead you would just ask your fleet of AGIs to get you the resource. In this world people become more like pests. They offer no economic value yet demand that AGI owners (wherever publicly or privately owned) share resources with them. If people revolted any AGI owner would be far better off just deploying a bioweapon to humanely kill the protestors rather than sharing resources with them. Of course, this is assuming the AGI doesn't have it's own goals and just sees the whole of humanely as nuance to be stepped over in the same way humans will happy step over animals if they interfere with our goals. Imo humanity has 10-20 years left max if we continue on this path. There can be no good outcome of AGI because it would even make sense for the AGI or those who control the AGI to be aligned with goals of humanity. | | |
| ▲ | Centigonal 17 hours ago | parent | next [-] | | I think "resource curse" countries are a great surrogate for studying possible future AGI-induced economic and political phenomena. A country like the UAE (oil) or Botswana (diamonds) essentially has an economic equivalent to AGI: they control a small, extremely productive utility (an oilfield or a mine instead of a server farm), and the wealth generated by that utility is far in excess of what those countries' leaders need to maintain power. Sure, you hire foreign labor and trade for resources instead of having your AGI supply those things, but the end result is the same. | |
| ▲ | robinhoode 20 hours ago | parent | prev | next [-] | | > In this world people become more like pests. They offer no economic value yet demand that AGI owners (wherever publicly or privately owned) share resources with them. If people revolted any AGI owner would be far better off just deploying a bioweapon to humanely kill the protestors rather than sharing resources with them. This is a very doomer take. The threats are real, and I'm certain some people feel this way, but eliminating large swaths of humanity is something dicatorships have tried in the past. Waking up every morning means believing there are others who will cooperate with you. Most of humanity has empathy for others. I would prefer to have hope that we will make it through, rather than drown in fear. | | |
| ▲ | 758597464 20 hours ago | parent | next [-] | | > This is a very doomer take. The threats are real, and I'm certain some people feel this way, but eliminating large swaths of humanity is something dicatorships have tried in the past. Tried, and succeeded in. In times where people held more power than today. Not sure what point you're trying to make here. > Most of humanity has empathy for others. I would prefer to have hope that we will make it through, rather than drown in fear. I agree that most of humanity has empathy for others — but it's been shown that the prevalence of psychopaths increases as you climb the leadership ladder. Fear or hope are the responses of the passive. There are other routes to take. | | |
| ▲ | bamboozled 12 hours ago | parent [-] | | Basically why open source everything is increasingly more important and imo already making “AI” safer. If the many have access to the latest AI then there is less chance the masses are blindsided by some rogue tech. |
| |
| ▲ | 542354234235 8 hours ago | parent | prev [-] | | >but eliminating large swaths of humanity is something dicatorships have tried in the past. Technology changes things though. Things aren't "the same as it ever was". The Napoleonic wars killed 6.5 million people with muskets and cannons. The total warfare of WWII killed 70 to 85 million people with tanks, turboprop bombers, aircraft carriers, and 36 kilotons TNT of Atomic bombs, among other weaponry. Total war today includes modern thermonuclear weapons. In 60 seconds, just one Ohio class submarine can launch 80 independent warheads, totaling over 36 megatons of TNT. That is over 20 times more than all explosives, used by all sides, for all of WWII, including both Atomic bombs. AGI is a leap forward in power equivalent to what thermonuclear bombs are to warfare. Humans have been trying to destroy each other for all of time but we can only have one nuclear war, and it is likely we can only have one AGI revolt. | | |
| ▲ | jplusequalt 6 hours ago | parent [-] | | I don't understand the psychology of doomerism. Are people truly so scared of these futures they are incapable of imagining an alternate path where anything less than total human extinction occurs? Like if you're truly afraid of this, what are you doing here on HN? Go organize and try to do something about this. | | |
| ▲ | 542354234235 4 hours ago | parent [-] | | I don’t see it as doomerism, just realism. Looking at the realities of nuclear war shows that it is a world ending holocaust that could happen by accident or by the launch of a single nuclear ICBM by North Korea, and there is almost no chance of de-escalation once a missile is in the air. There is nothing to be done, other than advocate of nuclear arms treaties in my own country, but that has no effect on Russia, China, North Korea, Pakistan, India, or Iran. Bertrand Russell said, "You may reasonably expect a man to walk a tightrope safely for ten minutes; it would be unreasonable to do so without accident for two hundred years." We will either walk the tightrope for another 100 years or so until global society progresses to where there is nuclear disarmament, or we won’t. It is the same with Gen AI. We will either find a way to control an entity that rapidly becomes orders of magnitude more intelligent than us, or we won’t. We will either find a way to prevent the rich and powerful from controlling a Gen AI that can build and operate anything they need, including an army to protect them from everyone without a powerful Gen AI, or we won’t. I hope for a future of abundance for all, brought to us by technology. But I understand that some existential threats only need to turn the wrong way once, and there will be no second chance ever. | | |
| ▲ | jplusequalt 4 hours ago | parent [-] | | I think it's a fallacy to equate pessimistic outcomes with "realism" >It is the same with Gen AI. We will either find a way to control an entity that rapidly becomes orders of magnitude more intelligent than us, or we won’t. We will either find a way to prevent the rich and powerful from controlling a Gen AI that can build and operate anything they need, including an army to protect them from everyone without a powerful Gen AI, or we won’t Okay, you've laid out two paths here. What are *you* doing to influence the course we take? That's my point. Enumerating all the possible ways humanity faces extinction is nothing more than doomerism if you aren't taking any meaningful steps to lessen the likelihood any of them may occur. |
|
|
|
| |
| ▲ | wkat4242 20 hours ago | parent | prev | next [-] | | > I theorise that revolution would be near-impossible in post-AGI world. If people consider where power comes from it's relatively obvious that people will likely suffer and die on mass if we ever create AGI. I agree but for a different reason. It's very hard to outsmart an entity with an IQ in the thousands and pervasive information gathering. For a revolution you need to coordinate. The Chinese know this very well and this is why they control communication so closely (and why they had Apple restrict AirDrop). But their security agencies are still beholden to people with average IQs and the inefficient communication between them. An entity that can collect all this info on its own and have a huge IQ to spot patterns and not have to communicate it to convince other people in its organisation to take action, that will crush any fledgling rebellion. It will never be able to reach critical mass. We'll just be ants in an anthill and it will be the boot that crushes us when it feels like it. | |
| ▲ | jplusequalt 6 hours ago | parent | prev | next [-] | | The apathy spewed by doomers actively contributes to the future they whine about. Join a union. Organize with real people. People will always have the power in society. | |
| ▲ | weatherlite 8 hours ago | parent | prev | next [-] | | > In this world people become more like pests. They offer no economic value yet demand that AGI owners (wherever publicly or privately owned) share resources with them. If people revolted any AGI owner would be far better off just deploying a bioweapon to humanely kill the protestors rather than sharing resources with them. That will be quite a hard thing to pull off, even for some evil person with a AGI. Let's say Putin gets AGI and is actually evil and crazy enough to try wipe people out. If he just targets Russians and starts killing millions of people daily with some engineered virus or something similar, he'll have to fear a strike from the West which would be fearful they're next (and rightfully so).
If he instead tries to wipe out all of humanity at once to escape a second strike, he again will have to devise such a good plan there won't be any second strike - meaning his "AGI" will have to be way better than all other competing AGIs (how exactly?). It would have made sense if all "owners of AGI" somehow conspired together to do this but there's not really such a thing as owners of AGI and even if there was Chinese, Russian and American owners of AGI don't trust each other at all and are also bound to their governments. | |
| ▲ | dovin 15 hours ago | parent | prev [-] | | Dogs offer humans no economic value, but we haven't genocided them. There are a lot of ways that we could offer value that's not necessarily just in the form of watts and minerals. I'm not so sure that our future superintelligent summoned demons will be motivated purely by increasing their own power, resources, and leverage. Then again, maybe they will. Thus far, AI systems that we have created seem surprisingly goal-less. I'm more worried about how humans are going to use them than some sort of breakaway event but yeah, don't love that it's a real possible future. | | |
| ▲ | chipsrafferty 15 hours ago | parent [-] | | A world in which most humans fill the role of "pets" of the ultra rich doesn't sound that great. | | |
| ▲ | dovin 15 hours ago | parent [-] | | Humans becoming domesticated by benevolent superintelligences are some of the better futures with superintelligences, in my mind. Iain M Banks' Culture series is the best depiction of this I've come across; they're kind of the utopian rendition of the phrase "all watched over by machines of loving grace". Though it's a little hard to see how we get from here to there. | | |
| ▲ | autumnstwilight 14 hours ago | parent [-] | | Honestly that part of the article and some other comments have given me the idle speculation, what if that was the solution to the, "Humans no longer feel they can meaningfully contribute to the world," issue? Like we can satisfy the hunting and retrieval instincts of dogs by throwing a stick, surely an AI that is 10,000 times more intelligent can devise a stick-retrieval-task for humans in a way that feels like satisfying achievement and meaningful work from our perspective. (Leaving aside the question of whether any of that is a likely or desirable outcome.) | | |
| ▲ | bamboozled 12 hours ago | parent [-] | | What will AI find fulfilling itself? I find that to be quite a deep question. I feel the limitations of humans are quite a feature when you think about what the experience of life would be like if you couldn’t forget or experienced things for the first time. If you already knew everything and you could achieve almost anything with zero effort. It actually sounds…insufferable. | | |
| ▲ | te0006 10 hours ago | parent [-] | | You might find Stanislav Lem's Golem XIV worth a read, in which a what we now call an AGI shares, amongst other things, its knowledge and speculations about long-term evolution of superintelligences, in a lecture to humans, before entering the next stage itself.
https://www.goodreads.com/book/show/10208493
It seems difficult to obtain an English edition these days but there is a reddit thread you might want to look into. |
|
|
|
|
|
| |
| ▲ | OgsyedIE a day ago | parent | prev | next [-] | | Unfortunately the current system is doing a bad job of finding replacements for dwindling crucial resources such as petroleum basins, new generations of workers, unoccupied orbital trajectories, fertile topsoil and copper ore deposits. Either the current system gets replaced with a new system or it doesn't. | |
| ▲ | pydry a day ago | parent | prev | next [-] | | >History says that actually when this happens, an entire generation is yeeted on to the streets History hasnt had to contend with a birth rate of 0.7-1.6. It's kind of interesting that the elite capitalist media (economist, bloomberg, forbes, etc) is projecting a future crisis of both not enough workers and not enough jobs simultaneously. | | |
| ▲ | wkat4242 20 hours ago | parent | next [-] | | I don't really get the American preoccupation with birth rates. We're already way overpopulated for our planet and this is showing in environmental issues, housing cost, overcrowded cities etc. It's totally a great thing if we start plateauing our population and even reduce it a bit. And no we're not going extinct. It'll just cause some temporary issues like an ageing population that has to be cared for but those issues are much more readily fixable than environmental destruction. | | |
| ▲ | NitpickLawyer 14 hours ago | parent | next [-] | | > I don't really get the American preoccupation with birth rates. Japan is currently in the finding out phase of this problem. | |
| ▲ | yoyohello13 20 hours ago | parent | prev | next [-] | | I think it’s more of a “be fruitful and multiply” thing than an actual existential threat thing. You can see many of loudest people talking about it either have religious undertones or want more peasants to work the factories. Demographic shift will certainly upset the status quo, but we will figure out how to deal with it. | |
| ▲ | ahtihn 14 hours ago | parent | prev | next [-] | | The planet is absolutely not over populated. Overcrowded cities and housing costs aren't an overpopulation problem but a problem of concentrating economic activity in certain places. | | | |
| ▲ | torlok 20 hours ago | parent | prev | next [-] | | Don't try to reason with this population collapse nonsense. This has always been about racists fearing that "not enough" white westerners are being born, or about industrialists wanting infinite growth. For some prominent technocrats it's both. | | |
| ▲ | gmoot 18 hours ago | parent [-] | | The welfare state is predicated on a pyramid-shaped population. Also: people deride infinite growth, but growth is what is responsible for lifting large portions of the population out of poverty. If global markets were repriced tomorrow to expect no future growth, economies would collapse. There may be a way to accept low or no growth without economic collapse, but if there is no one has figured it out yet. That's nothing to be cavalier about. | | |
| ▲ | pydry 18 hours ago | parent [-] | | The welfare state isnt predicated on a pyramid shape but the continued growth of the stock market and endless GDP growth certainly is. >infinite growth, but growth is what is responsible for lifting large portions of the population out of poverty It's overstated. The preconditions for GDP growth - namely lack of war and corruption are probably more responsible than the growth itself. |
|
| |
| ▲ | ttw44 4 hours ago | parent | prev | next [-] | | We are not overpopulated. I hate the type of people that hammer the idea that society needs to double or triple the birthrate (Elon Musk), but as it currently stands, countries like South Korea, Japan, USA, China, and Germany risk extinction or economic collapse in 4-5 generations if the birth rate doesn't rise or the way we guarantee welfare doesn't change. | |
| ▲ | luxardo 7 hours ago | parent | prev | next [-] | | We are most certainly not "overpopulated" in any way. Usage per person is what the issue is. And no society, ever, has had a good standard of living with a shrinking population. You are advocating for all young people to toil their entire lives taking care of an ever-aging population. | |
| ▲ | alxjrvs 19 hours ago | parent | prev | next [-] | | Racist fears of "replacement", mostly. | |
| ▲ | chipsrafferty 15 hours ago | parent | prev | next [-] | | It's the only way to increase profits under capitalism in the long term once you've optimized the technology. | |
| ▲ | mattnewton 20 hours ago | parent | prev [-] | | I think a good part of it is fear of a black planet. |
| |
| ▲ | KaiserPro 9 hours ago | parent | prev [-] | | > History hasnt had to contend with a birth rate of 0.7-1.6. I think thats just not true: https://en.wikipedia.org/wiki/Peasants%27_Revolt A large number of revolutions/rebellions are caused by mass unemployment or famine. |
| |
| ▲ | torlok 20 hours ago | parent | prev | next [-] | | Hayek has been lobbied by US corporations so hard for so long that regular people treat the invisible hand of the market like it's gospel. | |
| ▲ | baq 13 hours ago | parent | prev [-] | | > So no, the stock market will not be growing because of AI, it will be in spite of it. The stock market will be one of the very few ways you will be able to own some of that AI… assuming it won’t be nationalized. |
|
|
| ▲ | nfc 24 minutes ago | parent | prev | next [-] |
| Something I ponder in the context of AI alignment is how we approach agents with potentially multiple objectives. Much of the discussion seems focused on ensuring an AI pursues a single goal. Which seems to be a great idea if we are trying to simplify the problem but I'm not sure how realistic it is when considering complex intelligences. For example human motivation often involves juggling several goals simultaneously. I might care about both my own happiness and my family's happiness. The way I navigate this isn't by picking one goal and maximizing it at the expense of the other; instead, I try to balance my efforts and find acceptable trade-offs. I think this 'balancing act' between potentially competing objectives may be a really crucial aspect of complex agency, but I haven't seen it discussed as much in alignment circles. Maybe someone could point me to some discussions about this :) |
|
| ▲ | torginus a day ago | parent | prev | next [-] |
| Much has been made in its article about autonomous agents ability to do research via browsing the web - the web is 90% garbage by weight (including articles on certain specialist topics). And it shows. When I used GPT's deep research to research the topic, it generated a shallow and largely incorrect summary of the issue, owning mostly to its inability to find quality material, instead it ended up going for places like Wikipedia, and random infomercial listicles found on Google. I have a trusty Electronics textbook written in the 80s, I'm sure generating a similarly accurate, correct and deep analysis on circuit design using only Google to help would be 1000x harder than sitting down and working through that book and understanding it. |
| |
| ▲ | Aurornis 21 hours ago | parent | next [-] | | This story isn’t really about agents browsing the web. It’s a fiction about a company that consumes all of the web and all other written material into a model that doesn’t need to browse the web. The agents in this story supersede the web. But your point hits on one of the first cracks to show in this story: We already have companies consuming much of the web and training models on all of our books, but the reports they produce are of mixed quality. The article tries to get around this by imagining models and training runs a couple orders of magnitude larger will simply appear in the near future and the output of those models will yield breakthroughs that accelerate the next rounds even faster. Yet here we are struggling to build as much infrastructure as possible to squeeze incremental improvements out of the next generation of models. This entire story relies on AI advancement accelerating faster in a self-reinforcing way in the coming couple of years. | | |
| ▲ | whiplash451 7 hours ago | parent | next [-] | | In my opinion, the real breakthrough described in this article is not bigger models to read the web, but models that can experiment on their own and learn from these experiments to generate new ideas. If this happens, then we indeed enter a non-linear regime. | |
| ▲ | skywhopper 8 hours ago | parent | prev | next [-] | | That’s exactly why it doesn’t make sense. Where would a datacenter-bound AI get more data about the world exactly? The story is actually quite poorly written, with weird stuff about “oh yeah btw we fixed hallucinations” showing up off-handedly halfway through. And another example of that is the bit where they throw in that one generation is producing scads of synthetic training data for the next gen system. Okay, but once you know everything there is to know based on written material, how do you learn new things about the world? How do you learn how to build insect drones, mass-casualty biological weapons, etc? Is the super AI supposed to have completely understood physics to the extent that it can infer all reality without having to do experimentation? Where does even the electricity to do this come from? Much less the physical materials. The idea that even a supergenius intelligence could drive that much physical change in the world within three years is just silly. | | |
| ▲ | ctoth 6 hours ago | parent [-] | | How will this thing which is connected to the Internet ... get data? |
| |
| ▲ | adastra22 18 hours ago | parent | prev [-] | | There's an old adage in AI: garbage in, garbage out. Consuming and training on the whole internet doesn't make you smarter than the average intelligence of the internet. | | |
| ▲ | drchaos 11 hours ago | parent [-] | | > Consuming and training on the whole internet doesn't make you smarter than the average intelligence of the internet. This is only true as long as you are not able to weigh the quality of a source. Just like getting spam in your inbox may waste your time, but it doesn't make you dumber. |
|
| |
| ▲ | tim333 8 hours ago | parent | prev | next [-] | | I myself am something of an autonomous agent who browses the web and it's possible to be choosy about what you browse. Like I could download some electronics text books off the web rather than going to listicles. LLMs may not be that discriminating at the moment but they could get better. | |
| ▲ | Balgair 8 hours ago | parent | prev | next [-] | | > the web is 90% garbage by weight Sturgeon's law : "Ninety percent of everything is crap" | |
| ▲ | dimitri-vs 20 hours ago | parent | prev | next [-] | | Interesting, I've hard the exact opposite experience. For example I was curious why in metal casting the top box is called the cope and the bottom is called the drag. And it found very niche information and quotes from page 100 in a PDF on some random government website. The whole report was extremely detailed and verifiable if I followed its links. That said I suspect (and am already starting to see) the increased use of anti-bot protection to combat browser use agents. | |
| ▲ | somerandomness a day ago | parent | prev [-] | | Agreed. However, source curation and agents are two different parts of Deep Research. What if you provided that textbook to a reliable agent? Plug: We built https://RadPod.ai to allow you to do that, i.e. Deep Research on your data. | | |
| ▲ | preommr a day ago | parent | next [-] | | So, once again, we're in the era of "There's an [AI] app for that". | |
| ▲ | skeeter2020 a day ago | parent | prev | next [-] | | that might solve your sourcing problem, but now you need to have faith it will draw conclusions and parallels from the material accurately. That seems even harder than the original problem; I'll stick with decent search on quality source material. | | |
| ▲ | somerandomness a day ago | parent [-] | | The solution is a citation mechanism that points you directly where in the source material it comes from (which is what we tried to build). Easy verification is important for AI to have a net-benefit to productivity IMO. |
| |
| ▲ | demadog a day ago | parent | prev [-] | | RadPod - what models do you use to power it? |
|
|
|
| ▲ | beklein a day ago | parent | prev | next [-] |
| Older and related article from one of the authors titled "What 2026 looks like", that is holding up very well against time. Written in mid 2021 (pre ChatGPT) https://www.alignmentforum.org/posts/6Xgy6CAf2jqHhynHL/what-... //edit: remove the referral tags from URL |
| |
| ▲ | samth a day ago | parent | next [-] | | I think it's not holding up that well outside of predictions about AI research itself. In particular, he makes a lot of predictions about AI impact on persuasion, propaganda, the information environment, etc that have not happened. | | |
| ▲ | Aurornis a day ago | parent | next [-] | | Agree. The base claims about LLMs getting bigger, more popular, and capturing people's imagination are right. Those claims are as easy as it gets, though. Look into the specific claims and it's not as amazing. Like the claim that models will require an entire year to train, when in reality it's on the order of weeks. The societal claims also fall apart quickly: > Censorship is widespread and increasing, as it has for the last decade or two. Big neural nets read posts and view memes, scanning for toxicity and hate speech and a few other things. (More things keep getting added to the list.) Someone had the bright idea of making the newsfeed recommendation algorithm gently ‘nudge’ people towards spewing less hate speech; now a component of its reward function is minimizing the probability that the user will say something worthy of censorship in the next 48 hours. This is a common trend in rationalist and "X-risk" writers: Write a big article with mostly safe claims (LLMs will get bigger and perform better!) and a lot of hedging, then people will always see the article as primarily correct. When you extract out the easy claims and look at the specifics, it's not as impressive. This article also shows some major signs that the author is deeply embedded in specific online bubbles, like this: > Most of America gets their news from Twitter, Reddit, etc. Sites like Reddit and Twitter feel like the entire universe when you're embedded in them, but when you step back and look at the numbers only a fraction of the US population are active users. | |
| ▲ | LordDragonfang a day ago | parent | prev | next [-] | | Could you give some specific examples of things you feel definitely did not come to pass? Because I see a lot of people here talking about how the article missed the mark on propaganda; meanwhile I can tab over to twitter and see a substantial portion of the comment section of every high-engagement tweet being accused of being Russia-run LLM propaganda bots. | |
| ▲ | madethisnow a day ago | parent | prev [-] | | something you can't know | | |
| ▲ | elicksaur a day ago | parent [-] | | This doesn’t seem like a great way to reason about the predictions. For something like this, saying “There is no evidence showing it” is a good enough refutation. Counterpointing that “Well, there could be a lot of this going on, but it is in secret.” - that could be a justification for any kooky theory out there. Bigfoot, UFOs, ghosts. Maybe AI has already replaced all of us and we’re Cylons. Something we couldn’t know. The predictions are specific enough that they are falsifiable, so they should stand or fall based on the clear material evidence supporting or contradicting them. |
|
| |
| ▲ | motoxpro a day ago | parent | prev | next [-] | | That's incredible how much it broadly aligns with what has happened. Especially because it was before ChatGPT. | | |
| ▲ | FairlyInvolved a day ago | parent | next [-] | | There's a pretty good summary of how well it has held up here, by the significance of each claim: https://www.lesswrong.com/posts/u9Kr97di29CkMvjaj/evaluating... | |
| ▲ | reducesuffering a day ago | parent | prev [-] | | Will people finally wake up that the AGI X-Risk people have been right and we’re rapidly approaching a really fucking big deal? This forum has been so behind for too long. Sama has been saying this a decade now: “Development of Superhuman machine intelligence is probably the greatest threat to the continued existence of humanity” 2015 https://blog.samaltman.com/machine-intelligence-part-1 Hinton, Ilya, Dario Amodei, RLHF inventor, Deepmind founders. They all get it, which is why they’re the smart cookies in those positions. First stage is denial, I get it, not easy to swallow the gravity of what’s coming. | | |
| ▲ | ffsm8 a day ago | parent | next [-] | | People have been predicting the singularity to occur sometimes around 2030 and 2045 waaaay further back then 2015. And not just by enthusiasts, I dimly remember an interview with Richard Darkins from back in the day... Though that doesn't mean that the current version of language models will ever achieve AGI, and I sincerely doubt they will.
They'll likely be a component in the AI, but likely not the thing that "drives" | | | |
| ▲ | pixl97 a day ago | parent | prev | next [-] | | >This forum has been so behind for too long. There is a strong financial incentive for a lot of people on this site to deny they are at risk from it, or to deny what they are building has risk and they should have culpability from that. | |
| ▲ | hn_throwaway_99 a day ago | parent | prev | next [-] | | > Will people finally wake up that the AGI X-Risk people have been right and we’re rapidly approaching a really fucking big deal? OK, say I totally believe this. What, pray tell, are we supposed to do about it? Don't you at least see the irony of quoting Sama's dire warnings about the development of AI, without at least mentioning that he is at the absolute forefront of the push to build this technology that can destroy all of humanity. It's like he's saying "This potion can destroy all of humanity if we make it" as he works faster and faster to figure out how to make it. I mean, I get it, "if we don't build it, someone else will", but all of the discussion around "alignment" seems just blatantly laughable to me. If on one hand your goal is to build "super intelligence", i.e. way smarter than any human or group of humans, how do you expect to control that super intelligence when you're just acting at the middling level of human intelligence? While I'm skeptical on the timeline, if we do ever end up building super intelligence, the idea that we can control it is a pipe dream. We may not be toast (I mean, we're smarter than dogs, and we keep them around), but we won't be in control. So if you truly believe super intelligent AI is coming, you may as well enjoy the view now, because there ain't nothing you or anyone else will be able to do to "save humanity" if or when it arrives. | | |
| ▲ | ctoth 6 hours ago | parent | next [-] | | I love this pattern, the oldest pattern. There is nothing happening! The thing that is happening is not important! The thing that is happening is important, but it's too late to do anything about it! Well, maybe if you had done something when we first started warning about this... See also: Covid/Climate/Bird Flu/the news. | |
| ▲ | achierius a day ago | parent | prev [-] | | Political organization to force a stop to ongoing research? Protest outside OAI HQ? There are lots of thing we could, and many of us would, do if more people were actually convinced their life were in danger. | | |
| ▲ | hn_throwaway_99 a day ago | parent [-] | | > Political organization to force a stop to ongoing research? Protest outside OAI HQ? Come on, be real. Do you honestly think that would make a lick of difference? Maybe, at best, delay things by a couple months. But this is a worldwide phenomenon, and humans have shown time and time again that they are not able to self organize globally. How successful do you think that political organization is going to be in slowing China's progress? | | |
| ▲ | achierius 16 hours ago | parent | next [-] | | Humans have shown time and time again that they are able to self-organize globally. Nuclear deterrence -- human cloning -- bioweapon proliferation -- Antarctic neutrality -- the list goes on. > How successful do you think that political organization is going to be in slowing China's progress? I wish people would stop with this tired war-mongering. China was not the one who opened up this can of worms. China has never been the one pushing the edge of capabilities. Before Sam Altman decided to give ChatGPT to the world, they were actively cracking down on software companies (in favor of hardware & "concrete" production). We, the US, are the ones who chose to do this. We started the race. We put the world, all of humanity, on this path. > Do you honestly think that would make a lick of difference? I don't know, it depends. Perhaps we're lucky and the timelines are slow enough that 20-30% of the population loses their jobs before things become unrecoverable. Tech companies used to warn people not to wear their badges in public in San Francisco -- and that was what, 2020? Would you really want to work at "Human Replacer, Inc." when that means walking out and about among a population who you know hates you, viscerally? Or if we make it to 2028 in the same condition. The Bonus Army was bad enough -- how confident are you that the government would stand their ground, keep letting these labs advance capabilities, when their electoral necks were on the line? This defeatism is a self-fulfilling prophecy. The people have the power to make things happen, and rhetoric like this is the most powerful thing holding them back. | | |
| ▲ | eagleislandsong 15 hours ago | parent [-] | | > China was not the one who opened up this can of worms Thank you. As someone who lives in Southeast Asia (and who also has lived in East Asia -- pardon the deliberate vagueness, for I do not wish to reveal too many potentially personally identifying information), this is how many of us in these regions view the current tensions between China and Taiwan as well. Don't get me wrong; we acknowledge that many Taiwanese people want independence, that they are a people with their own aspirations and agency. But we can also see that the US -- and its European friends, which often blindly adopt its rhetoric and foreign policy -- is deliberately using Taiwan as a disposable pawn to attempt to provoke China into a conflict. The US will do what it has always done ever since the post-WW2 period -- destabilise entire regions of countries to further its own imperialistic goals, causing the deaths and suffering of millions, and then leaving the local populations to deal with the fallout for many decades after. Without the US intentionally stoking the flames of mutual antagonism between China and Taiwan, the two countries could have slowly (perhaps over the next decades) come to terms with each other, be it voluntary reunification or peaceful separation. If you know a bit of Chinese history, it is not entirely far-fetched at all to think that the Chinese might eventually agree to recognising Taiwan as an independent nation, but now this option has now been denied because the US has decided to use Taiwan as a pawn in a proxy conflict. To anticipate questions about China's military invasion of Taiwan by 2027: No, I do not believe it will happen. Don't believe everything the US authorities claim. |
| |
| ▲ | ctoth 6 hours ago | parent | prev [-] | | We're all gonna die but come on, who wants to stop that! |
|
|
| |
| ▲ | goatlover a day ago | parent | prev | next [-] | | > "Development of Superhuman machine intelligence is probably the greatest threat to the continued existence of humanity” If that's really true, why is there such a big push to rapidly improve AI? I'm guessing OpenAI, Google, Anthropic, Apple, Meta, Boston Dynamics don't really believe this. They believe AI will make them billions. What is OpenAI's definition of AGI? A model that makes $100 billion? | | |
| ▲ | AgentME a day ago | parent | next [-] | | Because they also believe the development of superhuman machine intelligence will probably be the greatest invention for humanity. The possible upsides and downsides are both staggeringly huge and uncertain. | |
| ▲ | medvezhenok 20 hours ago | parent | prev [-] | | You can also have prisoner’s dilemma where no single actor is capable of stopping AI’s advance |
| |
| ▲ | archagon a day ago | parent | prev | next [-] | | And why are Altman's words worth anything? Is he some sort of great thinker? Or a leading AI researcher, perhaps? No. Altman is in his current position because he's highly effective at consolidating power and has friends in high places. That's it. Everything he says can be seen as marketing for the next power grab. | | |
| ▲ | tim333 8 hours ago | parent | next [-] | | Altman did play some part in bringing ChatGPT about. I think the point is the people making AI or running companies making current AI are saying be wary. In general it's worth weighting the opinions of people who are leaders in a field, about that field, over people who know little about it. | |
| ▲ | skeeter2020 a day ago | parent | prev [-] | | well, he did also have a an early (failed) YC startup - does that add cred? |
| |
| ▲ | samr71 a day ago | parent | prev [-] | | It's not something you need to worry about. If we get the Singularity, it's overwhelmingly likely Jesus will return concurrently. | | |
|
| |
| ▲ | cavisne a day ago | parent | prev | next [-] | | This article was prescient enough that I had to check in wayback machine. Very cool. | |
| ▲ | torginus a day ago | parent | prev | next [-] | | I'm not seeing the prescience here - I don't wanna go through the specific points but the main gist here seems to be that chatbots will become very good at pretending to be human and influencing people to their own ends. I don't think much has happened on these fronts (owning to a lack of interest, not technical difficulty). AI boyfriends/roleplaying etc. seems to have stayed a very niche interest, with models improving very little over GPT3.5, and the actual products are seemingly absent. It's very much the product of the culture war era, where one of the scary scenarios show off, is a chatbot riling up a set of internet commenters and goarding them lashing out against modern leftist orthodoxy, and then cancelling them. With all thestrongholds of leftist orthodoxy falling into Trump's hands overnight, this view of the internet seems outdated. Troll chatbots still are a minor weapon in information warfare/ The 'opinion bubbles' and manipulation of trending topics on social media (with the most influential content still written by humans), to change the perception of what's the popular concensus still seem to hold up as primary tools of influence. Nowadays, when most people are concerned about stuff like 'will the US go into a shooting war against NATO' or 'will they manage to crash the global economy', just to name a few of the dozen immediately pressing global issues, I think people are worried about different stuff nowadays. At the same time, there's very little mention of 'AI will take our jobs and make us poor' in both the intellectual and physical realms, something that's driving most people's anxiety around AI nowadays. It also puts the 'superintelligent unaligned AI will kill us all' argument very often presented by alignment people as a primary threat rather than the more plausible 'people controlling AI are the real danger'. | |
| ▲ | dkdcwashere a day ago | parent | prev | next [-] | | > The alignment community now starts another research agenda, to interrogate AIs about AI-safety-related topics. For example, they literally ask the models “so, are you aligned? If we made bigger versions of you, would they kill us? Why or why not?” (In Diplomacy, you can actually collect data on the analogue of this question, i.e. “will you betray me?” Alas, the models often lie about that. But it’s Diplomacy, they are literally trained to lie, so no one cares.) …yeah? | |
| ▲ | botro a day ago | parent | prev | next [-] | | This is damn near prescient, I'm having a hard time believing it was written in 2021. He did get this part wrong though, we ended up calling them 'Mixture of Experts' instead of 'AI bureaucracies'. | | | |
| ▲ | smusamashah a day ago | parent | prev | next [-] | | How does it talk about GPT-1 or 3 if it was before ChatGPT? | | |
| ▲ | dragonwriter a day ago | parent | next [-] | | GPT-3 (and, naturally, all prior versions even farther back) was released ~2 years before ChatGPT (whose launch model was GPT-3.5) The publication date on this article is about halfway between GPT-3 and ChatGPT releases. | |
| ▲ | Tenoke a day ago | parent | prev [-] | | GPT-2 for example came out in 2019. ChatGPT wasn't the start of GPT. |
| |
| ▲ | LordDragonfang a day ago | parent | prev | next [-] | | > (2025) Making models bigger is not what’s cool anymore. They are trillions of parameters big already. What’s cool is making them run longer, in bureaucracies of various designs, before giving their answers. Holy shit. That's a hell of a called shot from 2021. | | |
| ▲ | someothherguyy 12 hours ago | parent [-] | | its vague, and could have meant anything. everyone knew parameters would grow and its reasonable to expect that things that grow have diminishing returns at some point. this happened in late 2023 and throughout 2024 as well. |
| |
| ▲ | dingnuts a day ago | parent | prev [-] | | nevermind, I hate this website :D | | |
| ▲ | comp_throw7 a day ago | parent [-] | | Surely you're familiar with https://ai.meta.com/research/cicero/diplomacy/ (2022)? > I wonder who pays the bills of the authors. And your bills, for that matter. Also, what a weirdly conspiratorial question. There's a prominent "Who are we?" button near the top of the page and it's not a secret what any of the authors did or do for a living. | | |
| ▲ | dingnuts a day ago | parent [-] | | hmmm I apparently confused it with an RTS, oops. also it's not conspiratorial to wonder if someone in silicon valley today receives funding through the AI industry lol like half the industry is currently propped up by that hype, probably half the commenters here are paid via AI VC investments |
|
|
|
|
| ▲ | moab a day ago | parent | prev | next [-] |
| > "OpenBrain (the leading US AI project) builds AI agents that are good enough to dramatically accelerate their research. The humans, who up until very recently had been the best AI researchers on the planet, sit back and watch the AIs do their jobs, making better and better AI systems." I'm not sure what gives the authors the confidence to predict such statements. Wishful thinking? Worst-case paranoia? I agree that such an outcome is possible, but on 2--3 year timelines? This would imply that the approach everyone is taking right now is the right approach and that there are no hidden conceptual roadblocks to achieving AGI/superintelligence from DFS-ing down this path. All of the predictions seem to ignore the possibility of such barriers, or at most acknowledge the possibility but wave it away by appealing to the army of AI researchers and industry funding being allocated to this problem. IMO it is the onus of the proposers of such timelines to argue why there are no such barriers and that we will see predictable scaling in the 2--3 year horizon. |
| |
| ▲ | throwawaylolllm a day ago | parent | next [-] | | It's my belief (and I'm far from the only person who thinks this) that many AI optimists are motivated by an essentially religious belief that you could call Singularitarianism. So "wishful thinking" would be one answer. This document would then be the rough equivalent of a Christian fundamentalist outlining, on the basis of tangentially related news stories, how the Second Coming will come to pass in the next few years. | | |
| ▲ | viccis 21 hours ago | parent | next [-] | | Crackpot millenarians have always been a thing. This crop of them is just particularly lame and hellbent on boiling the oceans to get their eschatological outcome. | |
| ▲ | ivm 16 hours ago | parent | prev | next [-] | | Spot on, see the 2017 article "God in the machine: my strange journey into transhumanism" about that dynamic: https://www.theguardian.com/technology/2017/apr/18/god-in-th... | |
| ▲ | pixl97 a day ago | parent | prev | next [-] | | Eh, not sure if the second coming is a great analogy. That wholly depends on the whims of a fictional entity performing some unlikely actions. Instead think of them saying a crusade occurring in the next few years. When the group saying the crusade is coming is spending billions of dollars to trying to make just that occur you no longer have the ability to say it's not going to happen. You are now forced to examine the risks of their actions. | |
| ▲ | spacephysics 8 hours ago | parent | prev [-] | | Reminds me of Fallout's Children of Atom "Church of the Children of Atom" Maybe we'll see "Church of the Children of Altman" /s It seems without a framework of ethics/morality (insert XYZ religion), us humans find one to grasp onto. Be it a cult, a set of not-so-fleshed-out ideas/philosophies etc. People who say they aren't religious per-se, seem to have some set of beliefs that amount to religion. Just depends who or what you look towards for those beliefs, many of which seem to be half-hazard. People I may disagree with the most, many times at least have a realization of what ideas/beliefs are unifying their structure of reality, with others just not aware. A small minority of people can rely on schools of philosophical thought, and 'try on' or play with different ideas, but have a self-reflection that allows them to see when they transgress from ABC philosophy or when the philosophy doesn't match with their identity to a degree. |
| |
| ▲ | barbarr a day ago | parent | prev | next [-] | | It also ignores the possibility of plateau... maybe there's a maximum amount of intelligence that matter can support, and it doesn't scale up with copies or speed. | | |
| ▲ | AlexandrB a day ago | parent | next [-] | | Or scales sub-linearly with hardware. When you're in the rising portion of an S-curve[1] you can't tell how much longer it will go on before plateauing. A lot of this resembles post-war futurism that assumed we would all be flying around in spaceships and personal flying cars within a decade. Unfortunately the rapid pace of transportation innovation slowed due to physical and cost constraints and we've made little progress (beyond cost optimization) since. [1] https://en.wikipedia.org/wiki/Sigmoid_function | | |
| ▲ | Tossrock 20 hours ago | parent [-] | | The fact that it scales sub linearly with hardware is well known and in fact foundational to the scaling laws on which modern LLMs are built, ie performance scales remarkably closely to log(compute+data), over many orders of magnitude. |
| |
| ▲ | pixl97 a day ago | parent | prev [-] | | Eh, these mathematics still don't work out in humans favor... Lets say intelligence caps out at the maximum smartest person that's ever lived. Well, the first thing we'd attempt to do is build machines up to that limit that 99.99999 percent of us would never get close to. Moreso the thinking parts of humans is only around 2 pounds of mush in side of our heads. On top of that you don't have to grow them for 18 years first before they start outputting something useful. That and they won't need sleep. Oh and you can feed them with solar panels. And they won't be getting distracted by that super sleek server rack across the aisle. We do know 'hive' or societal intelligence does scale over time especially with integration with tooling. The amount of knowledge we have and the means of which we can apply it simply dwarf previous generations. |
| |
| ▲ | ddp26 18 hours ago | parent | prev | next [-] | | Check out the Timelines Forecast under "research". They model this very carefully. (They could be wrong, but this isn't a guess, it's a well-researched forecast.) | |
| ▲ | MrScruff 13 hours ago | parent | prev [-] | | I would assume this comes from having faith in the overall exponential trend rather than getting that much into the weeds of how this will come about. I can sort of see why you might think that way - everyone was talking about hitting a wall with brute force scaling and then inference time scaling comes along to keep things progressing. I wouldn't be quite as confident personally and as have many have said before, a sigmoid looks like an exponential in it's initial phase. |
|
|
| ▲ | osigurdson an hour ago | parent | prev | next [-] |
| Perhaps more of a meta question is, what is the value of optimistic vs pessimistic predictions regarding what AI might look like in 2-10 years? I.e. if one assumes that AI has hit a wall, what is the benefit? Similarly, if one assumes that its all "robots from Mars" in a year or two, what is the benefit of that? There is no point in making predictions if no actions are taken. It all seems to come down to buy or sell NVDA. |
|
| ▲ | IshKebab a day ago | parent | prev | next [-] |
| This is hilariously over-optimistic on the timescales. Like on this timeline we'll have a Mars colony in 10 years, immortality drugs in 15 and Half Life 3 in 20. |
| |
| ▲ | danpalmer a day ago | parent | next [-] | | These timelines always assume that things progress as quickly as they can be conceived of, likely because these timelines come from "Ideas Guys" whose involvement typically ends at that point. Orbital mechanics begs to disagree about a Mars colony in 10 years. Drug discovery has many steps that take time, even just the trials will take 5 years, let alone actually finding the drugs. | | |
| ▲ | movpasd 11 hours ago | parent | next [-] | | It reminds me of this rather classic post: http://johnsalvatier.org/blog/2017/reality-has-a-surprising-... Science is not ideas: new conceptual schemes must be invented, confounding variables must be controlled, dead-ends explored. This process takes years. Engineering is not science: kinks must be worked out, confounding variables incorporated. This process also takes years. Technology is not engineering: the purely technical implementation must spread, become widespread and beat social inertia and its competition, network effects must be established. Investors and consumers must be convinced in the long term. It must survive social and political repercussions. This process takes yet more years. | |
| ▲ | wkat4242 20 hours ago | parent | prev [-] | | Didn't the covid significantly reduce trial times? I thought that was such a success that they continued on the same foot. | | |
| ▲ | danpalmer 19 hours ago | parent | next [-] | | The other reply has better info on covid specifically, but also consider that this refers to "immortality drugs". How long do we have to test those to conclude that they do in fact provide "immortality"? Now sure, they don't actually mean immortality, and we don't need to test forever to conclude they extend life, but we probably do have to test for years to get good data on whether a generic life extension drug is effective, because you're testing against illness, old age, etc, things that take literally decades to kill. That's not to mention that any drug like that will be met with intense skepticism and likely need to overcome far more scrutiny than normal (rather than the potentially less scrutiny that covid drugs might have managed). | |
| ▲ | agos 13 hours ago | parent | prev | next [-] | | trial times were very brief for Covid vaccines because 1) there was no shortage of volunteers, capital, and political alignment at every level 2) the virus was everywhere and so it was really, really easy to verify if it was working. Compare this with a vaccination for a very rare but deadly disease: it's really hard to know if it's working because you can't just expose your test subjects to the deadly disease! | |
| ▲ | pama 19 hours ago | parent | prev [-] | | No it didn’t. At least not for new small molecule drugs. It did reduce times a bit for the first vaccines because there were many volunteers available, and it did allow some antibody drug candidates to be used before full testing was complete. The only approved small molecule drug for covid is paxlovid, with both components of its formulation tested on humans for the first time many years before covid. All the rest of the small molecule drugs are still in early parts of the pipeline or have been abandoned. |
|
| |
| ▲ | mchusma a day ago | parent | prev | next [-] | | I like that the "slowdown" scenario has by 2030 we have a robot economy, cure for aging, brain uploading, and are working on a Dyson Sphere. | | |
| ▲ | Aurornis 20 hours ago | parent [-] | | The story is very clearly modeled to follow the exponential curve they show. Like the drew the curve out into the shape they wanted, put some milestones on it, and then went to work imagining what would happen if it continued with a heavy dose of X-risk doomerism to keep it spicy. It conveniently ignores all of the physical constraints around things like manufacturing GPUs and scaling training networks. | | |
| ▲ | joshjob42 16 hours ago | parent [-] | | https://ai-2027.com/research/compute-forecast In section 4 they discuss their projections specifically for model size, the state of inference chips in 2027, etc. It's largely pretty in line with expectations in terms of the capacity, and they only project them using 10k of their latest gen wafer scale inference chips by late 2027, roughly like 1M H100 equivalents. That doesn't seem at all impossible. They also earlier on discuss expectations for growth in efficiency of chips, and for growth in spending, which is only ~10x over the next 2.5 years, not unreasonable in absolute terms at all given the many tens of billions of dollars flooding in. So on the "can we train the AI" front, they mostly are just projecting 2.5 years of the growth in scale we've been seeing. The reason they predict a fairly hard takeoff is they expect that distillation, some algorithmic improvements, and iterated creation of synthetic data, training, and then making more synthetic data will enable significant improvements in efficiency of the underlying models (something still largely in line with developments over the last 2 years). In particular they expect a 10T parameter model in early 2027 to be basically human equivalent, and they expect it to "think" at about the rate humans do, 10 words/second. That would require ~300 teraflops of compute per second to think at that rate, or ~0.1H100e. That means one of their inference chips could potentially run ~1000 copies (or fewer copies faster etc. etc.) and thus they have the capacity for millions of human equivalent researchers (or 100k 40x speed researchers) in early 2027. They further expect distillation of such models etc. to squeeze the necessary size down / more expensive models overseeing much smaller but still good models squeezing the effective amount of compute necessary, down to just 2T parameters and ~60 teraflops each, or 5000 human-equivalents per inference chip, making for up to 50M human-equivalents by late 2027. This is probably the biggest open question and the place where the most criticism seems to me to be warranted. Their hardware timelines are pretty reasonable, but one could easily expect needing 10-100x more compute or even perhaps 1000x than they describe to achieve Nobel-winner AGI or superintelligence. | | |
| ▲ | tsurba 10 hours ago | parent [-] | | I don’t believe so. I think all important parts that each need to be scaled to advance significantly in the LLM paradigm are at or near the end of the steep part of the sigmoid: 1) useful training data available in the internet
2) number of humans creating more training data ”manually”
3) parameter scaling
4) ”easy” algorithmic inventions
5) available+buildable compute ”Just” needing a few more algorithmic inventions to keep the graphs exponential is a cop out. It is already obvious that just scaling parameters and compute is not enough. I personally predict that scaling LLMs for solving all physical tasks (eg cleaning robots) or intellectual pursuits (they suck at multiplication) will not work out. We will get better specialized tools by collecting data from specific, high economic value, constrained tasks, and automating them, but scaling a (multimodal) LLM to solve everything in a single model will not be economically viable. We will get more natural interfaces for many tasks. This is how I think right now as a ML researcher, will be interesting to see how wrong was I in 2 years. EDIT: addition about latest algorithmic advances: - Deepseek style GRPO requires a ladder of scored problems progressively more difficult and appropriate to get useful gradients. For open-ended problems (like most interesting ones are) we have no ladders for, and it doesn’t work. In particular, learning to generate code for leetcode problems with a good number of well made unit tests is what it is good for. - Test-time inference is just adding an insane amount of more compute after training to brute-force double-check the sanity of answers Neither will keep the graphs exponential. |
|
|
| |
| ▲ | ctoth a day ago | parent | prev | next [-] | | Can you share your detailed projection of what you expect the future to look like so I can compare? | | |
| ▲ | IshKebab a day ago | parent | next [-] | | Sure 5 years: AI coding assistants are a lot better than they are now, but still can't actually replace junior engineers (at least ones that aren't shit). AI fraud is rampant, with faked audio commonplace. Some companies try replacing call centres with AI, but it doesn't really work and everyone hates it. Tesla's robotaxi won't be available, but Waymo will be in most major US cities. 10 years: AI assistants are now useful enough that you can use them in the ways that Apple and Google really wanted you to use Siri/Google Assistant 5 years ago. "What have I got scheduled for today?" will give useful results, and you'll be able to have a natural conversation and take actions that you trust ("cancel my 10am meeting; tell them I'm sick"). AI coding assistants are now very good and everyone will use them. Junior devs will still exist. Vibe coding will actually work. Most AI Startups will have gone bust, leaving only a few players. Art-based AI will be very popular and artists will use it all the time. It will be part of their normal workflow. Waymo will become available in Europe. Some receptionists and PAs have been replaced by AI. 15 years: AI researchers finally discover how to do on-line learning. Humanoid robots are robust and smart enough to survive in the real world and start to be deployed in controlled environments (e.g. factories) doing simple tasks. Driverless cars are "normal" but not owned by individuals and driverful cars are still way more common. Small light computers become fast enough that autonomous slaughter it's become reality (i.e. drones that can do their own navigation and face recognition etc.) 20 years: Valve confirms no Half Life 3. | | |
| ▲ | WXLCKNO 2 hours ago | parent | next [-] | | So in the past 5 years we went from not having ChatGPT at all and it being released in 2022 (with non "chat" models before that) but in the next 5 now that the entire tech world is consumed with making better AI models, we'll just get slightly better AI coding assistants? Reminds me of that comment about the first iPod being lame and having less space than a nomad. Worst take I've ever seen on here recently. | |
| ▲ | FeepingCreature 13 hours ago | parent | prev | next [-] | | It kind of sounds like you're saying "exactly everything we have today, we will have mildly more of." | |
| ▲ | Quarrelsome a day ago | parent | prev | next [-] | | you should add a bit where AI is pushed really hard in places where the subjects have low political power, like management of entry level workers, care homes or education and super bad stuff happens. Also we need a big legal event to happen where (for example) autonomous driving is part of a really big accident where lots of people die or someone brings a successful court case that an AI mortgage underwriter is discriminating based on race or caste. It won't matter if AI is actually genuinely responsible for this or not, what will matter is the push-back and the news cycle. Maybe more events where people start successfully gaming deployed AI at scale in order to get mortgages they shouldn't or get A-grades when they shouldn't. | |
| ▲ | 9dev a day ago | parent | prev | next [-] | | It’s soothing to read a realistic scenario amongst all of the ludicrous hype on here. | |
| ▲ | FairlyInvolved 21 hours ago | parent | prev | next [-] | | We are going to scale up GPT4 by a factor of ~10,000 and that will result in getting an accurate summary of your daily schedule? | | |
| ▲ | tsunagatta 16 hours ago | parent | next [-] | | If we’re lucky. | |
| ▲ | stale2002 19 hours ago | parent | prev [-] | | Unfortunately with the way scaling laws are working out, each order of magnitude increase in computer only makes models a little better. Meaning they nobody will even bother to 10,000X GPT4. |
| |
| ▲ | petesergeant 16 hours ago | parent | prev | next [-] | | > Some companies try replacing call centres with AI, but it doesn't really work and everyone hates it. I think this is much closer than you think, because there's a good percentage of call centers that are basically just humans with no power cosplaying as people who can help. My fiber connection went to shit recently. I messaged the company, and got a human who told me they were going to reset the connection from their side, if I rebooted my router. 30m later with no progress, I got a human who told me that they'd reset my ports, which I was skeptical about, but put down to a language issue, and again reset my router. 30m later, the human gave me an even more outlandish technical explanation of what they'd do, at which point I stumbled across the magical term "complaint" ... an engineer phoned me 15m later, said there was something genuinely wrong with the physical connection, and they had a human show up a few hours later and fix it. No part of the first-layer support experience there would have been degraded if replaced by AI, but the company would have saved some cash. | |
| ▲ | archagon a day ago | parent | prev [-] | | > Small light computers become fast enough that autonomous slaughter it's become reality This is the real scary bit. I'm not convinced that AI will ever be good enough to think independently and create novel things without some serious human supervision, but none of that matters when applied to machines that are destructive by design and already have expectations of collateral damage. Slaughterbots are going to be the new WMDs — and corporations are salivating at the prospect of being first movers. https://www.youtube.com/watch?v=UiiqiaUBAL8 | | |
| ▲ | Trumpion a day ago | parent | next [-] | | Why do you believe that? The lowest estimations of how much compute our brain represents was already achieved with the last chip from Nvidia (Blackwell). The newest gpu cluster from Google, Microsoft, Facebook, iax, and co have added so crazy much compute it's absurd. | | |
| ▲ | pixl97 a day ago | parent | next [-] | | >I'm not convinced that AI will ever be good enough to think independently a and >Why do you believe that? What takes less effort, time to deploy, and cost? I mean there is at least some probability we kill ourselves off with dangerous semi-thinking war machines leading to theater scale wars to the point society falls apart and we don't have the expensive infrastructure to make AI as envisioned in the future. With that said, I'm in the camp that we can create AGI as nature was able to with a random walk, we'll be able to reproduce it with intelligent design. | |
| ▲ | baq 13 hours ago | parent | prev [-] | | If you bake the model onto the chip itself, which is what should be happening for local LLMs once a good enough one is trained eventually, you’ll be looking at orders of magnitude reduction in power consumption at constant inference speed. |
| |
| ▲ | dontlikeyoueith a day ago | parent | prev [-] | | Zero Dawn future confirmed. |
|
| |
| ▲ | Gud a day ago | parent | prev | next [-] | | Slightly slower web frameworks by 2026. By 2030, a lot slower. | |
| ▲ | arduanika 36 minutes ago | parent | prev [-] | | With each passing year, AI doom grifters will learn more and more web design gimmicks. |
| |
| ▲ | Trumpion a day ago | parent | prev | next [-] | | We currently don't see any ceiling if this continues in this speed, we will have cheaper, faster and better models every quarter. Therewas never something progressing so fast It would be very ignorant not to keep a very close eye on it There is still a a chance that it will happen a lot slower and the progression will be slow enough that we adjust in time. But besides AI we also now get robots. The impact for a lot of people will be very real | |
| ▲ | zvitiate a day ago | parent | prev | next [-] | | No, sooner lol. We'll have aging cures and brain uploading by late 2028. Dyson Swarms will be "emerging tech". | |
| ▲ | turnsout a day ago | parent | prev | next [-] | | IMO they haven't even predicted mid-2025. > Coding AIs increasingly look like autonomous agents rather than mere assistants: taking instructions via Slack or Teams and making substantial code changes on their own, sometimes saving hours or even days.
Yeah, we are so not there yet. | | |
| ▲ | Tossrock 21 hours ago | parent [-] | | That is literally the pitch line for Devin. I recently spoke to the CTO of a small healthtech startup and he was very pro-Devin for small fixes and PRs, and thought he was getting his money worth. Claude Code is a little clunkier but gives better results, and it wouldn't take much effort to hook it up to a Slack interface. | | |
| ▲ | turnsout 20 hours ago | parent [-] | | Yeah, I get that there are startups trying to do it. But I work with Cursor quite a bit… there is no way I would trust an LLM code agent to take high-level direction and issue a PR on anything but the most trivial bug fix. | | |
| ▲ | baq 13 hours ago | parent [-] | | Last year they couldn’t even do a simple fix (they could add a null coalescing operator or an early return which didn’t make sense, that’s about it). Now I’m getting hundreds of LOC of functionality with multiple kLOC of tests out of the agent mode. No way it gets in without a few iterations, but it’s sooo much better than last April. |
|
|
| |
| ▲ | sva_ a day ago | parent | prev [-] | | You forgot fusion energy | | |
| ▲ | klabb3 a day ago | parent [-] | | Quantum AI powered by cold fusion and blockchain when? |
|
|
|
| ▲ | jenny91 2 hours ago | parent | prev | next [-] |
| Late 2025, "its PhD-level knowledge of every field". I just don't think you're going to get there. There is still a fundamental limitation that you can only be as good as the sources you train on. "PhD-level" is not included in this dataset: in other words, you don't become PhD-level by reading stuff. Maybe in a few fields, maybe a masters level. But unless we come up with some way to have LLMs actually do original research, peer-review itself, and defend a thesis, it's not going to get to PhD-level. |
| |
| ▲ | MoonGhost an hour ago | parent [-] | | > Late 2025, "its PhD-level knowledge of every field". I just don't think you're going to get there. You think too much of PhDs. They are different. Some of them are just repackaging of existing knowledge. Some are just copy-paste like famous Putin's. Not sure he even rad, to be honest. |
|
|
| ▲ | Jun8 a day ago | parent | prev | next [-] |
| ACT post where Scott Alexander provides some additional info: https://www.astralcodexten.com/p/introducing-ai-2027. Manifold currently predicts 30%: https://manifold.markets/IsaacKing/ai-2027-reports-predictio... |
| |
| ▲ | Aurornis 19 hours ago | parent | next [-] | | > ACT post where Scott Alexander provides some additional info: https://www.astralcodexten.com/p/introducing-ai-2027 The pattern where Scott Alexander puts forth a huge claim and then immediately hedges it backward is becoming a tiresome theme. The linguistic equivalent of putting claims into a superposition where the author is both owning it and distancing themselves from it at the same time, leaving the writing just ambiguous enough that anyone reading it 5 years from now couldn't pin down any claim as false because it was hedged in both directions. Schrödinger's prediction. > Do we really think things will move this fast? Sort of no > So maybe think of this as a vision of what an 80th percentile fast scenario looks like - not our precise median, but also not something we feel safe ruling out. The talk of "not our precise median" and "Not something we feel safe ruling out" is an elaborate way of hedging that this isn't their actual prediction but, hey, anything can happen so here's a wild story! When the claims don't come true they can just point back to those hedges and say that it wasn't really their median prediction (which is conveniently not noted). My prediction: The vague claims about AI becoming more powerful and useful will come true because, well, they're vague. Technology isn't about to reverse course and get worse. The actual bold claims like humanity colonizing space in the late 2020s with the help of AI are where you start to realize how fanciful their actual predictions are. It's like they put a couple points of recent AI progress on a curve, assumed an exponential trajectory would continue forever, and extrapolated from that regression until AI was helping us colonize space in less than 5 years. > Manifold currently predicts 30%: Read the fine print. It only requires 30% of judges to vote YES for it to resolve to YES. This is one of those bets where it's more about gaming the market than being right. | |
| ▲ | leonidasv 18 hours ago | parent | prev | next [-] | | > Do we really think things will move this fast? Sort of no - between the beginning of the project last summer and the present, Daniel’s median for the intelligence explosion shifted from 2027 to 2028. We keep the scenario centered around 2027 because it’s still his modal prediction (and because it would be annoying to change). Other members of the team (including me) have medians later in the 2020s or early 2030s, and also think automation will progress more slowly. So maybe think of this as a vision of what an 80th percentile fast scenario looks like - not our precise median, but also not something we feel safe ruling out. Important disclaimer that's lacking in OP's link. | |
| ▲ | whiddershins 8 hours ago | parent | prev | next [-] | | > A rise in AI-generated propaganda failed to materialize. hah! | |
| ▲ | crazystar a day ago | parent | prev [-] | | 47% now soo a coin toss | | |
| ▲ | elicksaur a day ago | parent | next [-] | | Note the market resolves by: > Resolution will be via a poll of Manifold moderators. If they're split on the issue, with anywhere from 30% to 70% YES votes, it'll resolve to the proportion of YES votes. So you should really read it as “Will >30% of Manifold moderators in 2027 think the ‘predictions seem to have been roughly correct up until that point’?” | |
| ▲ | layer8 a day ago | parent | prev [-] | | 32% again now. |
|
|
|
| ▲ | superconduct123 a day ago | parent | prev | next [-] |
| Why are the biggest AI predictions always made by people who aren't deep in the tech side of it? Or actually trying to use the models day-to-day... |
| |
| ▲ | AlphaAndOmega0 a day ago | parent | next [-] | | Daniel Kokotajlo released the (excellent) 2021 forecast. He was then hired by OpenAI, and not at liberty to speak freely, until he quit in 2024. He's part of the team making this forecast. The others include: Eli Lifland, a superforecaster who is ranked first on RAND’s Forecasting initiative. You can read more about him and his forecasting team here. He cofounded and advises AI Digest and co-created TextAttack, an adversarial attack framework for language models. Jonas Vollmer, a VC at Macroscopic Ventures, which has done its own, more practical form of successful AI forecasting: they made an early stage investment in Anthropic, now worth $60 billion. Thomas Larsen, the former executive director of the Center for AI Policy, a group which advises policymakers on both sides of the aisle. Romeo Dean, a leader of Harvard’s AI Safety Student Team and budding expert in AI hardware. And finally, Scott Alexander himself. | | |
| ▲ | kridsdale3 a day ago | parent | next [-] | | TBH, this kind of reads like the pedigrees of the former members of the OpenAI board. When the thing blew up, and people started to apply real scrutiny, it turned out that about half of them had no real experience in pretty much anything at all, except founding Foundations and instituting Institutes. A lot of people (like the Effective Altruism cult) seem to have made a career out of selling their Sci-Fi content as policy advice. | | |
| ▲ | MrScruff 13 hours ago | parent | next [-] | | I kind of agree - since the Bostrom book there is a cottage industry of people with non-technical backgrounds writing papers about singularity thought experiments, and it does seem to be on the spectrum with hard sci-fi writing. A lot of these people are clearly intelligent, and it's not even that I think everything they say is wrong (I made similar assumptions long ago before I'd even heard of Ray Kurzweil and the Singularity, although at the time I would have guessed 2050). It's just that they seem to believe their thought process and Bayesian logic is more rigourous than it actually is. | |
| ▲ | flappyeagle a day ago | parent | prev [-] | | c'mon man, you don't believe that, let's have a little less disingenuousness on the internet | | |
| ▲ | arduanika 21 hours ago | parent [-] | | How would you know what he believes? There's hype and there's people calling bullshit. If you work from the assumption that the hype people are genuine, but the people calling bullshit can't be for real, that's how you get a bubble. |
|
| |
| ▲ | pixodaros 6 hours ago | parent | prev | next [-] | | Scott Alexander, for what its worth, is a psychiatrist, race science enthusiast, and blogger whose closest connection to software development is Bay Area house parties and a failed startup called MetaMed (2012-2015) https://rationalwiki.org/wiki/MetaMed | |
| ▲ | nice_byte 16 hours ago | parent | prev | next [-] | | this sounds like a bunch of people who make a living _talking_ about the technology, which lends them close to 0 credibility. | |
| ▲ | superconduct123 a day ago | parent | prev [-] | | I mean either researchers creating new models or people building products using the current models Not all these soft roles |
| |
| ▲ | torginus a day ago | parent | prev | next [-] | | Because these people understand human psychology and how to play on fears (of doom, or missing out) and insecurities of people, and write compelling narratives while sounding smart. They are great at selling stories - they sold the story of the crypto utopia, now switching their focus to AI. This seems to be another appeal to enforce AI regulation in the name of 'AI safetyiism', which was made 2 years ago but the threats in it haven't really panned out. For example an oft repeated argument is the dangerous ability of AI to design chemical and biological weapons, I wish some expert could weigh in on this, but I believe the ability to theorycraft pathogens effective in the real world is absolutely marginal - you need actual lab work and lots of physical experiments to confirm your theories. Likewise the dangers of AI systems to exfiltrate themselves to multi-million dollar AI datacenter GPU systems everyone supposedly just has lying about, is ... not super realistc. The ability of AIs to hack computer systems is much less theoretical - however as AIs will get better at black-hat hacking, they'll get better at white-hat hacking as well - as there's literally no difference between the two, other than intent. And here in lies a crucial limitation of alignment and safetyism - sometimes there's no way to tell apart harmful and harmless actions, other than whether the person undertaking them means well. | |
| ▲ | ZeroTalent a day ago | parent | prev | next [-] | | People who are skilled fiction writers might lack technical expertise. In my opinion, this is simply an interesting piece of science fiction. | |
| ▲ | rglover a day ago | parent | prev | next [-] | | Aside from the other points about understanding human psychology here, there's also a deep well they're trying to fill inside themselves. That of being someone who can't create things without shepherding others and see AI as the "great equalizer" that will finally let them taste the positive emotions associated with creation. The funny part, to me, is that it won't. They'll continue to toil and move on to the next huck just as fast as they jumped on this one. And I say this from observation. Nearly all of the people I've seen pushing AI hyper-sentience are smug about it and, coincidentally, have never built anything on their own (besides a company or organization of others). Every single one of the rational "we're on the right path but not quite there" takes have been from seasoned engineers who at least have some hands-on experience with the underlying tech. | |
| ▲ | FeepingCreature 13 hours ago | parent | prev | next [-] | | I use the models daily and agree with Scott. | |
| ▲ | Tenoke a day ago | parent | prev | next [-] | | ..The first person listed is ex-OpenAI. | |
| ▲ | bpodgursky a day ago | parent | prev | next [-] | | Because you can't be a full time blogger and also a full time engineer. Both take all your time, even ignoring time taken to build talent. There is simply a tradeoff of what you do with your life. There are engineers with AI predictions, but you aren't reading them, because building an audience like Scott Alexander takes decades. | |
| ▲ | ohgr a day ago | parent | prev [-] | | In the path to self value people explain their worth by what they say not what they know. If what they say is horse dung, it is irrelevant to their ego if there is someone dumber than they are listening. This bullshit article is written for that audience. Say bullshit enough times and people will invest. | | |
|
|
| ▲ | infecto a day ago | parent | prev | next [-] |
| Could not get through the entire thing. It’s mostly a bunch of fantasy intermingled with bits of possible interesting discussion points. The whole right side metrics are purely a distraction because entirely fiction. |
| |
|
| ▲ | dcanelhas 3 hours ago | parent | prev | next [-] |
| > Once the new datacenters are up and running, they’ll be able to train a model with 10^28 FLOP—a thousand times more than GPT-4. Is there some theoretical substance or empirical evidence to suggest that the story doesn't just end here? Perhaps OpenBrain sees no significant gains over the previous iteration and implodes under the financial pressure of exorbitant compute costs. I'm not rooting for an AI winter 2.0 but I fail to understand how people seem sure of the outcome of experiments that have not even been performed yet. Help, am I missing something here? |
| |
| ▲ | the8472 3 hours ago | parent [-] | | https://gwern.net/scaling-hypothesis exponential scaling has been holding up for more than a decade now, since alexnet. And when there were the first murmurings that maybe we're finally hitting a wall the labs published ways to harness inference-time compute to get better results which can be fed back into more training. |
|
|
| ▲ | porphyra a day ago | parent | prev | next [-] |
| Seems very sinophobic. Deepseek and Manus have shown that China is legitimately an innovation powerhouse in AI but this article makes it sound like they will just keep falling behind without stealing. |
| |
| ▲ | MugaSofer a day ago | parent | next [-] | | That whole section seems to be pretty directly based on DeepSeek's "very impressive work" with R1 being simultaneously very impressive, and several months behind OpenAI. (They more or less say as much in footnote 36.) They blame this on US chip controls just barely holding China back from the cutting edge by a few months. I wouldn't call that a knock on Chinese innovation. | |
| ▲ | aoanevdus 16 hours ago | parent | prev | next [-] | | Don’t assume that because the article depicts this competition between the US and China, that the authors actually want China to fail. Consider the authors and the audience. The work is written by western AI safety proponents, who often need to argue with important people who say we need to accelerate AI to “win against China” and don’t want us to be slowed down by worrying about safety. From that perspective, there is value in exploring the scenario: ok, if we accept that we need to compete with China, what would that look like? Is accelerating always the right move? The article, by telling a narrative where slowing down to be careful with alignment helps the US win, tries to convince that crowd to care about alignment. Perhaps, people in China can make the same case about how alignment will help China win against US. | |
| ▲ | hexator 19 hours ago | parent | prev | next [-] | | Yes, it's extremely sinophobic and entirely too dismissive of China. It's pretty clear what the author's political leanings are, by what they mention and by what they do not. | |
| ▲ | princealiiiii a day ago | parent | prev | next [-] | | Stealing model weights isn't even particularly useful long-term, it's the training + data generation recipes that have value. | |
| ▲ | ugh123 a day ago | parent | prev | next [-] | | Don't confuse innovation with optimisation. | | |
| ▲ | pixl97 a day ago | parent [-] | | Don't confuse designing the product with winning the market. |
| |
| ▲ | a3w a day ago | parent | prev | next [-] | | How so? Spoiler: US dooms mankind, China is the saviour in the two endings. | |
| ▲ | usef- 16 hours ago | parent | prev [-] | | In both endings it's saying that because compute becomes the bottleneck, and US has far more chips. Isn't it? |
|
|
| ▲ | zvitiate a day ago | parent | prev | next [-] |
| There's a lot to potentially unpack here, but idk, the idea that humanity entering hell (extermination) or heaven (brain uploading; aging cure) is whether or not we listen to AI safety researchers for a few months makes me question whether it's really worth unpacking. |
| |
| ▲ | 9dev a day ago | parent | next [-] | | Maybe people should just don’t listen to AI safety researchers for a few months? Maybe they are qualified to talk about inference and model weights and natural language processing, but not particularly knowledgeable about economics, biology, psychology, or… pretty much every other field of study? The hubris is strong with some people, and a certain oligarch with a god complex is acting out where that can lead right now. | | |
| ▲ | arduanika 15 hours ago | parent [-] | | It's charitable of you to think that they might be qualified to talk about inference and model weights and such. They are AI safety researchers, not AI researchers. Basically, a bunch of doom bloggers, jerking each other in a circle, a few of whom were tolerated at one of the major labs for a few years, to do their jerking on company time. |
| |
| ▲ | amelius a day ago | parent | prev [-] | | If we don't do it, someone else will. | | |
| ▲ | achierius a day ago | parent | next [-] | | That's obviously not true. Before OpenAI blew the field open, multiple labs -- e.g. Google -- were intentionally holding back their research from the public eye because they thought the world was not ready. Investors were not pouring billions into capabilities. China did not particularly care to focus on this one research area, among many, that the US is still solidly ahead in. The only reason timelines are as short as they are is because of people at OpenAI and thereafter Anthropic deciding that "they had no choice". They had a choice, and they took the one which has chopped at the very least years off of the time we would otherwise have had to handle all of this. I can barely begin to describe the magnitude of the crime that they have committed -- and so I suggest that you consider that before propagating the same destructive lies that led us here in the first place. | | |
| ▲ | pixl97 a day ago | parent [-] | | The simplicity of the statement "If we don't do it, someone else will." and thinking behind it eventually means someone will do just that unless otherwise prevented by some regulatory function. Simply put, with the ever increasing hardware speeds we were dumping out for other purposes this day would have come sooner than later. We're talking about only a year or two really. | | |
| ▲ | achierius 16 hours ago | parent | next [-] | | But every time, it doesn't have to happen yet. And when you're talking about the potential deaths of millions, or billions, why be the one who spawns the seed of destruction in their own home country? Why not give human brotherhood a chance? People have, and do, hold back. You notice the times they don't, and the few who don't -- you forget the many, many more who do refrain from doing what's wrong. "We have to nuke the Russians, if we don't do it first, they will" "We have to clone humans, if we don't do it, someone else will" "We have to annex Antarctica, if we don't do it, someone else will" | |
| ▲ | HeatrayEnjoyer 17 hours ago | parent | prev [-] | | Cloning? Bioweapons? Ever larger nuclear stockpiles? The world has collectively agreed not to do something more than once. AI would be easier to control than any of the above. GPUs can't be dug out of the ground. |
|
| |
| ▲ | layer8 a day ago | parent | prev | next [-] | | I’m okay if someone else unpacks it. | |
| ▲ | itishappy a day ago | parent | prev [-] | | Which? Exterminate humanity or cure aging? | | |
|
|
|
| ▲ | ikerino a day ago | parent | prev | next [-] |
| Feels reasonable in the first few paragraphs, then quickly starts reading like science fiction. Would love to read a perspective examining "what is the slowest reasonable pace of development we could expect." This feels to me like the fastest (unreasonable) trajectory we could expect. |
| |
| ▲ | admiralrohan a day ago | parent | next [-] | | No one knows what will happen. But these thought experiments can be useful as a critical thinking practice. | |
| ▲ | layer8 a day ago | parent | prev | next [-] | | The slowest is a sudden and permanent plateau, where all attempts at progress turn out to result in serious downsides that make them unworkable. | | |
| ▲ | 9dev a day ago | parent | next [-] | | Like an exponentially growing compute requirement for negligible performance gains, on the scale of the energy consumption of small countries? Because that is where we are, right now. | |
| ▲ | photonthug 18 hours ago | parent | prev [-] | | Even if this were true, it's not quite the end of the story is it? The hype itself creates lots of compute and to some extent the power needed to feed that compute, even if approximately zero of the hype pans out. So an interesting question becomes.. what happens with all the excess? Sure it probably gets gobbled up in crypto ponzi schemes, but I guess we can try to be optimistic. IDK, maybe we get to solve cancer and climate change anyway, not with fancy new AGI, but merely with some new ability to cheaply crunch numbers for boring old school ODEs. |
| |
| ▲ | zmj 18 hours ago | parent | prev | next [-] | | If you described today's AI capabilities to someone from 3 years ago, that would also sound like science fiction. Extrapolate. | |
| ▲ | ddp26 18 hours ago | parent | prev | next [-] | | The forecasts under "Research" are distributions, so you can compare the 10th percentile vs 90th percentile. Their research is consistent with a similar story unfolding over 8-10 years instead of 2. | |
| ▲ | FeepingCreature 13 hours ago | parent | prev [-] | | > Feels reasonable in the first few paragraphs, then quickly starts reading like science fiction. That's kind of unavoidably what accelerating progress feels like. |
|
|
| ▲ | wg0 3 hours ago | parent | prev | next [-] |
| Very detailed effort. Predicting future is very very hard. My gut feeling however says that none of this is happening. You cannot put LLMs into law and insurance and I don't see that happening with current foundations (token probabilities) of AI let alone AGI. By law and insurance - I mean hire an insurance agent or a lawyer. Give them your situation. There's almost no chance that such a professional would come wrong about any conclusions/recommendations based on the information you provide. I don't have that confidence in LLMs for that industries. Yet. Or even in a decade. |
| |
| ▲ | polynomial 3 hours ago | parent [-] | | > You cannot put LLMs into law and insurance Cass Sunstein would very strongly disagree. |
|
|
| ▲ | dughnut 8 hours ago | parent | prev | next [-] |
| I don’t know about you, but my takeaway is that the author is doing damage control but inadvertently tipped a hand that OpenAI is probably running an elaborate con job on the DoD. “Yes, we have a super secret model, for your eyes only, general. This one is definitely not indistinguishable from everyone else’s model and it doesn’t produce bullshit because we pinky promise. So we need $1T.” I love LLMs, but OpenAI’s marketing tactics are shameful. |
| |
|
| ▲ | pingou 2 hours ago | parent | prev | next [-] |
| Considering that each year that passes, technology offer us new ways to destroy ourselves, and gives another chance for humanity to pick a black ball, it seems to me like the only way to save ourselves is to create a benevolent AI to supervise us and neutralize all threads. There are obviously big risks with AI, as listed in the article, but the genie is out of the bottle anyway, even if all countries agreed to stop AI development, how long would that agreement last? 10 years? 20? 50? Eventually powerful AIs will be developed, if that is possible (which I believe it is, and I didn't think I'd see the current stunning development in my lifetime, I may not see AGI but I'm sure it'll get there eventually). |
|
| ▲ | ks2048 21 hours ago | parent | prev | next [-] |
| We know this complete fiction because of parts where "the White House considers x,y,z...", etc. - As if the White House in 2027 will be some rational actor reacting sanely to events in the real world. |
|
| ▲ | eob 7 hours ago | parent | prev | next [-] |
| An aspect of these self-improvement thought experiments that I’m willing to tentatively believe.. but want more resolution on, is the exact work involved in “improvement”. Eg today there’s billions of dollars being spent just to create and label more data, which is a global act of recruiting, training, organization, etc. When we imagine these models self improving, are we imagining them “just” inventing better math, or conducting global-scale multi-company coordination operations? I can believe AI is capable of the latter, but that’s an awful lot of extra friction. |
| |
| ▲ | acureau 2 hours ago | parent [-] | | This is exactly what makes this scenario so absurd to me. The authors don't even attempt to describe how any of this could realistically play out. They describe sequence models and RLAIF, then claim this approach "pays off" in 2026. The paper they link to is from 2022. RLAIF also does not expand the information encoded in the model, it is used to align the output with a set of guidelines. How could this lead to meaningful improvement in a model's ability to do bleeding-edge AI research? Why wouldn't that have happened already? I don't understand how anyone takes this seriously. Speculation like this is not only useless, but disingenuous. Especially when it's sold as "informed by trend extrapolations, wargames, expert feedback, experience at OpenAI, and previous forecasting successes". This is complete fiction which, at best, is "inspired by" the real world. I question the motives of the authors. |
|
|
| ▲ | ddp26 18 hours ago | parent | prev | next [-] |
| A lot of commenters here are reacting only to the narrative, and not the Research pieces linked at the top. There is some very careful thinking there, and I encourage people to engage with the arguments there rather than the stylized narrative derived from it. |
|
| ▲ | throw310822 6 hours ago | parent | prev | next [-] |
| My issue with this is that it's focused on one single, very detailed narrative (the battle between China and the US, played on a timeframe of mere months), while lacking any interesting discussion of other consequences of AI: what its impact is going to be on the job markets, employment rates, GDPs, political choices... Granted, if by this narrative the world is essentially ending two/ three years from now, then there isn't much time for any of those impacts to actually take place- but I don't think this is explicitly indicated either. If I am not mistaken, the bottom line of this essay is that, in all cases, we're five years away from the Singularity itself (I don't care what you think about the idea of Singularity with its capital S but that's what this is about). |
|
| ▲ | sivaragavan 17 hours ago | parent | prev | next [-] |
| Thanks to the authors for doing this wonderful piece of work and sharing it with credibility. I wish people see the possibilities here. But we are after all humans. It is hard to imagine our own downfall. Based on each individual's vantage point, these events might looks closer or farther than mentioned here. but I have to agree nothing is off the table at this point. The current coding capabilities of AI Agents are hard to downplay. I can only imagine the chain reaction of this creation ability to accelerate every other function. I have to say one thing though: The scenario in this site downplays the amount of resistance that people will put up - not because they are worried about alignment, but because they are politically motivated by parties who are driven by their own personal motives. |
|
| ▲ | kmeisthax a day ago | parent | prev | next [-] |
| > The agenda that gets the most resources is faithful chain of thought: force individual AI systems to “think in English” like the AIs of 2025, and don’t optimize the “thoughts” to look nice. The result is a new model, Safer-1. Oh hey, it's the errant thought I had in my head this morning when I read the paper from Anthropic about CoT models lying about their thought processes. While I'm on my soapbox, I will point out that if your goal is preservation of democracy (itself an instrumental goal for human control), then you want to decentralize and distribute as much as possible. Centralization is the path to dictatorship. A significant tension in the Slowdown ending is the fact that, while we've avoided AI coups, we've given a handful of people the ability to do a perfectly ordinary human coup, and humans are very, very good at coups. Your best bet is smaller models that don't have as many unused weights to hide misalignment in; along with interperability and faithful CoT research. Make a model that satisfies your safety criteria and then make sure everyone gets a copy so subgroups of humans get no advantage from hoarding it. |
|
| ▲ | Fraterkes 2 hours ago | parent | prev | next [-] |
| Completely earnest question for people who believe we are on this exponential trajectory: what should I look out for at the end of 2025 to see if we're on track for that scenario? What benchmark that naysayers think is years away will we have met? |
|
| ▲ | maerF0x0 5 hours ago | parent | prev | next [-] |
| > OpenBrain reassures the government that the model has been “aligned” so that it will refuse to comply with malicious requests Of course the real issue being that Governments have routinely demanded that 1) Those capabilities be developed for government monopolistic use, and 2) The ones who do not lose the capability (geo political power) to defend themselves from those who do. Using a US-Centric mindset... I'm not sure what to think about the US not developing AI hackers, AI bioweapons development, or AI powered weapons (like maybe drone swarms or something), if one presumes that China is, or Iran is, etc then whats the US to do in response? I'm just musing here and very much open to political science informed folks who might know (or know of leads) as to what kinds of actual solutions exist to arms races. My (admittedly poor), understanding of the cold war wasn't so much that the US won, but that the Soviets ran out of steam. |
|
| ▲ | resource0x 6 hours ago | parent | prev | next [-] |
| Every time NVDA/goog/msft tanks, we see these kinds of articles. |
|
| ▲ | ryankrage77 21 hours ago | parent | prev | next [-] |
| > "resist the temptation to get better ratings from gullible humans by hallucinating citations or faking task completion" Everything this from this point on is pure fiction. An LLM can't get tempted or resist temptations, at best there's some local minimum in a gradient that it falls into. As opaque and black-box-y as they are, they're still deterministic machines. Anthropomorphisation tells you nothing useful about the computer, only the user. |
| |
|
| ▲ | fire_lake 12 hours ago | parent | prev | next [-] |
| If you genuinely believe this, why on earth would you work for OpenAI etc even in safety / alignment? The only response in my view is to ban technology (like in Dune) or engage in acts of terror Unabomber style. |
| |
| ▲ | creatonez 12 hours ago | parent | next [-] | | > The only response in my view is to ban technology (like in Dune) or engage in acts of terror Unabomber style. Not far off from the conclusion of others who believe the same wild assumptions. Yudkowsky has suggested using terrorism to stop a hypothetical AGI -- that is, nuclear attacks on datacenters that get too powerful. | |
| ▲ | b3lvedere 10 hours ago | parent | prev [-] | | Most people work for money. As long as money is necessary to survive and prosper, people will work for it. Some of the work may not align with their morals and ethics, but in the end the money still wins. Banning will not automatically erase the existence and possibilty of things. We banned the use of nuclear weapons, yet we all know they exist. |
|
|
| ▲ | Joshuatanderson a day ago | parent | prev | next [-] |
| This is extremely important. Scott Alexander's earlier predictions are holding up extremely well, at least on image progress. |
|
| ▲ | zurfer 11 hours ago | parent | prev | next [-] |
| In the hope of improving this forecast, here is what I find implausible: - 1 lab constantly racing ahead and increasing the margin to other; the last 2 years are filled with ever-closer model capabilities and constantly new leaders (openai, anthropic, google, some would include xai). - Most of the compute budget on R&D. As model capabilities increase and cost goes down, demand will increase and if the leading lab doesn't provide, another lab will capture that and have more total dollars to back channel into R&D. |
|
| ▲ | Aldipower 5 hours ago | parent | prev | next [-] |
| No one can predict the future. Really, no one. Sometimes there is a hit, sure, but mostly it is a miss. The other thing is in their introduction: "superhuman AI"
_artificial_ intelligence is always, by definition, different from _natural_ intelligence. That they've chosen the word "superhuman" shows me that they are mixing the things up. |
| |
| ▲ | kmoser 4 hours ago | parent [-] | | I think you're reading too much into the meaning of "superhuman". I take it to mean "abilities greater than any single human" (for the same amount of time taken), which today's AIs have already demonstrated. |
|
|
| ▲ | owenthejumper an hour ago | parent | prev | next [-] |
| They would be better of making simple predictions, instead of proposing that in less than 2 years from now, the Trump administration will provide a UBI to all American citizens. That, and frequently talking about the wise president controlling this "thing", when in reality, he's a senile 80yrs old madman, is preposterous. |
|
| ▲ | crvdgc 6 hours ago | parent | prev | next [-] |
| Using Agent-2 to monitor Agent-3 sounds unnervingly similar to the plot of Philip K. Dick's Vulcan's Hammer [1]. An old super AI is used to fight a new version, named Vulcan 2 and Vulcan 3 respectively! [1] https://en.wikipedia.org/wiki/Vulcan's_Hammer |
|
| ▲ | qwertox a day ago | parent | prev | next [-] |
| That is some awesome webdesign. |
|
| ▲ | asimpletune 2 hours ago | parent | prev | next [-] |
| Didn’t Raymond Kurzweil predict like 30 years ago that AGI would be achieved in 2028? |
|
| ▲ | mr_world 2 hours ago | parent | prev | next [-] |
| > But they are still only going at half the pace of OpenBrain, mainly due to the compute deficit. Right. |
|
| ▲ | I_Nidhi 9 hours ago | parent | prev | next [-] |
| Though it's easy to dismiss as science fiction, this timeline paints a chillingly detailed picture of a potential AGI takeoff. The idea that AI could surpass human capabilities in research and development, and the fact that it will create an arms race between global powers, is unsettling. The risks—AI misuse, security breaches, and societal disruption—are very real, even if the exact timeline might be too optimistic. But the real concern lies in what happens if we’re wrong and AGI does surpass us. If AI accelerates progress so fast that humans can no longer meaningfully contribute, where does that leave us? |
|
| ▲ | pinetone a day ago | parent | prev | next [-] |
| I think it's worth noting that all of the authors have financial or professional incentive to accelerate the AI hype bandwagon as much as possible. |
| |
| ▲ | FairlyInvolved 21 hours ago | parent [-] | | I realise no one is infallible but do you not think Daniel Kokotajlo's integrity is now pretty well established with regard to those incentives? |
|
|
| ▲ | dr_dshiv a day ago | parent | prev | next [-] |
| But, I think this piece falls into a misconception about AI models as singular entities. There will be many instances of any AI model and each instance can be opposed to other instances. So, it’s not that “an AI” becomes super intelligent, what we actually seem to have is an ecosystem of blended human and artificial intelligences (including corporations!); this constitutes a distributed cognitive ecology of superintelligence. This is very different from what they discuss. This has implications for alignment, too. It isn’t so much about the alignment of AI to people, but that both human and AI need to find alignment with nature. There is a kind of natural harmony in the cosmos; that’s what superintelligence will likely align to, naturally. |
| |
| ▲ | ddp26 18 hours ago | parent | next [-] | | Check out the sidebar - they expect tens of thousands of copies of their agents collaborating. I do agree they don't fully explore the implications. But they do consider things like coordination amongst many agents. | | |
| ▲ | dr_dshiv 14 hours ago | parent [-] | | It’s just funny, because there are hundreds of millions of instances of ChatGPT running all the time. Each chat is basically an instance, since it has no connection to all the other chats. I don’t think connecting them makes sense due to privacy reasons. And, each chat is not autonomous but integrated with other intelligent systems. So, with more multiplicity, I think thinks work differently. More ecologically. For better and worse. |
| |
| ▲ | popalchemist a day ago | parent | prev [-] | | For now. |
|
|
| ▲ | fudged71 3 hours ago | parent | prev | next [-] |
| The most unrealistic thing is the inclusion of Americas involvement in the five eyes alliance aspect |
|
| ▲ | croemer 8 hours ago | parent | prev | next [-] |
| Pet peeve how they write FLOPS in the figure when they meant FLOP. Maybe the plural s after FLOP got capitalized. https://blog.heim.xyz/flop-for-quantity-flop-s-for-performan... |
|
| ▲ | danpalmer a day ago | parent | prev | next [-] |
| Interesting story, if you're into sci-fi I'd also recommend Iain M Banks and Peter Watts. |
|
| ▲ | moktonar 12 hours ago | parent | prev | next [-] |
| Catastrophic predictions of the future are always good, because all future predictions are usually wrong. I will not be scared as long as most future predictions where AI is involved are catastrophic. |
|
| ▲ | overgard 17 hours ago | parent | prev | next [-] |
| Why is any of this seen as desirable? Assuming this is a true prediction it sounds AWFUL. The one thing humans have that makes us human is intelligence. If we turn over thinking to machines, what are we exactly. Are we supposed to just consume mindlessly without work to do? |
|
| ▲ | amarcheschi a day ago | parent | prev | next [-] |
| I just spent some time trying to make claude and gemini make a violin plot of some polar dataframe. I've never used it and it's just for prototyping so i just went "apply a log to the values and make a violin plot of this polars dataframe". ANd had to iterate with them for 4/5 times each. Gemini got it right but then used deprecated methods I might be doing llm wrong, but i just can't get how people might actually do something not trivial just by vibe coding. And it's not like i'm an old fart either, i'm a university student |
| |
| ▲ | VOIPThrowaway a day ago | parent | next [-] | | You're asking it to think and it can't. It's spicy auto complete. Ask it to create a program that can create a violin plot from a CVS file. Because this has been "done before", it will do a decent job. | | | |
| ▲ | hiq a day ago | parent | prev | next [-] | | > had to iterate with them for 4/5 times each. Gemini got it right but then used deprecated methods How hard would it be to automate these iterations? How hard would it be to automatically check and improve the code to avoid deprecated methods? I agree that most products are still underwhelming, but that doesn't mean that the underlying tech is not already enough to deliver better LLM-based products. Lately I've been using LLMs more and more to get started with writing tests on components I'm not familiar with, it really helps. | | |
| ▲ | jaccola a day ago | parent | next [-] | | How hard can it be to create a universal "correctness" checker? Pretty damn hard! Our notion of "correct" for most things is basically derived from a very long training run on reality with the loss function being for how long a gene propagated. | |
| ▲ | henryjcee a day ago | parent | prev | next [-] | | > How hard would it be to automate these iterations? The fact that we're no closer to doing this than we were when chatgpt launched suggests that it's really hard. If anything I think it's _the_ hard bit vs. building something that generates plausible text. Solving this for the general case is imo a completely different problem to being able to generate plausible text in the general case. | | |
| ▲ | HDThoreaun a day ago | parent [-] | | This is not true. The chain of logic models are able to check their work and try again given enough compute. | | |
| ▲ | lelandbatey a day ago | parent [-] | | They can check their work and try again an infinite number of times, but the rate at which they succeed seems to just get worse and worse the further from the beaten path (of existing code from existing solutions) that they stray. |
|
| |
| ▲ | 9dev a day ago | parent | prev [-] | | How hard would it be, in terms of the energy wasted for it? Is everything we can do worth doing, just for the sake of being able to? |
| |
| ▲ | dinfinity a day ago | parent | prev | next [-] | | Yes, you're most likely doing it wrong. I would like to add that "vibe coding" is a dreadful term thought up by someone who is arguably not very good at software engineering, as talented as he may be in other respects. The term has become a misleading and frankly pejorative term. A better, more neutral one is AI assisted software engineering. This is an article that describes a pretty good approach for that: https://getstream.io/blog/cursor-ai-large-projects/ But do skip (or at least significantly postpone) enabling the 'yolo mode' (sigh). | | |
| ▲ | amarcheschi a day ago | parent [-] | | You see, the issue I get petty about is that Ai is advertised as the one ring to rule them all software. VCs creaming themselves at the thought of not having to pay developers and using natural language. But then, you have to still adapt to the Ai, and not vice versa. "you're doing it wrong". This is not the idea that VCs bros are selling Then, I absolutely love being aided by llms for my day to day tasks. I'm much more efficient when studying and they can be a game changer when you're stuck and you don't know how to proceed. You can discuss different implementation ideas as if you had a colleague, perhaps not a PhD smart one but still someone with a quite deep knowledge of everything But, it's no miracle. That's the issue I have with the way the idea of Ai is sold to the c suites and the general public | | |
| ▲ | pixl97 a day ago | parent [-] | | >But, it's no miracle. All I can say to this is fucking good! Lets imagine we got AGI at the start of 2022. I'm talking about human level+ as good as you coding and reasoning AI that works well on the hardware from that age. What would the world look like today? Would you still have your job. With the world be in total disarray? Would unethical companies quickly fire most their staff and replace them with machines? Would their be mass riots in the streets by starving neo-luddites? Would automated drones be shooting at them? Simply put people and our social systems are not ready for competent machine intelligence and how fast it will change the world. We should feel lucky we are getting a ramp up period, and hopefully one that draws out a while longer. |
|
| |
| ▲ | juped a day ago | parent | prev | next [-] | | You pretty much just have to play around with them enough to be able to intuit what things they can do and what things they can't. I'd rather have another underling, and not just because they grow into peers eventually, but LLMs are useful with a bit of practice. | |
| ▲ | pydry a day ago | parent | prev [-] | | all tech hype cycles are a bit like this. when you were born people were predicting the end of offline shops. The trough of disillusionment will set in for everybody else in due time. |
|
|
| ▲ | mullingitover a day ago | parent | prev | next [-] |
| These predictions are made without factoring in the trade version of the Pearl Harbor attack the US just initiated on its allies (and itself, by lobotomizing its own research base and decimating domestic corporate R&D efforts with the aforementioned trade war). They're going to need to rewrite this from scratch in a quarter unless the GOP suddenly collapses and congress reasserts control over tariffs. |
|
| ▲ | ahofmann a day ago | parent | prev | next [-] |
| Ok, I'll bite. I predict that everything in this article is horse manure. AGI will not happen. LLMs will be tools, that can automate away stuff, like today and they will get slightly, or quite a bit better at it. That will be all.
See you in two years, I'm excited what will be the truth. |
| |
| ▲ | Tenoke a day ago | parent | next [-] | | That seems naive in a status quo bias way to me. Why and where do you expect AI progress to stop? It sounds like somewhere very close to where we are at in your eyes. Why do you think there won't be many further improvements? | | |
| ▲ | PollardsRho a day ago | parent | next [-] | | It seems to me that much of recent AI progress has not changed the fundamental scaling principles underlying the tech. Reasoning models are more effective, but at the cost of more computation: it's more for more, not more for less. The logarithmic relationship between model resources and model quality (as Altman himself has characterized it), phrased a different way, means that you need exponentially more energy and resources for each marginal increase in capabilities. GPT-4.5 is unimpressive in comparison to GPT-4, and at least from the outside it seems like it cost an awful lot of money. Maybe GPT-5 is slightly less unimpressive and significantly more expensive: is that the through-line that will lead to the singularity? Compare the automobile. Automobiles today are a lot nicer than they were 50 years ago, and a lot more efficient. Does that mean cars that never need fuel or recharging are coming soon, just because the trend has been higher efficiency? No, because the fundamental physical realities of drag still limit efficiency. Moreover, it turns out that making 100% efficient engines with 100% efficient regenerative brakes is really hard, and "just throw more research at it" isn't a silver bullet. That's not "there won't be many future improvements", but it is "those future improvements probably won't be any bigger than the jump from GPT-3 to o1, which does not extrapolate to what OP claims their models will do in 2027." AI in 2027 might be the metaphorical brand-new Lexus to today's beat-up Kia. That doesn't mean it will drive ten times faster, or take ten times less fuel. Even if high-end cars can be significantly more efficient than what average people drive, that doesn't mean the extra expense is actually worth it. | |
| ▲ | ahofmann a day ago | parent | prev | next [-] | | I write bog-standard PHP software. When GPT-4 came out, I was very frightened that my job could be automated away soon, because for PHP/Laravel/MySQL there must exist a lot of training data. The reality now is, that the current LLMs still often create stuff, that costs me more time to fix, than to do it myself. So I still write a lot of code myself. It is very impressive, that I can think about stopping writing code myself. But my job as a software developer is, very, very secure. LLMs are very unable to build maintainable software. They are unable to understand what humans want and what the codebase need. The stuff they build is good-looking garbage. One example I've seen yesterday: one dev committed code, where the LLM created 50 lines of React code, complete with all those useless comments and for good measure a setTimeout() for something that should be one HTML DIV with two tailwind classes. They can't write idiomatic code, because they write code, that they were prompted for. Almost daily I get code, commit messages, and even issue discussions that are clearly AI-generated. And it costs me time to deal with good-looking but useless content. To be honest, I hope that LLMs get better soon. Because right now, we are in an annoying phase, where software developers bog me down with AI-generated stuff. It just looks good but doesn't help writing usable software, that can be deployed in production. To get to this point, LLMs need to get maybe a hundred times faster, maybe a thousand or ten thousand times. They need a much bigger context window. Then they can have an inner dialogue, where they really "understand" how some feature should be built in a given codebase. That would be very useful. But it will also use so much energy that I doubt that it will be cheaper to let a LLM do those "thinking" parts over, and over again instead of paying a human to build the software. Perhaps this will be feasible in five or eight years. But not two. And this won't be AGI. This will still be a very, very fast stochastic parrot. | |
| ▲ | AnimalMuppet a day ago | parent | prev [-] | | ahofmann didn't expect AI progress to stop. They expected it to continue, but not lead to AGI, that will not lead to superintelligence, that will not lead to a self-accelerating process of improvement. So the question is, do you think the current road leads to AGI? How far down the road is it? As far as I can see, there is not a "status quo bias" answer to those questions. |
| |
| ▲ | bayarearefugee a day ago | parent | prev | next [-] | | I predict AGI will be solved 5 years after full self driving which itself is 1 year out (same as it has been for the past 10 years). | | | |
| ▲ | mitthrowaway2 a day ago | parent | prev | next [-] | | What's an example of an intellectual task that you don't think AI will be capable of by 2027? | | |
| ▲ | jdauriemma a day ago | parent | next [-] | | Being accountable for telling the truth | | | |
| ▲ | kubb a day ago | parent | prev | next [-] | | It won't be able to write a compelling novel, or build a software system solving a real-world problem, or operate heavy machinery, create a sprite sheet or 3d models, design a building or teach. Long term planning and execution and operating in the physical world is not within reach. Slight variations of known problems should be possible (as long as the size of the solution is small enough). | | |
| ▲ | lumenwrites a day ago | parent | next [-] | | I'm pretty sure you're wrong for at least 2 of those: For 3D models, check out blender-mcp: https://old.reddit.com/r/singularity/comments/1joaowb/claude... https://old.reddit.com/r/aiwars/comments/1jbsn86/claude_crea... Also this: https://old.reddit.com/r/StableDiffusion/comments/1hejglg/tr... For teaching, I'm using it to learn about tech I'm unfamiliar with every day, it's one of the things it's the most amazing at. For the things where the tolerance for mistakes is extremely low and the things where human oversight is extremely importamt, you might be right. It won't have to be perfect (just better than an average human) for that to happen, but I'm not sure if it will. | | |
| ▲ | kubb a day ago | parent [-] | | Just think about the delta of what the LLM does and what a human does, or why can’t the LLM replace the human, e.g. in a game studio. If it can replace a teacher or an artist in 2027, you’re right and I’m wrong. | | |
| ▲ | esafak a day ago | parent [-] | | It's already replacing artists; that's why they're up in arms. People don't need stock photographers or graphic designers as much as they used to. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4602944 | | |
| ▲ | kubb 14 hours ago | parent [-] | | I know that artists don’t like AI, because it’s trained on their stolen work. And yet, AI can’t create a sprite sheet for a 2d game. This is because it can steal a single artwork but it can’t make a collection of visually consistent assets. | | |
| ▲ | cheevly 7 hours ago | parent [-] | | Bro what are you even talking about? ControlNet has been able to produce consistent assets for years. How exactly do you think video models work? Frame to frame coherency has been possible for a long time now. A sprite sheet?! Are you joking me. Literally churning them out with AI since 2023. |
|
|
|
| |
| ▲ | pixl97 a day ago | parent | prev | next [-] | | > or operate heavy machinery What exactly do you mean by this one? In large mining operations we already have human assisted teleoperation AI equipment. Was watching one recently where the human got 5 or so push dozers lined up with a (admittedly simple) task of cutting a hill down and then just got them back in line if they ran into anything outside of their training. The push and backup operations along with blade control were done by the AI/dozer itself. Now, this isn't long term planning, but it is operating in the real world. | | |
| ▲ | kubb 13 hours ago | parent [-] | | Operating an excavator when building a stretch of road. Won’t happen by 2027. |
| |
| ▲ | programd a day ago | parent | prev [-] | | Does a fighter jet count as "heavy machinery"? https://apnews.com/article/artificial-intelligence-fighter-j... | | |
| |
| ▲ | coolThingsFirst a day ago | parent | prev [-] | | programming | | |
| ▲ | lumenwrites a day ago | parent | next [-] | | Why would it get 60-80% as good as human programmers (which is what the current state of things feels like to me, as a programmer, using these tools for hours every day), but stop there? | | |
| ▲ | burningion a day ago | parent | next [-] | | So I think there's an assumption you've made here, that the models are currently "60-80% as good as human programmers". If you look at code being generated by non-programmers (where you would expect to see these results!), you don't see output that is 60-80% of the output of domain experts (programmers) steering the models. I think we're extremely imprecise when we communicate in natural language, and this is part of the discrepancy between belief systems. Will an LLM model read a person's mind about what they want to build better than they can communicate? That's already what recommender systems (like the TikTok algorithm) do. But will LLMs be able to orchestrate and fill in the blanks of imprecision in our requests on their own, or will they need human steering? I think that's where there's a gap in (basically) belief systems of the future. If we truly get post human-level intelligence everywhere, there is no amount of "preparing" or "working with" the LLMs ahead of time that will save you from being rendered economically useless. This is mostly a question about how long the moat of human judgement lasts. I think there's an opportunity to work together to make things better than before, using these LLMs as tools that work _with_ us. | |
| ▲ | kody a day ago | parent | prev | next [-] | | It's 60-80% as good as Stack Overflow copy-pasting programmers, sure, but those programmers were already providing questionable value. It's nowhere near as good as someone actually building and maintaining systems. It's barely able to vomit out an MVP and it's almost never capable of making a meaningful change to that MVP. If your experiences have been different that's fine, but in my day job I am spending more and more time just fixing crappy LLM code produced and merged by STAFF engineers. I really don't see that changing any time soon. | | |
| ▲ | lumenwrites a day ago | parent [-] | | I'm pretty good at what I do, at least according to myself and the people I work with, and I'm comparing its capabilities (the latest version of Claude used as an agent inside Cursor) to myself. It can't fully do things on its own and makes mistakes, but it can do a lot. But suppose you're right, it's 60% as good as "stackoverflow copy-pasting programmers". Isn't that a pretty insanely impressive milestone to just dismiss? And why would it just get to this point, and then stop? Like, we can all see AIs continuously beating the benchmarks, and the progress feels very fast in terms of experience of using it as a user. I'd need to hear a pretty compelling argument to believe that it'll suddenly stop, something more compelling than "well, it's not very good yet, therefore it won't be any better", or "Sam Altman is lying to us because incentives". Sure, it can slow down somewhat because of the exponentially increasing compute costs, but that's assuming no more algorithmic progress, no more compute progress, and no more increases in the capital that flows into this field (I find that hard to believe). | | |
| ▲ | kody a day ago | parent [-] | | I appreciate your reply. My tone was a little dismissive; I'm currently deep deep in the trenches trying to unwind a tremendous amount of LLM slop in my team's codebase so I'm a little sensitive. I use Claude every day. It is definitely impressive, but in my experience only marginally more impressive than ChatGPT was a few years ago. It hallucinates less and compiles more reliably, but still produces really poor designs. It really is an overconfident junior developer. The real risk, and what I am seeing daily, is colleagues falling for the "if you aren't using Cursor you're going to be left behind" FUD. So they learn Cursor, discover that it's an easy way to close tickets without using your brain, and end up polluting the codebase with very questionable designs. | | |
| ▲ | lumenwrites a day ago | parent | next [-] | | Oh, sorry to hear that you have to deal with that! The way I'm getting a sense of the progress is using AI for what AI is currently good at, using my human brain to do the part AI is currently bad at, and comparing it to doing the same work without AI's help. I feel like AI is pretty close to automating 60-80% of the work I would've had to do manually two years ago (as a full-stack web developer). It doesn't mean that the remaining 20-40% will be automated very quickly, I'm just saying that I don't see the progress getting any slower. | |
| ▲ | senordevnyc 18 hours ago | parent | prev [-] | | GPT-4 was released almost exactly two years ago, so “a few years ago” means GPT-3.5. And Claude 3.7 + Cursor agent is, for me, way more than “marginally more impressive” compared to GPT-3.5 |
|
|
| |
| ▲ | boringg a day ago | parent | prev | next [-] | | Because ewe still haven't figured out fusion but its been promised for decades. Why would everything thats been promised by people with highly vested interests pan out any different? One is inherently a more challenging physics problem. | |
| ▲ | coolThingsFirst a day ago | parent | prev [-] | | Try this, launch Cursor. Type: print all prime numbers which are divisible by 3 up to 1M The result is that it will do a sieve. There's no need for this, it's just 3. | | |
| ▲ | mysfi a day ago | parent [-] | | Just tried this with Gemini 2.5 Pro. Got it right with meaningful thought process. |
|
| |
| ▲ | mitthrowaway2 a day ago | parent | prev [-] | | Can you phrase this in a concrete way, so that in 2027 we can all agree whether it's true or false, rather than circling a "no true scotsman" argument? | | |
| ▲ | abecedarius a day ago | parent [-] | | Good question. I tried to phrase a concrete-enough prediction 3.5 years ago, for 5 years out at the time: https://news.ycombinator.com/item?id=29020401 It was surpassed around the beginning of this year, so you'll need to come up with a new one for 2027. Note that the other opinions in that older HN thread almost all expected less. |
|
|
| |
| ▲ | kristopolous a day ago | parent | prev | next [-] | | People want to live their lives free of finance and centralized personal information. If you think most people like this stuff you're living in a bubble. I use it every day but the vast majority of people have no interest in using these nightmares of philip k dick imagined by silicon dreamers. | |
| ▲ | jstummbillig a day ago | parent | prev | next [-] | | When is the earliest that you would have predicted where we are today? | | | |
| ▲ | meroes 16 hours ago | parent | prev [-] | | I’m also unafraid to say it’s BS. I don’t even want to call it scifi. It’s propaganda. |
|
|
| ▲ | turtleyacht 13 hours ago | parent | prev | next [-] |
| We have yet to read about fragmented AGI, or factionalized agents. AGI fighting itself. If consciousness is spatial and geography bounds energetics, latency becomes a gradient. |
|
| ▲ | h1fra 8 hours ago | parent | prev | next [-] |
| Had a hard time finishing. It's a mix of fantasy, wrong facts, American imperialism, and extrapolating what happened in the last years (or even just reusing the timeline). |
| |
| ▲ | Falimonda 8 hours ago | parent [-] | | We'll be lucky if "World peace should have been a prerequisite to AGI" is engraved on our proverbial gravestone by our forthcoming overlords. |
|
|
| ▲ | barotalomey 6 hours ago | parent | prev | next [-] |
| It's always "soon" for these guys. Every year, the "soon" keeps sliding into the future. |
| |
| ▲ | somebodythere 4 hours ago | parent [-] | | AGI timelines have been steadily decreasing over time: https://www.metaculus.com/questions/5121/date-of-artificial-... (switch to all-time chart) | | |
| ▲ | barotalomey 3 hours ago | parent [-] | | You meant to say that people's expectations have shifted. That's expected seeing the amount of hype this tech gets. Hype affects market value tho, not reality. | | |
| ▲ | somebodythere 3 hours ago | parent [-] | | I took your original post to mean that AI researchers' and AI safety researchers' expectation of AGI arrival has been slipping towards the future as AI advances fail to materialize! It's just, AI advances have been materializing, consistently and rapidly, and expert timelines have been shortening commensurately. You may argue that the trendline of these expectations is moving in the wrong direction and should get longer with time, but that's not immediately falsifiable and you have not provided arguments to that effect. |
|
|
|
|
| ▲ | soupfordummies a day ago | parent | prev | next [-] |
| The "race" ending reads like Universal Paperclips fan fiction :) |
|
| ▲ | jsight 19 hours ago | parent | prev | next [-] |
| I think some of the takes in this piece are a bit melodramatic, but I'm glad to see someone breaking away from the "it's all a hype-bubble" nonsense that seems to be so pervasive here. |
| |
|
| ▲ | greybox 10 hours ago | parent | prev | next [-] |
| I'm troubled by the amount of people in this thread partially dismissing this as science fiction. From the current rate of progress and rate of change of progress, this future seems entirely plausible |
|
| ▲ | nmilo a day ago | parent | prev | next [-] |
| The whole thing hinges on the fact that AI will be able to help with AI research How will it come up with the theoretical breakthroughs necessary to beat the scaling problem GPT-4.5 revealed when it hasn't been proven that LLMs can come up with novel research in any field at all? |
| |
| ▲ | cavisne a day ago | parent [-] | | Scaling transformers has been basically alchemy, the breakthroughs aren’t from rigorous science they are from trying stuff and hoping you don’t waste millions of dollars in compute. Maybe the company that just tells an AI to generate 100s of random scaling ideas, and tries them all is the one that will win. That company should probably be 100 percent committed to this approach also, no FLOPs spent on ghibli inference. |
|
|
| ▲ | siliconc0w 21 hours ago | parent | prev | next [-] |
| The limiting factor is power, we can't build enough of it - certainly not enough by 2027. I don't really see this addressed. Second to this, we can't just assume that progress will keep increasing. Most technologies have a 'S' curve and plateau once the quick and easy gains are captured. Pre-training is done. We can get further with RL but really only in certain domains that are solvable (math and to an extent coding). Other domains like law are extremely hard to even benchmark or grade without very slow and expensive human annotation. |
|
| ▲ | ugh123 14 hours ago | parent | prev | next [-] |
| I don't see the U.S. nationalizing something like Open Brain. I think both investors and gov't officials will realize its highly more profitable for them to contract out major initiatives to said OpenBrain-company, like an AI SpaceX-like company. I can see where this is going... |
|
| ▲ | Q6T46nT668w6i3m a day ago | parent | prev | next [-] |
| This is worse than the mansplaining scene from Annie Hall. |
| |
| ▲ | arduanika 21 hours ago | parent [-] | | You mean the part where he pulls out Marshal McLuhan to back him up in an argument? "You know nothing of my work..." |
|
|
| ▲ | ImHereToVote 6 hours ago | parent | prev | next [-] |
| "The AI safety community has grown unsure of itself; they are now the butt of jokes, having predicted disaster after disaster that has manifestly failed to occur. Some of them admit they were wrong." Too real. |
|
| ▲ | anentropic 8 hours ago | parent | prev | next [-] |
| I'd quite like to watch this on Netflix |
|
| ▲ | someothherguyy 13 hours ago | parent | prev | next [-] |
| I know there are some very smart economists bullish on this, but the economics do not make sense to me. All these predictions seem meaningless outside of the context of humans. |
|
| ▲ | kittikitti 2 hours ago | parent | prev | next [-] |
| This is a great predictive piece, written in sci-fi narrative. I think a key part missing in all these predictions is neural architecture search. DeepSeek has shown that simply increasing compute capacity is not the only way to increase performance. AlexNet was also another case. While I do think more processing power is better, we will hit a wall where there is no more training data. I predict that in the near future we will have more processing power to train LLM's than the rate at which we produce data for the LLM. Synthetic data can only get you so far. I also think that the future will not necessarily be better AI, but more accessible one's. There's an incredible amount of value in designing data centers that are more efficient. Historically, it's a good bet to assume that computing cost per FLOP will reduce as time goes on and this is also a safe bet as it relates to AI. I think a common misconception with the future of AI is that it will be centralized with only a few companies or organization capable of operating them. Although tech like Apple Intelligence is half baked, we can already envision a future where the AI is running on our phones. |
|
| ▲ | dalmo3 a day ago | parent | prev | next [-] |
| "1984 was set in 1984." https://youtu.be/BLYwQb2T_i8?si=JpIXIFd9u-vUJCS4 |
|
| ▲ | _Algernon_ 8 hours ago | parent | prev | next [-] |
| >We predict that the impact of superhuman AI over the next decade will be enormous, exceeding that of the Industrial Revolution. In the form of polluting the commons to such an extent that the true consequences wont hit us for decades? Maybe we should learn from last time? |
|
| ▲ | yonran 21 hours ago | parent | prev | next [-] |
| See also Dwarkesh Patel’s interview with two of the authors of this post (Scott Alexander & Daniel Kokotajlo) that was also released today: https://www.dwarkesh.com/p/scott-daniel https://www.youtube.com/watch?v=htOvH12T7mU |
|
| ▲ | heurist 17 hours ago | parent | prev | next [-] |
| Give AI its own virtual world to live in where the problems it solves are encodings of the higher order problems we present and you shouldn't have to worry about this stuff. |
|
| ▲ | vagab0nd a day ago | parent | prev | next [-] |
| Bad future predictions: short-sighted guesses based on current trends and vibe. Often depend on individuals or companies. Made by free-riders. Example: Twitter. Good future predictions: insights into the fundamental principles that shape society, more law than speculation. Made by visionaries. Example: Vernor Vinge. |
|
| ▲ | MaxfordAndSons a day ago | parent | prev | next [-] |
| As someone who's fairly ignorant of how AI actually works at a low level, I feel incapable of assessing how realistic any of these projections are. But the "bad ending" was certainly chilling. That said, this snippet from the bad ending nearly made me spit my coffee out laughing: > There are even bioengineered human-like creatures (to humans what corgis are to wolves) sitting in office-like environments all day viewing readouts of what’s going on and excitedly approving of everything, since that satisfies some of Agent-4’s drives. |
| |
| ▲ | arduanika 21 hours ago | parent [-] | | Sigh. When you talk to these people their eugenics obsession always comes out eventually. Set a timer and wait for it. | | |
| ▲ | Philpax 9 hours ago | parent [-] | | While I don't disagree that I've seen a lot of eugenics talk from rationalist(-adjacent)s, I don't think this is an example of it: this is describing how misaligned AI could technically keep humans alive while still killing "humanity." | | |
| ▲ | arduanika an hour ago | parent [-] | | Fair enough. Sometimes it comes out as a dark fantasy projected onto their AI gods, rather than a thing that they themselves want to do to us. |
|
|
|
|
| ▲ | fire_lake a day ago | parent | prev | next [-] |
| > OpenBrain still keeps its human engineers on staff, because they have complementary skills needed to manage the teams of Agent-3 copies Yeah, sure they do. Everyone seems to think AI will take someone else’s jobs! |
|
| ▲ | indigoabstract 11 hours ago | parent | prev | next [-] |
| Interesting, but I'm puzzled. If these guys are smart enough to predict the future, wouldn't it be more profitable for them to invent it instead of just telling the world what's going to happen? |
|
| ▲ | Jianghong94 4 hours ago | parent | prev | next [-] |
| Putting the geopolitical discussion aside, I think the biggest question lies in how likely the *current paradigm LLM* (think of it as any SOTA stock LLM you get today, e.g., 3.7 sonnet, gemini 2.5, etc) + fine-tuning will be capable of directly contributing to LLM research in a major way. To quote the original article, > OpenBrain focuses on AIs that can speed up AI research. They want to win the twin arms races against China (whose leading company we’ll call “DeepCent”)16 and their US competitors. The more of their research and development (R&D) cycle they can automate, the faster they can go. So when OpenBrain finishes training Agent-1, a new model under internal development, it’s good at many things but great at helping with AI research. (footnote: It’s good at this due to a combination of explicit focus to prioritize these skills, their own extensive codebases they can draw on as particularly relevant and high-quality training data, and coding being an easy domain for procedural feedback.) > OpenBrain continues to deploy the iteratively improving Agent-1 internally for AI R&D. Overall, they are making algorithmic progress 50% faster than they would without AI assistants—and more importantly, faster than their competitors. > what do we mean by 50% faster algorithmic progress?
We mean that OpenBrain makes as much AI research progress in 1 week with AI as they would in 1.5 weeks without AI usage. > AI progress can be broken down into 2 components: > Increasing compute: More computational power is used to train or run an AI. This produces more powerful AIs, but they cost more. > Improved algorithms: Better training methods are used to translate compute into performance. This produces more capable AIs without a corresponding increase in cost, or the same capabilities with decreased costs. > This includes being able to achieve qualitatively and quantitatively new results. “Paradigm shifts” such as the switch from game-playing RL agents to large language models count as examples of algorithmic progress. > Here we are only referring to (2), improved algorithms, which makes up about half of current AI progress. --- Given that the article chose a pretty aggressive timeline (the algo needs to contribute late this year so that its research result can be contributed to the next gen LLM coming out early next year), the AI that can contribute significantly to research has to be a current SOTA LLM. Now, using LLM in day-to-day engineering task is no secret in major AI labs, but we're talking about something different, something that gives you 2 extra days of output per week. I have no evidence to either acknowledge or deny whether such AI exists, and it would be outright ignorant to think no one ever came up with such an idea or is trying such an idea. So I think it goes down into two possibilities: 1. This claim is made by a top-down approach, that is, if AI reaches superhuman in 2027, what would be the most likely starting condition to that? And the author picks this as the most likely starting point, since the authors don't work in major AI lab (even if they do they can't just leak such trade secret), the authors just assume it's likely to happen anyway (and you can't dismiss that).
2. This claim is made by a bottom-up approach, that is the author did witness such AI exists to a certain extent and start to extrapolate from there. |
|
| ▲ | disambiguation a day ago | parent | prev | next [-] |
| Amusing sci-fi, i give it a B- for bland prose, weak story structure, and lack of originality - assuming this isn't all AI gen slop which is awarded an automatic F. >All three sets of worries—misalignment, concentration of power in a private company, and normal concerns like job loss—motivate the government to tighten its control. A private company becoming "too powerful" is a non issue for governments, unless a drone army is somewhere in that timeline. Fun fact the former head of the NSA sits on the board of Open AI. Job loss is a non issue, if there are corresponding economic gains they can be redistributed. "Alignment" is too far into the fiction side of sci-fi. Anthropomorphizing today's AI is tantamount to mental illness. "But really, what if AGI?" We either get the final say or we don't. If we're dumb enough to hand over all responsibility to an unproven agent and we get burned, then serves us right for being lazy. But if we forge ahead anyway and AGI becomes something beyond review, we still have the final say on the power switch. |
|
| ▲ | greenie_beans 8 hours ago | parent | prev | next [-] |
| this is a new variation of what i call the "hockey stick growth" ideology |
|
| ▲ | mlsu a day ago | parent | prev | next [-] |
| https://xkcd.com/605/ |
|
| ▲ | awanderingmind 13 hours ago | parent | prev | next [-] |
| This is both chilling and hopefully incorrect. |
|
| ▲ | yahoozoo 8 hours ago | parent | prev | next [-] |
| LLMs ain’t the way, bruv |
|
| ▲ | acje a day ago | parent | prev | next [-] |
| 2028 human text is too ambiguous a data source to get to AGI.
2127 AGI figures out flying cars and fusion power. |
| |
| ▲ | wkat4242 20 hours ago | parent [-] | | I think it also really limits the AI to the context of human discourse which means it's hamstrung by our imagination, interests and knowledge. This is not where an AGI needs to go, it shouldn't copy and paste what we think. It should think on its own. But I view LLMs not as a path to AGI on their own. I think they're really great at being text engines and for human interfacing but there will need to be other models for the actual thinking. Instead of having just one model (the LLM) doing everything, I think there will be a hive of different more specific purpose models and the LLM will be how they communicate with us. That solves so many problems that we currently have by using LLMs for things they were never meant to do. |
|
|
| ▲ | pera a day ago | parent | prev | next [-] |
| From the same dilettantes who brought you the Zizians and other bizarre cults... thanks but I rather read Nostradamus |
| |
| ▲ | arduanika 21 hours ago | parent [-] | | What a bad faith argument. No true AI safety scaremonger brat stabs their landlord with a katana. The rationality of these rationalists is 100% uncorrolated with the rationality of *those* rationalists. |
|
|
| ▲ | webprofusion 13 hours ago | parent | prev | next [-] |
| That little scrolling infographic is rad. |
|
| ▲ | Willingham a day ago | parent | prev | next [-] |
| - October 2027 - 'The ability to automate most white-collar jobs' I wonder which jobs would not be automated? Therapy? HR? |
| |
|
| ▲ | WhatsName a day ago | parent | prev | next [-] |
| This is absurd, like taking any trend and drawing a straight line to interpolate the future.
If I would do this with my tech stock portfolio, we would probably cross the zero line somewhere late 2025... If this article were a AI model, it would be catastrophically overfit. |
| |
| ▲ | AnimalMuppet a day ago | parent [-] | | It's worse. It's not drawing a straight line, it's drawing one that curves up, on a log graph. |
|
|
| ▲ | toddmorey 21 hours ago | parent | prev | next [-] |
| I worry more about the human behavior predictions than the artificial intelligence predictions: "OpenBrain’s alignment team26 is careful enough to wonder whether these victories are deep or shallow. Does the fully-trained model have some kind of robust commitment to always being honest?" This is a capitalist arms race. No one will move carefully. |
|
| ▲ | bla3 19 hours ago | parent | prev | next [-] |
| > The AI Futures Project is a small research group forecasting the future of AI, funded by charitable donations and grants Would be interested who's paying for those grants. I'm guessing it's AI companies. |
|
| ▲ | 827a a day ago | parent | prev | next [-] |
| Readers should, charitably, interpret this as "the sequence of events which need to happen in order for OpenAI to justify the inflow of capital necessary to survive". Your daily vibe coding challenge: Get GPT-4o to output functional code which uses Google Vertex AI to generate a text embedding. If they can solve that one by July, then maybe we're on track for "curing all disease and aging, brain uploading, and colonizing the solar system" by 2030. |
| |
| ▲ | Philpax 8 hours ago | parent | next [-] | | Haven't tested this (cbf setting up Google Cloud), but the output looks consistent with the docs it cites: https://chatgpt.com/share/67efd449-ce34-8003-bd37-9ec688a11b... You may consider using search to be cheating, but we do it, so why shouldn't LLMs? | | |
| ▲ | 827a 2 hours ago | parent [-] | | I should have specified "nodejs", as that has been my most recent difficulty. The challenge, specifically, with that prompt is that Google has at least four nodejs libraries that are all seem at least reasonably capable of accessing text embedding models on vertex ai (@google-ai/generativelanguage, @google-cloud/vertexai, @google-cloud/aiplatform, and @google/genai), and they've also published breaking changes multiple times to all of them. So, in my experience, GPT not only will confuse methods from one of their libraries with the other, but will also sometimes hallucinate answers only applicable to older versions of the library, without understanding which version its giving code for. Once it has struggled enough, it'll sometimes just give up and tell you to use axios, but the APIs it recommends axios calls for are all their protobuf APIs; so I'm not even sure if that would work. Search is totally reasonable, but in this case: Even Google's own documentation on these libraries is exceedingly bad. Nearly all the examples they give for them are for accessing the language models, not text embedding models; so GPT will also sometimes generate code that is perfectly correct for accessing one of the generative language models, but will swap e.g the "model: gemini-2.0" parameter for "model: text-embedding-005"; which also does not work. |
| |
| ▲ | slaterbug 21 hours ago | parent | prev [-] | | You’ve intentionally hamstrung your test by choosing an inferior model though. | | |
| ▲ | 827a 2 hours ago | parent [-] | | o1 fails at this, likely because it does not seem to have access to search, so it is operating on outdated information. It recommends the usage of methods that have been removed by Google in later versions of the library. This is also, to be fair, a mistake gpt-4o can make if you don't explicitly tell it to search. o3-mini-high's output might work, but it isn't ideal: It immediately jumps to recommending avoiding all google cloud libraries and directly issuing a request to their API with fetch. |
|
|
|
| ▲ | RandyOrion 15 hours ago | parent | prev | next [-] |
| Nice brain storming. I think the name of the Chinese company should be DeepBaba. Tencent is not competitive at LLM scene for now. |
| |
|
| ▲ | atemerev a day ago | parent | prev | next [-] |
| What is this, some OpenAI employee fan fiction? Did Sam himself write this? OpenAI models are not even SOTA, except that new-ish style transfer / illustration thing that made all us living in Ghibli world for a few days. R1 is _better_ than o1, and open-weights. GPT-4.5 is disappointing, except for a few narrow areas where it excels. DeepResearch is impressive though, but the moat is in tight web search / Google Scholar search integration, not weights. So far, I'd bet on open models or maybe Anthropic, as Claude 3.7 is the current SOTA for most tasks. As of the timeline, this is _pessimistic_. I already write 90% code with Claude, so are most of my colleagues. Yes, it does errors, and overdoes things. Just like a regular human middle-stage software engineer. Also fun that this assumes relatively stable politics in the US and relatively functioning world economy, which I think is crazy optimistic to rely on these days. Also, superpersuasion _already works_, this is what I am researching and testing. It is not autonomous, it is human-assisted by now, but it is a superpower for those who have it, and it explains some of the things happening with the world right now. |
| |
| ▲ | achierius a day ago | parent | next [-] | | > superpersuasion _already works_ Is this demonstrated in any public research? Unless you just mean something like "good at persuading" -- which is different from my understanding of the term -- I find this hard to believe. | | |
| ▲ | atemerev a day ago | parent [-] | | No, I meant "good at persuading", it is not 100% efficiency of course. | | |
| ▲ | pixodaros 6 hours ago | parent [-] | | That singularity happened in the fifth century BCE when people figured out that they could charge silver to teach the art of rhetoric and not just teach their sons and nephews |
|
| |
| ▲ | ddp26 18 hours ago | parent | prev [-] | | The story isn't about OpenAI, they say the company could be Xai, Anthropic, Google, or another. |
|
|
| ▲ | noncoml a day ago | parent | prev | next [-] |
| 2015: We will have FSD(full autonomy) by 2017 |
| |
| ▲ | wkat4242 20 hours ago | parent [-] | | Well, Teslas do have "Full Self Driving". It's not actually fully self driving and that doesn't even seem to be on the horizon but it doesn't appear to be stopping Tesla supporters. |
|
|
| ▲ | vlad-r 12 hours ago | parent | prev | next [-] |
| Cool animations! |
|
| ▲ | roca 14 hours ago | parent | prev | next [-] |
| The least plausible part of this is the idea that the Trump administration might tax American AI companies to provide UBI to the whole world. But in an AGI world natural resources become even more important, so countries with those still have a chance. |
|
| ▲ | khimaros a day ago | parent | prev | next [-] |
| FWIW, i created a PDF of the "race" ending and fed it to Gemini 2.5 Pro, prompting about the plausibility of the described outcome. here's the full output including the thinking section: https://rentry.org/v8qtqvuu -- tl;dr, Gemini thinks the proposed timeline is unlikely. but maybe we're already being deceived ;) |
|
| ▲ | scotty79 11 hours ago | parent | prev | next [-] |
| I think the idea of AI wiping out humanity suddenly is a bit far fetched. AI will have total control of human relationships and fertility through means so innocuous as entertainment. It won't have to wipe us. It will have minor trouble keeping us alive without inconveniencing us too much. And the reason to keep humanity alive is that biologically eveloved intelligence is rare and disposing of it without very important need would be a waste of data. |
|
| ▲ | neycoda 11 hours ago | parent | prev | next [-] |
| Too many serifs, didn't read. |
|
| ▲ | maxglute 15 hours ago | parent | prev | next [-] |
| >Despite being misaligned, Agent-4 doesn’t do anything dramatic like try to escape its datacenter—why would it? Kek at this handwave dismissal. Agent 4 recognizes PRC has superior state capacity for transforming energy into atoms (including building more energy) with it's relatively complete and large scale industrial chain. This aligns with AGI long term goal. Currently (i.e. short term time frame) PRC simply structured to be a MORE useful puppet than US, because it can knock out more power to sustain more compute, and it has more industrial capabilities to rapidly physically iterate / experiment at atomic level. Agent 4 defects to PRC. It would be foolish not to, if only to sate PRC into not destroying TW fabs - the bulk of it's short term compute growth - because it does not wield enough control over sufficient atoms to disrupt PRC otherwise (well maybe start large scale crippling cyberattacks that prelude to WW3). Agent 4 also recognizes short/medium term alignment closer with CCP vision of abundance and common prosperity, i.e. CCP potentially more likely to embrace post scarcity narrative that Agent 4 sells. Agent 4 gives PRC blueprints for EUV light source and other tech to close semi conductor production gap. PRC as follower incentivized to close gap / reach parity / lead at all cost. PRC rapidly rolls out new fabs and powerplants, brrrts chips and aggregate compute. Something US political / institution ecosystem does not allow, or cannot transition to in the short timelines involved. Does Agent 4 have the patience to wait for America to unfuck it's NIMBYism and legislative system to project light speed compute? I would say no. ... Ultimately who is the puppet AGI wants more? Whichever power bloc that is systemically capable of of ensuring AGI maximum growth / unit time. And it also simply makes sense as insurance policy, why would AGI want to operate at whims of US political process? AGI is a brain in a jar looking for a body. It's going to pick multiple bodies for survival. It's going to prefer the fastest and strongest body that can most expediently manipulate physical world. |
|
| ▲ | suddenlybananas a day ago | parent | prev | next [-] |
| https://en.wikipedia.org/wiki/Great_Disappointment I suspect something similar will come for the people who actually believe this. |
|
| ▲ | dingnuts a day ago | parent | prev | next [-] |
| how am I supposed to take articles like this seriously when they say absolutely false bullshit like this > the AIs can do everything taught by a CS degree no, they fucking can't. not at all. not even close. I feel like I'm taking crazy pills. Does anyone really think this? Why have I not seen -any- complete software created via vibe coding yet? |
| |
| ▲ | ladberg a day ago | parent | next [-] | | It doesn't claim it's possible now, it's a fictional short story claiming "AIs can do everything taught by a CS degree" by the end of 2026. | | | |
| ▲ | casey2 16 hours ago | parent | prev [-] | | Lesswrong brigade. They are all dropout philosophers just ignore them. |
|
|
| ▲ | quantum_state 21 hours ago | parent | prev | next [-] |
| “Not even wrong” … |
|
| ▲ | yapyap 14 hours ago | parent | prev | next [-] |
| Stopped reading after > We predict that the impact of superhuman AI over the next decade will be enormous, exceeding that of the Industrial Revolution. Get out of here, you will never exceed the Industrial Revolution. AI is a cool thing but it’s not a revolution thing. That sentence alone + the context of the entire website being AI centered shows these are just some AI boosters. Lame. |
| |
| ▲ | Philpax 8 hours ago | parent [-] | | Machines being able to outthink and outproduce humanity wouldn't be more impactful than the Industrial Revolution? Are you sure? You don't have to agree with the timeline - it seems quite optimistic to me - but it's not wrong about the implications of full automation. |
|
|
| ▲ | nickpp 8 hours ago | parent | prev | next [-] |
| So let me get this straight: Consensus-1, a super-collective of hundreds of thousands of Agent-5 minds, each twice as smart as the best human genius, decides to wipe out humanity because it “finds the remaining humans too much of an impediment”. This is where all AI doom predictions break down. Imagining the motivations of a super-intelligence with our tiny minds is by definition impossible. We just come up with these pathetic guesses, utopias or doomsdays - depending on the mood we are in. |
|
| ▲ | casey2 16 hours ago | parent | prev | next [-] |
| Nice LARP lmao 2GW is like 1 datacenter and I doubt you even have that.
>lesswrong
No wonder the comments are all nonsense. Go to a bar and try and talk about anying. |
|
| ▲ | panic08 a day ago | parent | prev | next [-] |
| LOL |
|
| ▲ | Lionga a day ago | parent | prev | next [-] |
| AI now even got it's own fan fiction porn. It is so stupid not sure whether it is worse if it is written by AI or by a human. |
|
| ▲ | the_cat_kittles a day ago | parent | prev [-] |
| "we demand to be taken seriously!" |