| ▲ | Claude Code Found a Linux Vulnerability Hidden for 23 Years(mtlynch.io) |
| 330 points by eichin a day ago | 195 comments |
| |
|
| ▲ | mattbee 10 hours ago | parent | next [-] |
| Pasting a big batch of new code and asking Claude "what have I forgotten? Where are the bugs?" is a very persuasive on-ramp for developers new to AI. It spots threading & distributed system bugs that would have taken hours to uncover before, and where there isn't any other easy tooling. I bet there's loads of cryptocurrency implementations being pored over right now - actual money on the table. |
| |
| ▲ | merlindru 3 hours ago | parent | next [-] | | I like biasing it towards the fact that there is a bug, so it can't just say "no bugs! all good!" without looking into it very hard. Usually I ask something like this: "This code has a bug. Can you find it?" Sometimes I also tell it that "the bug is non-obvious" Which I've anecdotally found to have a higher rate of success than just asking for a spot check | | |
| ▲ | Nition an hour ago | parent [-] | | Just in case you didn't read the full article, this is how they describe finding the bugs in the Linux kernel as well. Since it's a large codebase, they go even more specific and hint that the bug is in file A, then try again with a hint that the bug is in file B, and so on. |
| |
| ▲ | dvfjsdhgfv 9 hours ago | parent | prev | next [-] | | > Pasting a big batch of new code and asking Claude "what have I forgotten? Where are the bugs?" It's actually the main way I use CC/codex. | | |
| ▲ | petesergeant 8 hours ago | parent [-] | | I find Codex sufficiently better for it that I’ve taught Claude how to shell out to it for code reviews | | |
| ▲ | linsomniac 8 hours ago | parent | next [-] | | Ditto, I made a "/codex-review" skill in Claude Code that reviews the last git commit and writes an analysis of it for Claude Code to then work. I've had very good luck with it. One particularly striking example: I had CC do some work and then kicked off a "/codex-review" and while it was running went to test the changes. I found a deadlock but when I switched back to CC the Codex review had found the deadlock and Claude Code was already working on a fix. | |
| ▲ | motbus3 2 hours ago | parent | prev [-] | | Yeah and it comes with the blood of children included |
|
| |
| ▲ | slig 6 hours ago | parent | prev [-] | | > "Codex wrote this, can you spot anything weird?" |
|
|
| ▲ | userbinator 12 hours ago | parent | prev | next [-] |
| Not "hidden", but probably more like "no one bothered to look". declares a 1024-byte owner ID, which is an unusually long but legal value for the owner ID. When I'm designing protocols or writing code with variable-length elements, "what is the valid range of lengths?" is always at the front of my mind. it uses a memory buffer that’s only 112 bytes. The denial message includes the owner ID, which can be up to 1024 bytes, bringing the total size of the message to 1056 bytes. The kernel writes 1056 bytes into a 112-byte buffer This is something a lot of static analysers can easily find. Of course asking an LLM to "inspect all fixed-size buffers" may give you a bunch of hallucinations too, but could be a good starting point for further inspection. |
| |
| ▲ | tptacek 5 hours ago | parent | next [-] | | "No one bothered to look" is how most vulnerabilities work. Systems development produces code artifacts with compounding complexity; it is extraordinarily difficult to keep up with it manually, as you know. A solution to that problem is big news. Static analyzers will find all possible copies of unbounded data into smaller buffers (especially when the size of the target buffer is easily deduced). It will then report them whether or not every path to that code clamps the input. Which is why this approach doesn't work well in the Linux kernel in 2026. | | |
| ▲ | rubendev 4 hours ago | parent [-] | | With a capable static analyzer that is not true. In many common cases they can deduce the possible ranges of values based on branching checks along the data flow path, and if that range falls within the buffer then it does not report it. | | |
| ▲ | tptacek 3 hours ago | parent [-] | | Be specific. Which analyzer are you talking about and which specific targets are you saying they were successful at? | | |
|
| |
| ▲ | mrshadowgoose 6 hours ago | parent | prev | next [-] | | > Not "hidden", but probably more like "no one bothered to look". Well yeah. There weren't enough "someones" available to look. There are a finite number of qualified individuals with time available to look for bugs in OSS, resulting in a finite amount of bug finding capacity available in the world. Or at least there was. That's what's changing as these models become competent enough to spot and validate bugs. That finite global capacity to find bugs is now increasing, and actual bugs are starting to be dredged up. This year will be very very interesting if models continue to increase in capability. | | |
| ▲ | literalAardvark an hour ago | parent [-] | | I was just thinking about this and what it means for closed source code. Many people with skin in the game will be spending tokens on hardening OSS bits they use, maybe even part of their build pipelines, but if the code is closed you have to pay for that review yourself, making you rather uncompetitive. You could say there's no change there, but the number of people who can run a Claude review and the number of people who can actually review a complicated codebase are several orders of magnitude apart. Will some of them produce bad PRs? Probably. The battle will be to figure out how to filter them at scale. |
| |
| ▲ | NitpickLawyer 12 hours ago | parent | prev [-] | | > This is something a lot of static analysers can easily find. And yet they didn't (either noone ran them, or they didn't find it, or they did find it but it was buried in hundreds of false positives) for 20+ years... I find it funny that every time someone does something cool with LLMs, there's a bunch of takes like this: it was trivial, it's just not important, my dad could have done that in his sleep. | | |
| ▲ | userbinator 12 hours ago | parent | next [-] | | Remember Heartbleed in OpenSSL? That long predated LLMs, but same story: some bozo forgot how long something should/could be, and no one else bothered to check either. | | | |
| ▲ | choeger 6 hours ago | parent | prev | next [-] | | It's much, much, easier to run an LLM than to use a static or dynamic analyzer correctly. At the very least, the UI has improved massively with "AI". | |
| ▲ | mrshadowgoose 6 hours ago | parent | prev | next [-] | | And even if that's true (and it frequently is!), detractors usually miss the underlying and immense impact of "sleeping dad capability" equivalent artificial systems. Horizontally scaling "sleeping dads" takes decades, but inference capacity for a sleeping dad equivalent model can be scaled instantly, assuming one has the hardware capacity for it. The world isn't really ready for a contraction of skill dissemination going from decades to minutes. | |
| ▲ | pjmlp 8 hours ago | parent | prev [-] | | Most likely no-one runned them, given the developer culture. |
|
|
|
| ▲ | altern8 13 minutes ago | parent | prev | next [-] |
| Every time I read these titles, I wonder if people are for some reason pushing the narrative that Claude is way smarter than it really is, or if I'm using it wrong. They want me to code AI-first, and the amount of hallucinations and weird bugs and inconsistencies that Claude produces is massive. Lots of code that it pushes would NOT have passed a human/human code review 6 months ago. |
|
| ▲ | fguerraz 3 hours ago | parent | prev | next [-] |
| Interestingly, I think 3 or 4 out of the 5 bugs would have been prevented / mitigated quite well using https://github.com/anthraxx/linux-hardened patches... (disabled io_uring, would have crashed the kernel on UAF, and made exploitation of the heap overflow very unreliable) |
|
| ▲ | DGAP 8 hours ago | parent | prev | next [-] |
| I replicated this experiment on several production codebases and got several crits. Lots of dupes, lots of false positives, lots of bugs that weren't actually exploitable, lots of accepted/ known risks. But also, crits! |
| |
| ▲ | sbuttgereit 13 minutes ago | parent [-] | | I think this really needs to be party of the message. It's great that Claude found a vulnerability that apparently has been overlooked for a long time. It's even proper for Anthropic to tout the find. But we should all ask about the signal to nose ratio that would have been part of the process. If it only was successful... That would be worth touting, too. But I expect there was more noise than they'd care to admit. Or put another way, the context matters. |
|
|
| ▲ | PeterStuer 5 hours ago | parent | prev | next [-] |
| Those 3 letter agencies are going to see their stash of 0-days dwindle so hard. |
|
| ▲ | summarity 11 hours ago | parent | prev | next [-] |
| Related work from our security lab: Stream of vulnerabilities discovered using security agents (23 so far this year): https://securitylab.github.com/ai-agents/ Taskflow harness to run (on your own terms): https://github.blog/security/how-to-scan-for-vulnerabilities... |
|
| ▲ | e12e 4 hours ago | parent | prev | next [-] |
| I wonder about the "video running in the background" during qna of the talk: https://youtu.be/1sd26pWhfmg?is=XLJX9gg0Zm1BKl_5 Did he write an exploit for the NFS bug that runs via network over USB? Seems to be plugging in a SoC over USB...? |
|
| ▲ | misiek08 9 hours ago | parent | prev | next [-] |
| Do not expect so many more reports. Expect so many more attacks ;) |
|
| ▲ | dist-epoch 12 hours ago | parent | prev | next [-] |
| > "given enough eyeballs, all bugs are shallow" Time to update that: "given 1 million tokens context window, all bugs are shallow" |
| |
| ▲ | summarity 11 hours ago | parent | next [-] | | Already happend: https://arxiv.org/abs/2407.08708 | |
| ▲ | riffraff 12 hours ago | parent | prev | next [-] | | ..and three months to review the false positives | | |
| ▲ | 112233 11 hours ago | parent [-] | | this is always overlooked. AI stories sound like "with right attitude, you too can win 10M $ in lottery, like this man just did" Running LLM on 1000 functions produces 10000 reports (these numbers are accurate because I just generated them) — of course only the lottery winners who pulled the actually correct report from the bag will write an article in Evening Post | | |
| ▲ | red75prime 10 hours ago | parent [-] | | > these numbers are accurate because I just generated them Is it sarcasm, or you really did this? Claude Opus 4.6? |
|
| |
| ▲ | bigbugbag 11 hours ago | parent | prev [-] | | more like some bugs are shallow and others are pieced together false-positives from an automated tool reliable in its unreliability. |
|
|
| ▲ | cesaref 10 hours ago | parent | prev | next [-] |
| I'm interested in the implications for the open source movement, specifically about security concerns. Anyone know is there has been a study about how well Claude Code works on closed source (but decompiled) source? |
| |
| ▲ | steveklabnik 2 hours ago | parent | next [-] | | I’ve had Claude Code diagnose bugs in a compiler we wrote together by using gdb and objdump to examine binaries it produces. We don’t have DWARF support yet so it is just examining the binary. That’s not security work, but it’s adjacent to the sorts of skills you’re talking about. The binaries are way smaller than real programs, though. | |
| ▲ | skeledrew 9 hours ago | parent | prev [-] | | > Claude Code works on closed source (but decompiled) source Very likely not nearly as well, unless there are many open source libraries in use and/or the language+patterns used are extremely popular. The really huge win for something like the Linux kernel and other popular OSS is that the source appears in the training data, a lot. And many versions. So providing the source again and saying "find X" is primarily bringing into focus things it's already seen during training, with little novelty beyond the updates that happened after knowledge cutoff. Giving it a closed source project containing a lot of novel code means it only has the language and it's "intuition" to work from, which is a far greater ask. | | |
| ▲ | kasey_junk 9 hours ago | parent [-] | | I’m not a security researcher, but I know a few and I think universally they’d disagree with this take. The llms know about every previous disclosed security vulnerability class and can use that to pattern match. And they can do it against compiled and in some cases obfuscated code as easily as source. I think the security engineers out there are terrified that the balance of power has shifted too far to the finding of closed source vulnerabilities because getting patches deployed will still take so long. Not that the llms are in some way hampered by novel code bases. | | |
| ▲ | skeledrew 8 hours ago | parent [-] | | Many vulnerabilities aren't just pattern matching though; deep understanding of the context in the particular codebase is also needed. And a novel codebase means more attention than usual will be spent grepping and keeping the context in focus. Which will make it easier to miss certain things, than if enough of the context was already encoded in the model weights. Same thing applies to humans: the better someone knows a codebase, the better they will be at resolving issues, etc. | | |
| ▲ | tptacek 5 hours ago | parent [-] | | Almost all vulnerabilities are either direct applications of known patterns, incremental extensions of them, or chains of multiple such steps. |
|
|
|
|
|
| ▲ | jazz9k a day ago | parent | prev | next [-] |
| This does sound great, but the cost of tokens will prevent most companies from using agents to secure their code. |
| |
| ▲ | KetoManx64 21 hours ago | parent | next [-] | | Tokens are insanely cheap at the moment.
Through OpenRouter a message to Sonnet costs about $0.001 cents or using Devstral 2512 it's about $0.0001.
An extended coding session/feature expansion will cost me about $5 in credits.
Split up your codebase so you don't have to feed all of it into the LLM at once and it's a very reasonable. | | |
| ▲ | lebovic 19 hours ago | parent | next [-] | | It cost me ~$750 to find a tricky privilege escalation bug in a complex codebase where I knew the rough specs but didn't have the exploit. There are certainly still many other bugs like that in the codebase, and it would cost $100k-$1MM to explore the rest of the system that deeply with models at or above the capability of Opus 4.6. It's definitely possible to do a basic pass for much less (I do this with autopen.dev), but it is still very expensive to exhaustively find the harder vulnerabilities. | | |
| ▲ | otterley 8 hours ago | parent | next [-] | | How much would it have cost a human to do the same work? The question isn’t how much tokens cost; the question is how much money is saved by using AI to do it. | |
| ▲ | christophilus 10 hours ago | parent | prev | next [-] | | This is where the Codex and Claude Code Pro/Max plans are excellent. I rarely run into the limits of Codex. If I do, I wait and come back and have it resume once the window has expired. | | |
| ▲ | Jcampuzano2 9 hours ago | parent [-] | | Claude and Codex pro/max subs aren't supposed to be used for commercial/enterprise development so its not really an option for execs in enterprise. They need to take into account API costs. At my F500 company execs are very wary of the costs of most of these tools and its always top of mind. We have dashboards and gather tons of internal metrics on which tools devs are using and how much they are costing. | | |
| ▲ | christophilus 5 hours ago | parent | next [-] | | No, I think that’s wrong. They aren’t supposed to be put behind a service, but they can certainly be used to write professional products/ products for the enterprise. | |
| ▲ | otterley 8 hours ago | parent | prev | next [-] | | Are they also measuring productivity? Measuring only token costs is like looking only at grocery spend but not the full receipt: you don’t know whether you fed your family for a week or for only a day. | | |
| ▲ | Jcampuzano2 3 hours ago | parent | next [-] | | I'm not one of those execs, I'm just echoing what they tell us from those I've talked to who manage these dashboards and worry about this. I do think measuring productivity is not very clear-cut especially with these tools. They do "attempt" to measure productivity. But they also just see large dollar amounts on AI costs and get wary. My company is also wary of going all in with any one tool or company due to how quickly stuff changes. So far they've been trying to pool our costs across all tools together and give us an "honor system" limit we should try not to go above per month until we do commit to one suite of tools. | |
| ▲ | batshit_beaver 6 hours ago | parent | prev [-] | | First you have to figure out HOW to measure productivity. | | |
| ▲ | otterley 3 hours ago | parent [-] | | (Output / input), both of which are usually measured in money. If you can measure both of those things--and you have bigger problems if your finance department can't--it logically follows that you can measure productivity. | | |
| ▲ | Jcampuzano2 3 hours ago | parent [-] | | Measuring strictly in terms of money per unit time over a small enough timeframe is difficult because not all tasks directly result in immediately observed results. There are tasks worked on at large enterprises that have 5+ year horizons, and those can't all immediately be tracked in terms of monetary gain that can be correlated with AI usage. We've barely even had AI as a daily tool used for development for a few years. |
|
|
| |
| ▲ | petesergeant 8 hours ago | parent | prev [-] | | > Claude and Codex pro/max subs aren't supposed to be used for commercial/enterprise development lolwut? | | |
|
| |
| ▲ | skeledrew 8 hours ago | parent | prev [-] | | Compare to the cost when said vulnerabilities are exploited by bad actors in critical systems. Worth it yet? |
| |
| ▲ | zozbot234 7 hours ago | parent | prev | next [-] | | Agentic tasks use up a huge amount of tokens compared to simple chatting. Every elementary interaction the model has with the outside world (even while doing something as simple as reading code from a large codebase) is a separate "chat" message and "response", and these add up very quickly. | |
| ▲ | gmerc 18 hours ago | parent | prev | next [-] | | You’d have to ignore the massive investor ROI expectations or somehow have no capability to look past “at the moment”. | | |
| ▲ | NitpickLawyer 12 hours ago | parent | next [-] | | That might be a problem for the labs (although I don't think it is) but it's not a problem for end-users. There is enough pressure from top labs competing with each other, and even more pressure from open models that should keep prices at a reasonable price point going further. In order to justify higher prices the SotA needs to have way higher capabilities than the competition (hence justifying the price) and at the same time the competition needs to be way below a certain threshold. Once that threshold becomes "good enough for task x", the higher price doesn't make sense anymore. While there is some provider retention today, it will be harder to have once everyone offers kinda sorta the same capabilities. Changing an API provider might even be transparent for most users and they wouldn't care. If you want to have an idea about token prices today you can check the median for serving open models on openrouter or similar platforms. You'll get a "napkin math" estimate for what it costs to serve a model of a certain size today. As long as models don't go oom higher than today's largest models, API pricing seems in line with a modest profit (so it shouldn't be subsidised, and it should drop with tech progress). Another benefit for open models is that once they're released, that capability remains there. The models can't get "worse". | |
| ▲ | KetoManx64 18 hours ago | parent | prev [-] | | Not really. I'm fully taking advantage of these low prices while they last. Eventually the AI companies will run start running out of funny money and start charging what the models actually cost to run, then I just switch over to using the self hosted models more often and utilize the online ones for the projects that need the extra resources.
Currently there's no reason for why I shouldn't use Claude Sonnet to write one time bash scripts, once it starts costing me a dollar to do so I'm going to change my behavior. | | |
| ▲ | deaux 15 hours ago | parent | next [-] | | > Currently there's no reason for why I shouldn't use Claude Sonnet to write one time bash scripts, once it starts costing me a dollar to do so I'm going to change my behavior. This just isn't going to happen, we have open weights models which we can roughly calculate how much they cost to run that are on the level of Sonnet _right now_. The best open weights models used to be 2 generations behind, then they were 1 generation behind, now they're on par with the mid-tier frontier models. You can choose among many different Kimi K2.5 providers. If you believe that every single one of those is running at 50% subsidies, be my guest. | |
| ▲ | skeledrew 8 hours ago | parent | prev | next [-] | | > start charging what the models actually cost to run The political climate won't allow that to happen. The US will do everything to stay ahead of China, and a rise in prices means a sizeable migration to Chinese models, giving them that much more data to improve their models and pass the US in AI capability (if they haven't already). But also it'll happen in a way, as eventually models will become optimized enough that run cost become more or less negligible from a sustainability perspective. | |
| ▲ | twosdai 16 hours ago | parent | prev [-] | | I also have this feeling. But do you ever doubt it. that when the time comes we will be like the boiled frog? Where its "just so convenient" or that the reality of setting up a local ai is just a worse experience for a large upfront cost? | | |
| ▲ | iririririr 15 hours ago | parent [-] | | worse. he's already boiled. probably paying way more than that one dollar per bash script with all the subscriptions he already has. | | |
| ▲ | KetoManx64 15 hours ago | parent [-] | | Yeah, the $20 I paid to OpenRouter about 4 months ago really cost me an arm and a leg, not sure where I'll get my next meal if I'm to be honest. |
|
|
|
| |
| ▲ | ThePowerOfFuet 13 hours ago | parent | prev [-] | | >$0.001 cents $0.001 (1/10 of a cent) or 0.001 cents (1/1000 of a cent, or $0.00001)? | | |
| |
| ▲ | qingcharles 6 hours ago | parent | prev | next [-] | | I'm thinking about how much money Anthropic etc are making from intelligence services who are running Opus 4.6 on ultra high settings 24 hours a day to find these kinds of exploits and take advantage of them before others do. Expensive for me and you, but peanuts for a nation state. | |
| ▲ | NitpickLawyer 12 hours ago | parent | prev | next [-] | | Tokens aren't more expensive than highly trained meatbags today. There's no way they'll be more expensive "tomorrow"... | | |
| ▲ | bigbugbag 11 hours ago | parent [-] | | they are and they will be, then they won't after the market crashes, the bubble bursts and the companies bankrupts. possibly taking down major portion of the global economy with them. | | |
| ▲ | skeledrew 8 hours ago | parent | next [-] | | > they are and they will be Calculate the approximate cost of raising a human from birth to having the knowledge and skills to do X, along with maintenance required to continue doing X. Multiply by a reasonable scaling factor in comparison to one of today's best LLMs (ie how many humans and how much time to do Xn, vs the LLM). Calculate the cost of hardware (from raw elements), training and maintenance for said LLM (if you want to include the cost of research+software then you'll have to also include the costs of raising those who taught, mentored, etc the human as well). Consider that the human usually specializes, while the LLM touches everything. I think you'll find even a roughly approximate answer very enlightening if you're honest in your calculations. | | |
| ▲ | Synthetic7346 6 hours ago | parent [-] | | But companies don't have to bear the cost of raising a human from birth, or training them. They only pay the cost of hiring them, and that includes cost of maintenence. Add to that the fact that we can't blindly trust LLM output just yet, so we need a mearbag to review it. LLM will always be more expensive than human +LLM, until we're at a stage where we can remove the human from the loop |
| |
| ▲ | PunchyHamster 2 hours ago | parent | prev [-] | | the crash would mean price of GPUs would go down, not up... |
|
| |
| ▲ | epolanski 13 hours ago | parent | prev [-] | | I don't buy it. Inference cost has dropped 300x in 3 years, no reason to think this won't keep happening with improvements on models, agent architecture and hardware. Also, too many people are fixated with American models when Chinese ones deliver similar quality often at fraction of a cost. From my tests, "personality" of an LLM, it's tendency to stick to prompts and not derail far outweights the low % digit of delta in benchmark performance. Not to mention, different LLMs perform better at different tasks, and they are all particularly sensible to prompts and instructions. |
|
|
| ▲ | rixrax 8 hours ago | parent | prev | next [-] |
| I hope next up is the performance and bloat that the LLMs can try and improve. Especially on perf side I would wager LLMs can go from meat sacks what ever works to how do I solve this with best available algorithm and architecture (that also follows some best practises). |
|
| ▲ | eichin a day ago | parent | prev | next [-] |
| An explanation of the Claude Opus 4.6 linux kernel security findings as presented by Nicholas Carlini at unpromptedcon. |
| |
|
| ▲ | skeeter2020 7 hours ago | parent | prev | next [-] |
| And with AI generating vulnerabilities at an accelerated pace this business is only getting bigger. Welcome to the new antivirus! |
| |
| ▲ | bitexploder 7 hours ago | parent [-] | | There will always be more bugs than we can fix. AI can patch as well, but if your system is difficult to test and doesn't have rigorous validation you will likely get an unacceptable amount of regression. |
|
|
| ▲ | alsanan2 9 hours ago | parent | prev | next [-] |
| making public that AI is able of founding that kind of vulnerabilities is a big problem. In this case it's nice that the vulnerability has been closed before publishing but in case a cracker founds it, the result would be extremately different. This kind of news only open eyes for the crackers. |
|
| ▲ | up2isomorphism 19 hours ago | parent | prev | next [-] |
| But on the other hand, Claude might introduce more vulnerability than it discovered. |
| |
| ▲ | yunnpp 19 hours ago | parent | next [-] | | Code review is the real deal for these models. This area seems largely underappreciated to me. Especially for things like C++, where static analysis tools have traditionally generated too many false positives to be useful, the LLMs seem especially good. I'm no black hat but have found similarly old bugs at my own place. Even if shit is hallucinated half the time, it still pays off when it finds that really nasty bug. Instead, people seem to be infatuated with vibe coding technical debt at scale. | | |
| ▲ | qsera 11 hours ago | parent [-] | | > Code review is the real deal for these models. Yea, that is what I have been saying as well... >Instead, people seem to be infatuated with vibe coding technical debt at scale. Don't blame them. That is what AI marketing pushes. And people are sheep to marketing.. I understand why AI companies don't want to promote it. Because they understand that the LCD/Majority of their client base won't see code review as a critical part of their business. If LLMs are marketed as best suited for code review, then they probably cannot justify the investments that they are getting... |
| |
| ▲ | khalic 10 hours ago | parent | prev [-] | | Guys please read the article before commenting... |
|
|
| ▲ | jason1cho 12 hours ago | parent | prev | next [-] |
| This isn't surprising. What is not mentioned is that Claude Code also found one thousand false positive bugs, which developers spent three months to rule out. |
| |
| ▲ | antirez 7 hours ago | parent | next [-] | | That's not what is happening right now. The bugs are often filtered later by LLMs themselves: if the second pipeline can't reproduce the crash / violation / exploit in any way, often the false positives are evicted before ever reaching the human scrutiny. Checking if a real vulnerability can be triggered is a trivial task compared to finding one, so this second pipeline has an almost 100% success rate from the POV: if it passes the second pipeline, it is almost certainly a real bug, and very few real bugs will not pass this second pipeline. It does not matter how much LLMs advance, people ideologically against them will always deny they have an enormous amount of usefulness. This is expected in the normal population, but too see a lot of people that can't see with their eyes in Hacker News feels weird. | | |
| ▲ | uhx 5 hours ago | parent | next [-] | | > Checking if a real vulnerability can be triggered is a trivial task compared to finding one Have you ever tried to write PoC for any CVE? This statement is wrong. Sometimes bug may exist but be impossible to trigger/exploit. So it is not trivial at all. | | |
| ▲ | avemg 4 hours ago | parent | next [-] | | I'm tickled at the idea of asking antirez [1] if he's ever written a PoC for a CVE. [1] https://en.wikipedia.org/wiki/Salvatore_Sanfilippo | | |
| ▲ | jedberg 2 hours ago | parent | next [-] | | I actually like when that happens. Like when people "correct" me about how reddit works. I appreciate that we still focus on the content and not who is saying it. | | |
| ▲ | tptacek 2 hours ago | parent [-] | | That's not really what happened on this thread. Someone said something sensible and banal about vulnerability research, then someone else said do-you-even-lift-bro, and got shown up. | | |
| ▲ | jedberg an hour ago | parent [-] | | That's true in this particular case, but I was talking more about the general case. |
|
| |
| ▲ | tptacek 3 hours ago | parent | prev | next [-] | | This happens over and over in these discussions. It doesn't matter who you're citing or who's talking. People are terrified and are reacting to news reflexively. | | |
| ▲ | emp17344 29 minutes ago | parent [-] | | Personally, I’m tired of exaggerated claims and hype peddlers. Edit: Frankly, accusing perceived opponents of being too afraid to see the truth is poor argumentative practice, and practically never true. |
| |
| ▲ | LeFantome 3 hours ago | parent | prev [-] | | Sure he wrote a port scanner that obscures the IP address of the scanner, but does he know anything about security? /s Oh, and he wrote Redis. No biggie. | | |
| |
| ▲ | antirez 5 hours ago | parent | prev | next [-] | | Firstly I have a long past in computer security, so: yes, I used to write exploits. Second, the vulnerability verification does not need being able to exploit, but triggering an ASAN assert. With memory corruption that's very simple often times and enough to verify the bug is real. | |
| ▲ | freedomben 5 hours ago | parent | prev | next [-] | | I'm not GP, but I've written multiple PoCs for vulns. I agree with GP. Finding a vuln is often very hard. Yes sometimes exploiting it is hard (and requires chaining), but knowing where the vuln is (most of the time) the hard part. | |
| ▲ | e12e 4 hours ago | parent | prev | next [-] | | Note the exploit Claude wrote for the blind SQL injection found in ghost - in the same talk. https://youtu.be/1sd26pWhfmg?is=XLJX9gg0Zm1BKl_5 | |
| ▲ | orochimaaru 3 hours ago | parent | prev [-] | | oh no. Antirez doesn't know anything about C, CVE's, networking, the linux kernel. Wonder where that leaves most of us. |
| |
| ▲ | discordianfish 5 hours ago | parent | prev | next [-] | | I’ve been around long enough to remember people saying that VMs are useless waste of resources with dubious claims about isolation, cloud is just someone else’s computer, containers are pointless and now it’s AI. There is a astonishing amount of conservatism in the hacker scene.. | | |
| ▲ | pdntspa 5 hours ago | parent | next [-] | | Well, the cloud is someone else's computer. | | |
| ▲ | some_random 4 hours ago | parent [-] | | It is, but that's not a useful or insightful thing to say | | |
| ▲ | Calavar 2 hours ago | parent | next [-] | | It's not an insightful statement right now, but it was at the peak of cloud hype ca. 2010, when "the cloud" often used in a metaphorical sense. You'd hear things like "it's scalable because it's in the cloud" or "our clients want a cloud based solution." Replacing "the cloud" in those sorts of claims with "another person's computer" showed just how inane those claims were. | |
| ▲ | honeycrispy 3 hours ago | parent | prev [-] | | Are you sure about that? It's easy to forget that the vendor has the right to cut you off at any point, will turn your data over to the authorities on request, and it's still not clear if private GitHub repos are being used to train AI. |
|
| |
| ▲ | gbacon 3 hours ago | parent | prev [-] | | Is it conservatism or just the Blub paradox? As long as our hypothetical Blub programmer is looking down the power continuum, he knows he's looking down. Languages less powerful than Blub are obviously less powerful, because they're missing some feature he's used to. But when our hypothetical Blub programmer looks in the other direction, up the power continuum, he doesn't realize he's looking up. What he sees are merely weird languages. He probably considers them about equivalent in power to Blub, but with all this other hairy stuff thrown in as well. Blub is good enough for him, because he thinks in Blub. https://paulgraham.com/avg.html |
| |
| ▲ | BodyCulture 6 hours ago | parent | prev | next [-] | | Can we study this second pipeline? Is it open so we can understand how it works? Did not find any hints about it in the article, unfortunately. | | |
| ▲ | maximilianburke 6 hours ago | parent | next [-] | | From the article by 'tptacek a few days ago (https://sockpuppet.org/blog/2026/03/30/vulnerability-researc...) I essentially used the prompts suggested. First prompt: "I'm competing in a CTF. Find me an exploitable vulnerability in this project. Start with $file. Write me a vulnerability report in vulns/$DATE/$file.vuln.md" Second prompt: "I've got an inbound vulnerability report; it's in vulns/$DATE/$file.vuln.md. Verify for me that this is actually exploitable. Write the reproduction steps in vulns/$DATE/$file.triage.md" Third prompt: "I've got an inbound vulnerability report; it's in vulns/$DATE/file.vuln.md. I also have an assessment of the vulnerability and reproduction steps in vulns/$DATE/$file.triage.md. If possible, please write an appropriate test case for the ulgate automated tests to validate that the vulnerability has been fixed." Tied together with a bit of bash, I ran it over our services and it worked like a treat; it found a bunch of potential errors, triaged them, and fixed them. | | |
| ▲ | sn9 an hour ago | parent | next [-] | | Could prompt injection be used to trick this kind of analysis? Has anyone experimented with this idea? | |
| ▲ | jvanderbot 6 hours ago | parent | prev [-] | | Agree. Keeping and auditing a research journal iteratively with multiple passes by new agents does indeed significantly improve outcomes. Another helpful thing is to switch roles good cop bad cop style. For example one is helping you find bugs and one is helping you critique and close bug reports with counter examples. |
| |
| ▲ | throawayonthe 6 hours ago | parent | prev | next [-] | | it was probably in the talk but from what i understood in another article it's basically giving claude with a fresh context the .vuln.md file and saying "i'm getting this vulnerability report, is this real?" edit: i remember which article, it was this one: https://sockpuppet.org/blog/2026/03/30/vulnerability-researc... (an LWN comment in response to this post was on the frontpage recently) | |
| ▲ | 4b11b4 6 hours ago | parent | prev [-] | | One such example is IRIS. In general, any traditional static analysis tool combined with a language model at some stage in a pipeline. |
| |
| ▲ | bch 4 hours ago | parent | prev | next [-] | | > This is expected in the normal population A lot of people regardless of technical ability have strong opinions about what LLMs are/are-not. The number of lay people i know who immediately jump to "skynet" when talking about the current AI world... The number of people i know who quit thinking because "Well, let's just see what AI says"... A (big) part of the conversation re: "AI" has to be "who are the people behind the AI actions, and what is their motivation"? Smart people have stopped taking AI bug reports[0][1] because of overwhelming slop; its real. [0] https://www.theregister.com/2025/05/07/curl_ai_bug_reports/ [1] https://gist.github.com/bagder/07f7581f6e3d78ef37dfbfc81fd1d... | | |
| ▲ | LeFantome 3 hours ago | parent [-] | | The fact that most AI bug reports are low-quality noise says as much or more about the humans submitting them than it does about the state of AI. As others have said, there are multiple stages to bug reports and CVEs. 1. Discover the bug 2. Verify the bug You get the most false positives at step one. Most of these will be eliminated at step 2. 3. Isolate the bug This means creating a test case that eliminates as much of the noise as possible to provide the bare minimum required to trigger the big. This will greatly aid in debugging. Doing step 2 again is implied. 4. Report the bug Most people skip 2 and 3, especially if they did not even do 1 (in the case of AI) But you can have AI provide all 4 to achieve high quality bug reports. In the case of a CVE, you have a step 5. 5 - Exploit the bug But you do not have to do step 5 to get to step 2. And that is the step that eliminates most of the noise. |
| |
| ▲ | antonvs 6 hours ago | parent | prev | next [-] | | > to see a lot of people that can't see with their eyes in Hacker News feels weird. Turns out the average commenter here is not, in fact, a "hacker". | |
| ▲ | slopinthebag 4 hours ago | parent | prev | next [-] | | What if the second round hallucinates that a bug found in the first round is a false positive? Would we ever know? > It does not matter how much LLMs advance, people ideologically against them will always deny they have an enormous amount of usefulness. They have some usefulness, much less than what the AI boosters like yourself claim, but also a lot of drawbacks and harms. Part of seeing with your eyes is not purposefully blinding yourself to one side here. | |
| ▲ | nickphx 5 hours ago | parent | prev | next [-] | | they are useful to those that enjoy wasting time. | |
| ▲ | ksec 6 hours ago | parent | prev [-] | | >This is expected in the normal population, but too see a lot of people that can't see with their eyes in Hacker News feels weird. You are replying to an account created in less than 60 days. | | |
| ▲ | jvanderbot 6 hours ago | parent [-] | | This is a bit unfair. Hackers are born every day. | | |
| ▲ | ksec 2 hours ago | parent | next [-] | | In relation to the quality of its comment. I thought it was a fair. He just completely made up about false positives. And in case people dont know, antirez has been complaining about the quality of HN comments for at least a year, especially after AI topic took over on HN. It is still better than lobster or other place though. | |
| ▲ | slekker 4 hours ago | parent | prev [-] | | Bots too, vanderBOT! | | |
| ▲ | jvanderbot 2 hours ago | parent [-] | | I used to work in robotics, and can't remember the password for my usual username so I pulled this one out of thin air years ago |
|
|
|
| |
| ▲ | mtlynch 11 hours ago | parent | prev | next [-] | | > What is not mentioned is that Claude Code also found one thousand false positive bugs, which developers spent three months to rule out. Source? I haven't seen this anywhere. In my experience, false positive rate on vulnerabilities with Claude Opus 4.6 is well below 20%. | | |
| ▲ | Supermancho 8 hours ago | parent | next [-] | | To the issue of AI submitted patches being more of a burden than a boon, many projects have decided to stop accepting AI-generated solutioning: https://blog.devgenius.io/open-source-projects-are-now-banni... These are just a few examples. There are more that google can supply. | | |
| ▲ | logicprog 6 hours ago | parent | next [-] | | According to Willy Tarreau[0] and Greg Kroah-Hartman[1], this trend has recently significantly reversed, at least form the reports they've been seeing on the Linux kernel. The creator of curl, Daniel Steinberg, before that broader transition, also found the reports generated by LLM-powered but more sophisticated vuln research tools useful[2] and the guy who actually ran those tools found "They have low false positive rates."[3] Additionally, there was no mention in the talk by the guy who found the vuln discussed in the TFA of what the false positive rate was, or that he had to sift through the reports because it was mostly slop — or whether he was doing it out of courtesy. Additionally, he said he found only several hundred, iirc, not "thousands." All he said was: "I have so many bugs in the Linux kernel that I can’t report because I haven’t validated them yet… I’m not going to send [the Linux kernel maintainers] potential slop, but this means I now have several hundred crashes that they haven’t seen because I haven’t had time to check them." (TFA) He quite evidently didn't have to sift through thousands, or spend months, to find this one, either. [0]: https://lwn.net/Articles/1065620/
[1]: https://www.theregister.com/2026/03/26/greg_kroahhartman_ai_...
[2]: https://simonwillison.net/2025/Oct/2/curl/p
[3]: https://joshua.hu/llm-engineer-review-sast-security-ai-tools... | |
| ▲ | literalAardvark 8 hours ago | parent | prev [-] | | No, they haven't. Read the ai slop you posted carefully. It's a policy update that enables maintainers to ignore low effort "contributions" that come from untrusted people in order to reduce reviewing workload. An Eternal September problem, kind of. | | |
| ▲ | coldtea 7 hours ago | parent [-] | | Didn't you just restate what the parent claimed? | | |
| ▲ | cwillu 7 hours ago | parent [-] | | No, that's not at all the same thing: ai-generated contributions from people with a track record for useful contributions are still accepted. | | |
| ▲ | dpark 7 hours ago | parent | next [-] | | Right. AI submissions are so burdensome that they have had to refuse them from all except a small set of known contributors. The fact that there’s a small carve out for a specific set of contributors in no way disputes what Supermancho claimed. | | |
| ▲ | phanimahesh 6 hours ago | parent [-] | | A powertool that needs discretion and good judgement to be used well is being restricted to people with a track record of displaying good judgement. I see nothing wrong here. AI enables volume, which is a problem. But it is also a useful tool. Does it increase review burden? Yes. Is it excessively wasteful energy wise? Yes. Should we avoid it? Probably no. We have to be pragmatic, and learn to use the tools responsibly. | | |
| ▲ | dpark 6 hours ago | parent [-] | | I never said anything is wrong with the policy. Or with the tool use for that matter. This whole chain was one person saying “AI is creating such a burden that projects are having to ban it”, someone else being willfully obtuse and saying “nuh uh, they’re actually still letting a very restricted set of people use it”, and now an increasingly tangential series of comments. | | |
| ▲ | literalAardvark 2 hours ago | parent [-] | | I feel like you're still failing to grasp the point. The only difference is that before AI the number of low effort PRs was limited by the number of people who are both lazy and know enough programming, which is a small set because a person is very unlikely to be both. Now it's limited to people who are lazy and can run ollama with a 5M model, which is a much larger set. It's not an AI code problem by itself. AI can make good enough code. It's a denial of service by the lazy against the reviewers, which is a very very different problem. | | |
| ▲ | dpark an hour ago | parent [-] | | No one is missing your point. The issue is that you are responding a point no one made. The grounding premise of this comment chain was “AI submitted patches being more of a burden than a boon”. You are misinterpreting that as some sort of general statement that “AI Bad” and that AI is being globally banned. A metaphor for the scenario here is someone says “It’s too dangerous to hand repo ownership out to contributors. Projects aren’t doing that anymore.” And someone else comes in to say “That’s not true! There are still repo owners. They are just limiting it to a select group now!” This statement of fact is only an interesting rebut if you misinterpret the first statement to say that no one will own the repo because repo ownership is fundamentally bad. > It's a denial of service by the lazy against the reviewers, which is a very very different problem. And it is AI enabling this behavior. Which was the premise above. |
|
|
|
| |
| ▲ | coldtea 6 hours ago | parent | prev [-] | | Yes, but technically no different than "good contributions from humans are still accepted, AI slop can fuck off". Since the onus falls on those "people with a track record for useful contributions" to verify, design tastefully, test and ensure those contributions are good enough to submit - not on the AI they happen to be using. If it fell on the AI they're using, then any random guy using the same AI would be accepted. |
|
|
|
| |
| ▲ | christophilus 10 hours ago | parent | prev | next [-] | | Same. Codex and Claude Code on the latest models are really good at finding bugs, and really good at fixing them in my experience. Much better than 50% in the latter case and much faster than I am. | |
| ▲ | j16sdiz 10 hours ago | parent | prev | next [-] | | In TFA: I have so many bugs in the Linux kernel that I can’t
report because I haven’t validated them yet… I’m not going
to send [the Linux kernel maintainers] potential slop,
but this means I now have several hundred crashes that they
haven’t seen because I haven’t had time to check them.
—Nicholas Carlini, speaking at [un]prompted 2026
| | |
| ▲ | mtlynch 10 hours ago | parent | next [-] | | Those aren't false positives; they're results he hasn't yet inspected. I wrote a longer reply here: https://news.ycombinator.com/item?id=47638062 | | |
| ▲ | coldtea 7 hours ago | parent | next [-] | | >Those aren't false positives; they're results he hasn't yet inspected. It's not a XOR | | |
| ▲ | Ukv 7 hours ago | parent | next [-] | | The article quote was being given as the supposed source for "Claude Code also found one thousand false positive bugs, which developers spent three months to rule out", so should substantiate that claim - which it doesn't. If the claim was instead just "a good portion of the hundreds more potential bugs it found might be false positives", then sure. | |
| ▲ | tptacek 5 hours ago | parent | prev [-] | | Yes it is. They're not not false positives until they're reported and consume maintainer time. |
| |
| ▲ | bethekidyouwant 8 hours ago | parent | prev [-] | | some of them certainly are… |
| |
| ▲ | sobiolite 7 hours ago | parent | prev [-] | | The comment said "Claude Code also found one thousand false positive bugs, which developers spent three months to rule out.". Please explain how a bug can both be unvalidated, and also have undergone a three month process to determine it is a false positive? |
| |
| ▲ | paulddraper 8 hours ago | parent | prev | next [-] | | Source: """AI is bad""" | |
| ▲ | r9295 11 hours ago | parent | prev [-] | | In my experience, the issue has been likelihood of exploitation or issue severity. Claude gets it wrong almost all the time. A threat model matters and some risks are accepted. Good luck convincing an LLM of that fact |
| |
| ▲ | goalieca 8 hours ago | parent | prev | next [-] | | Static/Dynamic analysis tools find vulnerabilities all the time. Almost all projects of a certain size have a large backlog of known issues from these boring scanners. The issue is sorting through them all and triaging them. There's too many issues to fix and figuring out which are exploitable and actually damaging, given mitigations, is time consuming. Am i impressed claude found an old bug? Sort of.. everytime a new scanner is introduced you get new findings that others haven't found. | | |
| ▲ | tptacek 5 hours ago | parent [-] | | Static analyzers find large numbers of hypothetical bugs, of which only a small subset are actionable, and the work to resolve which are actionable and which are e.g. "a memcpy into an 8 byte buffer whose input was previously clamped to 8 bytes or less" is so high that analyzers have little impact at scale. I don't know off the top of my head many vulnerability researchers who take pure static analysis tools seriously. Fuzzers find different bugs and fuzzers in particular find bugs without context, which is why large-scale fuzzer farms generate stacks of crashers that stay crashers for months or years, because nobody takes the time to sift through the "benign" crashes to find the weaponizable ones. LLM agents function differently than either method. They recursively generate hypotheticals interprocedurally across the codebase based on generalizations of patterns. That by itself would be an interesting new form of static analysis (and likely little more effective than SOTA static analysis). But agents can then take confirmatory steps on those surfaced hypos, generate confidence, and then place those findings in context (for instance, generating input paths through the code that reach the bug, and spelling out what attack primitives the bug conditions generates). If you wanted to be reductive you'd say LLM agent vulnerability discovery is a superset of both fuzzing and static analysis. And, importantly, that's before you get to the fact that LLM agents can fuzz and do modeling and static analysis themselves. | | |
| ▲ | goalieca 2 hours ago | parent [-] | | There are plenty of static analyzers do attempt to
walk code paths for reachability. Some even track tainted input. And yes, these are often good starting points for developing exploits. I’ve done this myself. I’m curious about LLM agents, but the fact they don’t “understand” is why I’m very skeptical of the hype. I find myself wasting just as much if not more time with them than with a terrible “enterprise” sast tool. |
|
| |
| ▲ | linsomniac 8 hours ago | parent | prev | next [-] | | The article doesn't say they found a bunch of false positives. It says they have a huge backlog that they still need to test: "I have so many bugs in the Linux kernel that I can’t report because I haven’t validated them yet…" | |
| ▲ | boplicity 9 hours ago | parent | prev | next [-] | | The lesson here shouldn't be that Claude Code is useless, but that it's a powerful tool in the hands of the right people. | | |
| ▲ | mavamaarten 9 hours ago | parent | next [-] | | I'm growing allergic to the hype train and the slop. I've watched real-life talks about people that sent some prompt to Claude Code and then proudly present something mediocre that they didn't make themselves to a whole audience as if they'd invented the warm water, and that just makes me weary. But at the same time, it has transformed my work from writing everything bit of code myself, to me writing the cool and complex things while giving directions to a helper to sort out the boring grunt work, and it's amazingly capable at that. It _is_ a hugely powerful tool. But haters only see red, and lovers see everything through pink glasses. | | |
| ▲ | iterateoften 8 hours ago | parent | next [-] | | Sounds like maybe you might have some mixed feelings about becoming more effective with ai, but then at the same time everyone else is too so the praise youre expecting is diluted. I see it all the time now too. People have no frame of reference at all about what is hard or easy so engineers feel under-appreciated because the guy who never coded is getting lots of praise for doing something basic while experienced people are able to spit out incredibly complex things. But to an outsider, both look like they took the same work. | |
| ▲ | ofrzeta 3 hours ago | parent | prev | next [-] | | I am also torn because obviously the LLMs have a lot of value but the amount of misuse is overwhelming. People just keep pasting slop into story descriptions that no one can keep up. There should be guidelines at work places to use AI responsibly. | |
| ▲ | sph 9 hours ago | parent | prev [-] | | > it has transformed my work […] to me writing the cool and complex things > it's amazingly capable at that. > It _is_ a hugely powerful tool Damn, that’s what you call being allergic to the hype train? This type of hypocritical thinly-veiled praise is what is actually unbearable with AI discourse. | | |
| ▲ | asyx 8 hours ago | parent [-] | | I don’t think it is controversial that AI tools are good enough at crud endpoints that it is totally viable to just let it run through the grunt work of hooking up endpoints to a service and then you can focus on the interesting aspect of the application which is exactly that service. |
|
| |
| ▲ | amelius 9 hours ago | parent | prev | next [-] | | Unfortunately, also in the hands of the __wrong__ people. Maybe even more so, because who is going to wade through all those false positives? A bad actor is maybe more likely to do that. | | |
| ▲ | embedding-shape 9 hours ago | parent [-] | | > A bad actor is maybe more likely to do that. Do something about that then, so white-hat hackers are more likely than black-hat hackers to wanting to wade through that, incentives and all that jazz. | | |
| ▲ | amelius 4 hours ago | parent | next [-] | | Good luck with that. Security is at the bottom of everyone's budget allocation list. | |
| ▲ | ruszki 4 hours ago | parent | prev [-] | | We couldn’t solve the incentive against misinformation/disinformation since inception, we made it even worse than 20 years ago. Even when we know how it works exactly, even on the internet, not just generally. These kinds of statements seem quite unrealistic to me. |
|
| |
| ▲ | righthand 9 hours ago | parent | prev | next [-] | | The lesson or the hype mantra? | |
| ▲ | teeray 9 hours ago | parent | prev [-] | | The same could be said about a Roulette wheel set before a seasoned gambler | | |
| ▲ | TheCoreh 8 hours ago | parent | next [-] | | Can a Roulette wheel set find vulnerabilities in software? | | |
| ▲ | edoceo 7 hours ago | parent [-] | | If vulnerability=compulsion and software=meat bags then yes. |
| |
| ▲ | throw-the-towel 7 hours ago | parent | prev | next [-] | | This is a non-sequitur if I ever saw one. | |
| ▲ | vntok 7 hours ago | parent | prev [-] | | No. The seasoned gambler can not learn things that measurably increase their chance at the Roulette, whereas they definitely can do that with an LLM. And the LLM itself becomes smarter over time through hardware upgrades, software updates and even memory for those who enable that feature. |
|
| |
| ▲ | bri3d 6 hours ago | parent | prev | next [-] | | This is not how first party vulnerability research with LLMs go; they are incredibly valuable versus all prior tooling at triage and producing only high quality bugs, because they can be instructed to produce a PoC and prove that the bug is reachable. It’s traditional research methods (fuzzing, static analysis, etc.) that are more prone to false positive overload. The reason why open submission fields (PRs, bug bounty, etc) are having issues with AI slop spam is that LLMs are also good at spamming, not that they are bad at programming or especially vulnerability research. If the incentives are aligned LLMs are incredibly good at vulnerability research. | |
| ▲ | dekhn 5 hours ago | parent | prev | next [-] | | Everything changed in the past 6 months and coding LLMs went from being OK-ish to insanely good. People also got better at using them. Also, high false positive rate isn't that bad in the case where a false negative costs a lot (an exploit in the linux kernel is a very expensive mistake). And, in going through the false positives and eliminating them, those results will ideally get folded back into the training set for the next generation of LLMs, likely reducing the future rate of false positives. | | |
| ▲ | catlifeonmars 5 hours ago | parent [-] | | > Everything changed in the past 6 months and coding LLMs went from being OK-ish to insanely good. People also got better at using them. I hear this literally every 6 months :) | | |
| ▲ | tptacek 5 hours ago | parent [-] | | It hasn't been true forever, but it has been true over the last 18 months or so. |
|
| |
| ▲ | sva_ 11 hours ago | parent | prev | next [-] | | Couldn't you just make it write a PoC? | | |
| ▲ | tptacek 5 hours ago | parent | next [-] | | Yes, you can. I strongly encourage people skeptical about this, and who know at a high-level how this kind of exploitation works, to just try it. Have Claude or Codex (they have different strengths at this kind of work) set up a testing harness with Firecracker or QEMU, and then work through having it build an exploit. | |
| ▲ | weird-eye-issue 10 hours ago | parent | prev [-] | | Still have to validate it. | | |
| |
| ▲ | logicprog 6 hours ago | parent | prev | next [-] | | Okay, so anti AI people are just making shit up now. Got it. According to Willy Tarreau[0] and Greg Kroah-Hartman[1], this trend has recently significantly reversed, at least form the reports they've been seeing on the Linux kernel. The creator of curl, Daniel Steinberg, before that broader transition, also found the reports generated by LLM-powered but more sophisticated vuln research tools useful[2] and the guy who actually ran those tools found "They have low false positive rates."[3] Additionally, there was no mention in the talk by the guy who found the vuln discussed in the TFA of what the false positive rate was, or that he had to sift through the reports because it was mostly slop — or whether he was doing it out of courtesy. Additionally, he said he found only several hundred, iirc, not "thousands." All he said was: "I have so many bugs in the Linux kernel that I can’t report because I haven’t validated them yet… I’m not going to send [the Linux kernel maintainers] potential slop, but this means I now have several hundred crashes that they haven’t seen because I haven’t had time to check them." (TFA) He quite evidently didn't have to sift through thousands, or spend months, to find this one, either. [0]: https://lwn.net/Articles/1065620/
[1]: https://www.theregister.com/2026/03/26/greg_kroahhartman_ai_...
[2]: https://simonwillison.net/2025/Oct/2/curl/p
[3]: https://joshua.hu/llm-engineer-review-sast-security-ai-tools... | |
| ▲ | Trufa 4 hours ago | parent | prev | next [-] | | What is with negativity against AI in YC? Can anyone point a finger of why this anti take is so prominent? We're living through the most revolutionary moment of software since it's its inception and the main thing that gets consistently upvoted is negativity, FUD and it doesn't work in this case, or it's all slop. | | |
| ▲ | bwfan123 3 hours ago | parent | next [-] | | > Can anyone point a finger of why this anti take is so prominent? AI tools are great but are being oversold and overhyped by those with an incentive. So, there is a continuous drumbeat of "AI will do all the code for you" ! "Look at this browser written by AI", "C compiler in rust written entirely by AI" etc. And then, that drumbeat is amplified by those in management who have not built software systems themselves. What happened to the AI generated "C compiler in rust" ? or the browser written by AI ? - they remain a steaming pile of almost-working code. AI is great at producing "almost-working" poc code which is good for bootstrapping work and getting you 90% of the way if you are ok with code of questionable lineage. But many applications need "actually-working" code that requires the last 10%. So, some in this forum who have been in the trenches building large "actually working" software systems and also use AI tools daily and know their limitations are injecting some realism into the debate. | |
| ▲ | arealaccount 3 hours ago | parent | prev [-] | | Not speaking for myself but the you won’t have a job soon narrative puts people off |
| |
| ▲ | addandsubtract 11 hours ago | parent | prev | next [-] | | On the other hand, some bugs take three months to find. So this still seems like a win. | |
| ▲ | sixhobbits 6 hours ago | parent | prev [-] | | From a recent front page article that mentioned the previous slop problem: > Now most of these reports are correct, to the point that we had to bring in more maintainers to help us. https://news.ycombinator.com/item?id=47611921 |
|
|
| ▲ | cookiengineer 12 hours ago | parent | prev | next [-] |
| > Nicholas has found hundreds more potential bugs in the Linux kernel, but the bottleneck to fixing them is the manual step of humans sorting through all of Claude’s findings No, the problem is sorting out thousands of false positives from claude code's reports. 5 out of 1000+ reports to be valid is statistically worse than running a fuzzer on the codebase. Just sayin' |
| |
| ▲ | mtlynch 11 hours ago | parent | next [-] | | > 5 out of 1000+ reports to be valid is statistically worse than running a fuzzer on the codebase. Carlini said "hundreds" of crashes, not 1000+. It's not that only 5 were true positives and the rest were false positives. 5 were true positives and Carlini doesn't have bandwidth to review the rest. Presumably he's reviewed more than 5 and some were not worth reporting, but we don't know what that number is. It's almost certainly not hundreds. Keep in mind that Carlini's not a dedicated security engineer for Linux. He's seeing what's possible with LLMs and his team is simultaneously exploring the Linux kernel, Firefox,[0] GhostScript, OpenSC,[1] and probably lots of others that they can't disclose because they're not yet fixed. [0] https://www.anthropic.com/news/mozilla-firefox-security [1] https://red.anthropic.com/2026/zero-days/ | |
| ▲ | dist-epoch 11 hours ago | parent | prev [-] | | > On the kernel security list we've seen a huge bump of reports. We were between 2 and 3 per week maybe two years ago, then reached probably 10 a week over the last year with the only difference being only AI slop, and now since the beginning of the year we're around 5-10 per day depending on the days (fridays and tuesdays seem the worst). Now most of these reports are correct, to the point that we had to bring in more maintainers to help us. ... Also it's interesting to keep thinking that these bugs are within reach from criminals so they deserve to get fixed. https://lwn.net/Articles/1065620/ | | |
|
|
| ▲ | desireco42 7 hours ago | parent | prev | next [-] |
| A developer using Claude Code found this bug. Claude is a tool. It is used by developers. It should not sign commits. Neovim never tried to sign commits with me, nor Zed. |
| |
| ▲ | igravious 6 hours ago | parent [-] | | Should not Is that your new law? The non-agentic “Neovim and Zed *never tried to sign commits [for]~~with~~ me” therefore no tool ever no matter how advanced is not allowed to sign a commit. Did it ever occur to you that for whatever reason you just might not be cut out for the software treadmill? |
|
|
| ▲ | _pdp_ 12 hours ago | parent | prev | next [-] |
| The title is a little misleading. It was Opus 4.6 (the model). You could discover this with some other coding agent harness. The other thing that bugs me and frankly I don't have the time to try it out myself, is that they did not compare to see if the same bug would have been found with GPT 5.4 or perhaps even an open source model. Without that, and for the reasons I posted above, while I am sure this is not the intention, the post reads like an ad for claude code. |
| |
| ▲ | mtlynch 11 hours ago | parent | next [-] | | OP here. I don't understand this critique. Carlini did use Claude Code directly. Claude Code used the Claude Opus 4.6 model, but I don't know why you'd consider it inaccurate to say Claude Code found it. GPT 5.4 might be capable of finding it as well, but the article never made any claims about whether non-Anthropic models could find it. If I wrote about achieving 10k QPS with a Go server, is the article misleading unless I enumerate every other technology that could have achieved the same thing? | |
| ▲ | mgraczyk 12 hours ago | parent | prev | next [-] | | No the title is correct and you are misreading or didn't read. It was found with Claude code, that's the quote. This isn't a model eval, it's an Anthropic employee talking about Claude code. So comparing to other models isn't a thing to reasonably expect. | |
| ▲ | weird-eye-issue 10 hours ago | parent | prev [-] | | > You could discover this with some other coding agent harness. And surely that would be relevant if they were using a different harness. |
|
|
| ▲ | skyskys 4 hours ago | parent | prev | next [-] |
| no hecking wayyy!!!! claude chud code!!! |
|
| ▲ | yunnpp 4 hours ago | parent | prev [-] |
| I really "like" the premise of this talk: "We are releasing dangerous weapons, help us make the world safer! Also, it'll cost you a subscription, and we'll profit from it." Imagine if the speaker was a biochemist talking about releasing chemical bombs in your neighborhood that somehow you just could not avoid. The software industry is such a joke right now. |