| Looking back with fresh eyes, I definitely think I could’ve presented what I’m trying to say better. On a purely technical play, you’re right that I’m drawing a distinction that may not hold up purely on technical grounds. Maybe the better framing is: I trust constrained, single purpose models with somewhat verifiable outputs (seeing text go in, translated text go out, compare its consistency) more than I trust general purpose models with broad access to my browsing context, regardless of whether they’re both neural networks under the hood. WRT to the “scope”, maybe I have picked up the wrong end of the stick with what Mozilla are planning to do - but they’ve already picked all the low hanging fruit with AI integration with the features you’ve mentioned and the fact they seem to want to dig their heels in further, at least to me, signals that they want deeper integration? Although who knows, the post from the new CEO may also be a litmus test to see what the response to that post elicits, and then go from there. |
| |
| ▲ | MrAlex94 5 days ago | parent [-] | | I disagree that it’s fear mongering. Have we not had numerous articles on HN about data exfiltration in recent memory? Why would an LLM that is in the drivers seat of a browser (not talking about current feature status in Firefox wrt to sanitised data being interacted with) not have the same pitfalls? Seems as if we’d be 3 for 3 in the “agents rule of 2” in the context of the web and a browser? > [A] An agent can process untrustworthy inputs > [B] An agent can have access to sensitive systems or private data > [C] An agent can change state or communicate externally https://simonwillison.net/2025/Nov/2/new-prompt-injection-pa... Even if we weren’t talking about such malicious hypotheticals, hallucinations are a common occurrence as are CLI agents doing things it thinks best, sometimes to the detriment of the data it interacts with. I personally wouldn’t want my history being modified or deleted, same goes with passwords and the like. It is a bit doomerist, I doubt it’ll have such broad permissions but it just doesn’t sit well which I suppose is the spirit of the article and the stance Waterfox takes. | | |
| ▲ | dkdcio 5 days ago | parent | next [-] | | > Have we not had numerous articles on HN about data exfiltration in recent memory? there’s also an article on the front page of HN right now claiming LLMs are black boxes and we don’t know how they work, which is plainly false. this point is hardly evidence of anything and equivalent to “people are saying” | | |
| ▲ | FeepingCreature 5 days ago | parent [-] | | This is true though. While we know what they do on a mechanistic level, we cannot reliably analyze why the model outputs any particular answer in functional terms without a heroic effort at the "arxiv paper" level. | | |
| ▲ | dkdcio 5 days ago | parent [-] | | that’s true of analyzing individual atoms in a combustion engine — yet I doubt you’d claim we don’t know how they work also this went from “we can’t analyze” to “we can’t analyze reliably [without a lot of effort]” quite quickly | | |
| ▲ | twosdai 5 days ago | parent | next [-] | | In the digital world, we should be able to go back from output to input unless the intention of the function is to "not do that". Like hashing. Llms not being able to go from output back to input deterministically and for us to understand why is very important, most of our issues with llms stem from this issue. Its why mechanistic interpretabilty research is so hot right now. The car analogy is not good because models are digital components and a car is a real world thing. They are not comparable. | | |
| ▲ | dkdcio 5 days ago | parent [-] | | ah I forgot digital components are not real world things |
| |
| ▲ | FeepingCreature 5 days ago | parent | prev [-] | | I mean, fluid dynamics is an unsolved issue. But even so we know *considerably* less about how LLMs work in functional terms than about how combustion engines work. | | |
| ▲ | dkdcio 5 days ago | parent [-] | | I outright disagree; we know how LLMs work | | |
| ▲ | int_19h 5 days ago | parent [-] | | We know how neural nets work. We don't know how a specific combination of weights in the net is capable of coherently asking questions asked in a natural language, though. If we did, we could replicate what it does without training it. | | |
| ▲ | dkdcio 4 days ago | parent [-] | | > We know how neural nets work. We don't know how a specific combination of weights in the net is capable of coherently asking questions asked in a natural language, though. these are the same thing. the neural network is trained to predict the most likely next word (rather token, etc.) — that’s how it works. that’s it. you train a neural network on data, it learns the function you trained it to, it “acts” like the data. have you actually studied neural networks? do you know how they work? I’m confused why you and so many others are seemingly so confused by this. what fundamentally are you asking for to meet the criteria of knowing how LLMs work? some algorithm that can look at weights and predict if the net will output “coherent” text? > If we did, we could replicate what it does without training it. not sure what this is supposed to mean | | |
| ▲ | FeepingCreature 4 days ago | parent [-] | | It's like you're describing a compression program as "it takes a big file and returns a smaller file by exploiting regularities in the data." Like, you have accurately described what it does, but you have in no way answered the question of how it does that. If you then explain the function of a CPU and how ELF binaries work (which is the equivalent of trying to answer the question by explaining how neural networks work), you then have still not answered the actually important question! Which is "what are the algorithms that LLMs have learnt that allow them to (apparently) converse and somewhat reason like humans?" | | |
| ▲ | dkdcio 4 days ago | parent [-] | | …except we know what every neuron in a neural network is doing. I ask again, what criteria do we need to meet for you to claim we know how LLMs work? we know the equations, we know the numbers going through a network, we know the universal approximation theorem —- what’re you looking for exactly? I’ve answered the “what have they learnt” bit; a function that predicts the next token based on data. what more do you need? | | |
| ▲ | FeepingCreature 4 days ago | parent [-] | | Yes, in the analogy it's equivalent to saying you know "what" every instruction in the compression program is doing. push decrements rsp, xor rax, rax zeroes out the register. You know every step. But you don't know the algorithm that those instructions are implementing, and that's the same situation we're in with LLMs. We can describe their actions numerically, but we cannot describe them behaviorally, and they're doing things that we don't know how to otherwise do with numerical methods. They've clearly learnt algorithms but we cannot yet formalize what they are. The universal approximation theorem actually works against your argument here, because it's too powerful- they could be implementing anything. edit: We know the data that their function outputs, it's a "blurry jpeg of the internet" because that's what they're trained on. But we do not know what the function is, and being able to blurrily compress the internet into a tb or whatever is utterly beyond any other compression algorithm known to man. |
|
|
|
|
|
|
|
|
| |
| ▲ | yunohn 5 days ago | parent | prev [-] | | I believe you are conflating multiple concepts to prove a flaky point. Again, unless your agent has access to a function that exfiltrates data, it is impossible for it to do so. Literally! You do not need to provide any tools to an LLM that summarizes or translates websites, manages your open tabs, etc. This can be done fully locally in a sandbox. Linking to simonw does not make your argument valid. He makes some great points, but he does not assert what you are claiming at any point. Please stop with this unnecessary fear mongering and make a better argument. | | |
| ▲ | nazgul17 5 days ago | parent [-] | | Thinking aloud, but couldn't someone create a website with some malicious text that, when quoted in a prompt, convinces the LLM to expose certain private data to the web page, and couldn't the webpage send that data to a third party, without the need for the LLM to do so? This is probably possible to mitigate, but I fear what people more creative, motivated and technically adept could come up with. | | |
|
|
|