| ▲ | jaccola 4 hours ago |
| All of the latest models I've tried actually pass this test. What I found interesting was all of the success cases were similar to: e.g. "Drive. Most car washes require the car to be present to wash,..." Only most?! They have an inability to have a strong "opinion" probably because their post training, and maybe the internet in general, prefer hedged answers.... |
|
| ▲ | Waterluvian 4 hours ago | parent | next [-] |
| Here’s my take: boldness requires the risk of being wrong sometimes. If we decide being wrong is very bad (which I think we generally have agreed is the case for AIs) then we are discouraging strong opinions. We can’t have it both ways. |
| |
|
| ▲ | hansmayer 3 hours ago | parent | prev | next [-] |
| > They have an inability to have a strong "opinion" probably What opinion? It's evaluation function simply returned the word "Most" as being the most likely first word in similar sentences it was trained on. It's a perfect example showing how dangerous this tech could be in a scenario where the prompter is less competent in the domain they are looking an answer for. Let's not do the work of filling in the gaps for the snake oil salesmen of the "AI" industry by trying to explain its inherent weaknesses. |
| |
| ▲ | 2 hours ago | parent | next [-] | | [deleted] | |
| ▲ | wilg 2 hours ago | parent | prev | next [-] | | Presumably the OP scare quoted "opinion" precisely to avoid having to get into this tedious discussion. | |
| ▲ | lkeskull 3 hours ago | parent | prev [-] | | this example worked in 2021, it's 2026. wake up. these models are not just "finding the most likely next word based on what they've seen on the internet". | | |
| ▲ | strix_varius 3 hours ago | parent | next [-] | | Well, yes, definitionally they are doing exactly that. It just turns out that there's quite a bit of knowledge and understanding baked into the relationships of words to one another. LLMs are heavily influenced by preceding words. It's very hard for them to backtrack on an earlier branch. This is why all the reasoning models use "stop phrases" like "wait" "however" "hold on..." It's literally just text injected in order to make the auto complete more likely to revise previous bad branches. | |
| ▲ | jaccola 2 hours ago | parent | prev | next [-] | | The person above was being a bit pedantic, and zealous in their anti-anthropomorphism. But they are literally predicting the next token. They do nothing else. Also if you think they were just predicting the next token in 2021, there has been no fundamental architecture change since then. All gains have been via scale and efficiency optimisations (not to discount that, an awful lot of complexity in both of these) | | |
| ▲ | nearbuy 2 hours ago | parent [-] | | That's not what they said. They said: > It's evaluation function simply returned the word "Most" as being the most likely first word in similar sentences it was trained on. Which is false under any reasonable interpretation. They do not just return the word most similar to what they would find in their training data. They apply reasoning and can choose words that are totally unlike anything in their training data. If you prompt it: > Complete this sentence in an unexpected way: Mary had a little... It won't say lamb. Any if you think whatever it says was in the training data, just change the constraints until you're confident it's original. (E.g. tell it every word must start with a vowel and it should mention almonds.) "Predicting the next token" is also true but misleading. It's predicting tokens in the same sense that your brain is just minimizing prediction error under predictive coding theory. | | |
| ▲ | hansmayer an hour ago | parent [-] | | You are actually proving my point with your example, if you think about it a bit more. |
|
| |
| ▲ | csomar 2 hours ago | parent | prev [-] | | Unless LLMs architecture have changed, that is exactly what they are doing. You might need to learn more how LLMs work. | | |
| ▲ | andy12_ an hour ago | parent [-] | | Unless the LLM is a base model or just a finetuned base model, it definitely doesn't predict words just based on how likely they are in similar sentences it was trained on. Reinforcement learning is a thing and all models nowadays are extensively trained with it. If anything, they predict words based on a heuristic ensemble of what word is most likely to come next in similar sentences and what word is most likely to give a final higher reward. | | |
| ▲ | csomar 7 minutes ago | parent | next [-] | | > If anything, they predict words based on a heuristic ensemble of what word is most likely to come next in similar sentences and what word is most likely to give a final higher reward. So... "finding the most likely next word based on what they've seen on the internet"? | |
| ▲ | hansmayer 44 minutes ago | parent | prev [-] | | You know that when A. Karpathy released NanoLLM (or however it was called), he said it was mainly coded by hand as the LLMs were not helpful because "the training dataset was way off". So yeah, your argumentation actually "reinforces" my point. | | |
| ▲ | andy12_ 35 minutes ago | parent [-] | | No, your opinion is wrong because the reason some models don't seem to have some "strong opinion" on anything is not related to predicting words based on how similar they are to other sentences in the training data. It's most likely related to how the model was trained with reinforcement learning, and most specifically, to recent efforts by OpenAI to reduce hallucination rates by penalizing guessing under uncertainty[1]. [1] https://cdn.openai.com/pdf/d04913be-3f6f-4d2b-b283-ff432ef4a... | | |
| ▲ | hansmayer 22 minutes ago | parent [-] | | Well, you do understand the "penalising" or as the ML scientific community likes to call it - "adjusting the weights downwards" - is part of setting up the evaluation functions, for gasp - calculating the next most likely tokens, or to be more precise, tokens with the highest possible probability? You are effectively proving my point, perhaps in a bit hand-wavy fashion, that nevertheless still can be translated into the technical language. | | |
| ▲ | andy12_ 3 minutes ago | parent [-] | | You do understand that the mechanism through which an auto-regressive transformer works (predicting one token at a time) is completely unrelated to how a model with that architecture behaves or how it's trained, right? You can have both: - An LLM that works through completely different mechanisms, like predicting masked words, predicting the previous word, or predicting several words at a time. - A normal traditional program, like a calculator, encoded as an autoregressive transformer that calculates its output one word at a time (compiled neural networks) [1][2] So saying "it predicts the next word" is a nothing-burger. That a program calculates its output one token at a time tells you nothing about its behavior. [1] https://arxiv.org/pdf/2106.06981 [2] https://wengsyx.github.io/NC/static/paper_iclr.pdf |
|
|
|
|
|
|
|
|
| ▲ | andersmurphy 3 hours ago | parent | prev | next [-] |
| Did you try several times per model? In my experience it's luck of the draw. All the models I tried managed to get it wrong at least once. The models that had access to search got ot right.But, then were just dealing with an indirect version of Google. (And they got it right for the wrong reasons... I.e this is a known question designed to confuse LLMs) |
|
| ▲ | 6 minutes ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | jl6 3 hours ago | parent | prev | next [-] |
| I guess it didn’t want to rule out the existence of ultra-powerful water jets that can wash a car in sniper mode. |
|
| ▲ | madeofpalk 2 hours ago | parent | prev | next [-] |
| I enjoyed the Deepseek response that said “If you walk there, you'll have to walk back anyway to drive the car to the wash.” There’s a level of earnestness here that tickles my brain. |
|
| ▲ | deevus 3 hours ago | parent | prev | next [-] |
| I tried with Opus 4.6 Extended and it failed. LLMs are non deterministic so I'm guessing if I try a couple of times it might succeed. |
|
| ▲ | 2 hours ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | nozzlegear 3 hours ago | parent | prev | next [-] |
| Opus 4.6 answered with "Drive." Opus 4.6 in incognito mode (or whatever they call it) answered with "Walk." |
|
| ▲ | sneak 44 minutes ago | parent | prev | next [-] |
| There are car wash services that will come to where your car is and wash it. It’s not wrong! |
|
| ▲ | yanis_t 3 hours ago | parent | prev | next [-] |
| > Most car washes...
I read it as slight-sarcasm answer |
|
| ▲ | Puts 3 hours ago | parent | prev | next [-] |
| > Only most?! What if AI developed sarcasm without us knowing… xD |
| |
|
| ▲ | dyauspitr 3 hours ago | parent | prev [-] |
| There are mobile car washes that come to your house. |
| |
| ▲ | Loocid 3 hours ago | parent | next [-] | | That still requires a car present to be washed though. | | |
| ▲ | column 2 hours ago | parent [-] | | but you can walk over to them and tell them to go wash the car that is 50 meters away. no driving involved. | | |
| |
| ▲ | andersmurphy 3 hours ago | parent | prev [-] | | Do they involve you walking to them first? | | |
| ▲ | learingsci 3 hours ago | parent [-] | | You could, but presumably most people call. I know of such a place. They wash cars on the premises but you could walk in and arrange to have a mobile detailing appointment later on at some other location. |
|
|