▲ | hackinthebochs a day ago | ||||||||||||||||||||||||||||
LLMs aren't just modeling word co-occurrences. They are recovering the underlying structure that generates word sequences. In other words, they are modeling the world. This model is quite low fidelity, but it should be very clear that they go beyond language modeling. We all know of the pelican riding a bicycle test [1]. Here's another example of how various language models view the world [2]. At this point it's just bad faith to claim LLMs aren't modeling the world. [1] https://simonwillison.net/2025/Aug/7/gpt-5/#and-some-svgs-of... [2] https://www.lesswrong.com/posts/xwdRzJxyqFqgXTWbH/how-does-a... | |||||||||||||||||||||||||||||
▲ | SR2Z a day ago | parent | next [-] | ||||||||||||||||||||||||||||
The "pelican on a bicycle" test has been around for six months and has been discussed a ton on the internet; that second example is fascinating but Wikipedia has infoboxes containing coordinates like 48°51′24″N 2°21′8″E (Paris, notoriously on land). How much would you bet that there isn't a CSV somewhere in the training set exactly containing this data for use in some GIS system? I think that "modeling the world" is a red herring, and that fundamentally an LLM can only model its input modalities. Yes, you could say this about human beings, but I think a more useful definition of "model the world" is that a model needs to realize any facts that would be obvious to a person. The fact that frontier models can easily be made to contradict themselves is proof enough to me that they cannot have any kind of sophisticated world model. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
▲ | homarp a day ago | parent | prev [-] | ||||||||||||||||||||||||||||
and we can say that a bastardized version of the Sapir-Worf hypothesis applies: what's in the training set shapes or limits LLM's view of the world | |||||||||||||||||||||||||||||
|