▲ | vidarh 4 days ago | |||||||||||||||||||||||||
To say any specific LLM can reason is a somewhat significant claim. To say LLMs as a class is architecturally able to be trained to reason is - in the complete absence of evidence to suggest humans can compute functions outside the Turing computable - is effectively only an argument that they can implement a minimal Turing machine given the context is used as IO. Given the size of the rules needed to implement the smallest known Turing machines, it'd take a really tiny model for them to be unable to. Now, you can then argue that it doesn't "count" if it needs to be fed a huge program step by step via IO, but if it can do something that way, I'd need some really convincing evidence for why the static elements those steps could not progressively be embedded into a model. | ||||||||||||||||||||||||||
▲ | wizzwizz4 3 days ago | parent [-] | |||||||||||||||||||||||||
No such evidence exists: we can construct such a model manually. I'd need some quite convincing evidence that any given training process is approximately equivalent to that, though. | ||||||||||||||||||||||||||
|