| ▲ | gpm 7 hours ago | ||||||||||||||||
What are you doing in your loop? As typically deployed [1] LLMs are not turing complete. They're closer to linear bounded automaton, but because transformers have a strict maximum input size they're actually a subset of the weaker class of deterministic finite automaton. These aren't like python programs or something that can work on as much memory as you supply them, their architecture works on a fixed maximum amount of memory. I'm not particularly convinced turing complete is the relevant property though. I'm rather convinced that I'm not turing complete either... my head is only so big after all. [1] i.e. in a loop that appends output tokens to the input and has some form of sliding context window (perhaps with some inserted instructions to "compact" and then sliding the context window right to after those instructions once the LLM emits some special "done compacting" tokens). [2] Common sampling procedures make them mildly non-deterministic, but I don't believe they do so in a way that changes the theoretical class of these machines from DFAs. | |||||||||||||||||
| ▲ | vidarh 6 hours ago | parent | next [-] | ||||||||||||||||
Context effectively provifes an IO port, and so all the loop needs to do is to simulate the tape head, and provide a single token of state. You can not be convinced Turing complete is relevant all you want - we don't know of any more expansive category of computable functions, and so given that an LLM in the setup described is Turing complete no matter that they aren't typically deployed that way is irrelevant. They trivially can be, and that is enough to make the shallow dismissal of pointing out they're "just" predicting the next token meaningless. | |||||||||||||||||
| ▲ | roywiggins 6 hours ago | parent | prev | next [-] | ||||||||||||||||
Turing Machines don't need access to the entire tape all at once, it's sufficient for it to see one cell at a time. You could certainly equip an LLM with a "read cell", "write cell", and "move left/right" tool and now you have a Turing machine. It doesn't need to keep any of its previous writes or reads in context. A sliding context window is more than capacious enough for this. | |||||||||||||||||
| |||||||||||||||||
| ▲ | empath75 5 hours ago | parent | prev [-] | ||||||||||||||||
No physically realizable machine is technically turing complete. But it is trivially possible to give systems-including-LLMs external storage that is accessible on demand. | |||||||||||||||||