▲ | daxfohl a day ago | |||||||||||||||||||||||||||||||||||||||||||
Yeah, I can't even get them to retain a simple state. I've tried having them run a maze, but instead of giving them the whole maze up front, I have them move one step at a time, tell them which directions are open from that square and ask for the next move, etc. After a few moves they get hopelessly lost and just start wandering back and forth in a loop. Even when I prompt them explicitly to serialize a state representation of the maze after each step, and even if I prune the old context so they don't get tripped up on old state representations, they still get flustered and corrupt the state or lose track of things eventually. They get the concept: if I explain the challenge and ask to write a program to solve such a maze step-by-step like that, they can do that successfully first-try! But maintaining it internally, they still seem to struggle. | ||||||||||||||||||||||||||||||||||||||||||||
▲ | nomadpenguin a day ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||
There are specialized architectures (the Tolman-Eichenbaum Machine)* that are able to complete this kind of task. Interestingly, once trained, their activations look strikingly similar to place and grid cells in real brains. The team were also able to show (in a separate paper) that the TEM is mathematically equivalent to a transformer. * https://www.sciencedirect.com/science/article/pii/S009286742... | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||
▲ | kqr 14 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||
My experience in trying to get them to play text adventures[1] is similar. I had to prompt with very specific leading questions to give them a decent chance of even recognising the main objective after the first few steps. [1]: https://entropicthoughts.com/getting-an-llm-to-play-text-adv... | ||||||||||||||||||||||||||||||||||||||||||||
▲ | warrenm a day ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||
>I've tried having them run a maze, but instead of giving them the whole maze up front, I have them move one step at a time, tell them which directions are open from that square and ask for the next move, etc. Presuming these are 'typical' mazes (like you find in a garden or local corn field in late fall), why not have the bot run the known-correct solving algorithm (or its mirror)? | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||
▲ | yberreby 17 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||
It took me a second to realize you were talking about prompting a LLM. This is fundamentally different from what the parent is doing. "AI" is so much more than "talking to a pretrained LLM." |