| ▲ | shagie 2 days ago | ||||||||||||||||||||||||||||||||||||||||
Make up a fanciful problem and ask it to solve it. For example, https://chatgpt.com/s/t_691f6c260d38819193de0374f090925a is unlikely to be found in the training data - I just made it up. Another example of wizards and witches and warriors and summoning... https://chatgpt.com/share/691f6cfe-cfc8-8011-b8ca-70e2c22d36... - I doubt that was in the training data either. Make up puzzles of your own and see if it is able to solve it or not. The blanket claim of "cannot solve problems that are not in its training data" seems to be something that can be disproven by making up a puzzle from your own human creativity and seeing if it can solve it - or for that matter, how it attempts to solve it. It appears that there is some ability for it to reason about new things. I believe that much of this "an LLM can't do X" or "an LLM is parroting tokens that it was trained on" comes from trying to claim that all the material that it creates was created before, by a human and any use of an LLM is stealing from some human and thus unethical to use. ( ... and maybe if my block world or wizards and warriors and witches puzzle was in the training data somewhere, I'm unconsciously copying something somewhere else and my own use of it is unethical. ) | |||||||||||||||||||||||||||||||||||||||||
| ▲ | wadadadad 2 days ago | parent | next [-] | ||||||||||||||||||||||||||||||||||||||||
This is an interesting idea, but as you stated, it's all logic; it's hard to come up with an idea where you don't have to explain concepts yet still is dissimilar enough to be in the training. In your second example with the wizards- did you notice that it failed to follow the rules? Step 3, the witch was summoned by the wizard. I'm curious as to why you didn't comment either way on this. On a related note, instead of puzzles, what about presenting riddles? I would argue that riddles are creative, pulling bits and pieces of meaning from words to create an answer. If AI can solve riddles not seen before, would that count as creative and not solving problems in their dataset? Here's one I created and presented (the first incorrect answer I got was Escape Room; I gave it 10 attempts and it didn't get the answer I was thinking of): --- Solve the riddle: Chaos erupts around The shape moot The goal is key | |||||||||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||||||||
| ▲ | Ardren 2 days ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||||||||
I think your example works, as it does try to solve a problem it hasn't seen (though it is very similar to existing problems). ... But, CharGPT makes several mistakes :-) > Wizard Teleport: Wz1 teleports himself and Wz2 to Castle Beta. This means Wz1 has used his only teleport power. Good. > Witch Summon: From Castle Beta, Wi1 at Castle Alpha is summoned by Wz1. Now Wz1 has used his summon power. Wizzard1 cannot summon. > Wizard Teleport: Now, Wz2 (who is at Castle Beta) teleports back to Castle Alpha, taking Wa1 with him. Warrior1 isn't at Castle beta > Wizard Teleport: Wz2, from Castle Alpha, teleports with Wa2 to Castle Beta. Wizzard2 has already teleported | |||||||||||||||||||||||||||||||||||||||||