▲ | marcus_holmes 4 days ago | ||||||||||||||||||||||
I notice there's no prompt saying "you should try to win the game" yet the results are measured by how much the LLM wins. Is this implicit in the "you are a grandmaster chess player" prompt? Is there some part of the LLM training that does "if this is a game, then I will always try to win"? Could the author improve the LLM's odds of winning just by telling it to try and win? | |||||||||||||||||||||||
▲ | tinco 4 days ago | parent | next [-] | ||||||||||||||||||||||
I think you're putting too much weight on its intentions, it doesn't have intentions it is a mathematical model that is trained to give the most likely outcome. In almost all examples and explanations it has seen from chess games, each player would be trying to win, so it is simply the most logical thing for it to make a winning move. So I wouldn't expect explicitly prompting it to win to improve its performance by much if at all. The reverse would be interesting though, if you would prompt it to make losing/bad moves, would it be effective in doing so, and would the moves still be mostly legal? That might reveal a bit more about how much relies on concepts it's seen before. | |||||||||||||||||||||||
| |||||||||||||||||||||||
▲ | Nashooo 4 days ago | parent | prev | next [-] | ||||||||||||||||||||||
IMO this is clearly implicit in the "you are a grandmaster chess player" prompt. As that should make generating best possible move tokens more likely. | |||||||||||||||||||||||
| |||||||||||||||||||||||
▲ | tananan 3 days ago | parent | prev | next [-] | ||||||||||||||||||||||
It would surely just be fluff in the prompt. The model's ability to generate chess sequences will be bounded by the expertise in the pool of games in the training set. Even if the pool was poisoned by games in which some players are trying to lose (probably insignificant), no one annotates player intent in chess games, and so prompting it to win or lose doesn't let the LLM pick up on this. You can try this by asking an LLM to play to lose. ChatGPT ime tries to set itself up for scholar's mate, but if you don't go for it, it will implicitly start playing to win (e.g. taking your unprotected pieces). If you ask it "why?", it gives you the usual bs post-hoc rationalization. | |||||||||||||||||||||||
| |||||||||||||||||||||||
▲ | montjoy 4 days ago | parent | prev | next [-] | ||||||||||||||||||||||
I came to the comments to say this too. If you were prompting it to generate code, you generally get better results when you ask it for a result. You don’t just tell it, “You are a python expert and here is some code”. You give it a direction you want the code to go. I was surprised that there wasn’t something like, “and win”, or, “black wins”, etc. | |||||||||||||||||||||||
▲ | boredhedgehog 3 days ago | parent | prev [-] | ||||||||||||||||||||||
Further, the prompt also says to "choose the next move" instead of the best move. It would be fairly hilarious if the reinforcement training has made the LLM unwilling to make the human feel bad through losing a game. |