▲ | moyix a day ago | |||||||||||||||||||||||||||||||
Note that MuZero did better than AlphaGo, without access to preprogrammed rules: https://en.wikipedia.org/wiki/MuZero | ||||||||||||||||||||||||||||||||
▲ | smokel a day ago | parent | next [-] | |||||||||||||||||||||||||||||||
Minor nitpick: it did not use preprogrammed rules for scanning through the search tree, but it does use preprogrammed rules to enforce that no illegal moves are made during play. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
▲ | CGamesPlay 14 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||
This is true, and MuZero's paper notes that it did better with less computation than AlphaZero. But it still used about 10x more computation to get there than AlphaGo, which was "bootstrapped" with human expert moves. I think this is very important context to anyone who is trying to implement an AI for their own game. |