▲ | westurner 3 days ago | |||||||||||||||||||||||||
> That's fair, they are relabelling colours and rotating the boards. Photometric augmentation, Geometric augmentation > I meant more like mass generation of novel puzzles to try and train specific patterns. What is the difference between Synthetic Data Generation and Self Play (like AlphaZero)? Don't self play simulations generate synthetic training data as compared to real observations? | ||||||||||||||||||||||||||
▲ | bubblyworld 3 days ago | parent [-] | |||||||||||||||||||||||||
I don't know the jargon, but for me the main thing is the distinction between humans injecting additional bits of information into the training set vs the algorithm itself discovering those bits of information. So self-play is very interesting (it's automated as part of the algorithm) but stuff like generating tons of novel sudoku puzzles and adding them to the training set is less interesting (the information is being fed into the training set "out-of-band", so to speak). In this case I was wrong, the authors are clearly adding bits of information themselves by augmenting the dataset with symmetries (I propose "symmetry augmentation" as a much more sensible phrase for this =P). Since symmetries share a lot of mutual information with each other, I don't think this is nearly as much of a crutch as adding novel data points into the mix before training, but ideally no augmentation would be needed. I guess you could argue that in some sense it's fair play - when humans are told the rules of sudoku the symmetry is implicit, but here the AI is only really "aware" of the gradient. | ||||||||||||||||||||||||||
|