▲ | fooker 4 days ago | |||||||||||||||||||||||||
Fine tuning and training are very different beasts. | ||||||||||||||||||||||||||
▲ | kouteiheika 4 days ago | parent [-] | |||||||||||||||||||||||||
No they're not? The process is essentially exactly the same, just with a much lower total FLOPs budget, since if you're not training from scratch then you don't need to train for as long. I can use *exactly* the same code that I used to fine-tune a model to train a new model from scratch; literally the only difference is whether I initialize the initial weights randomly or with an existing model, a couple of hyperparameters (e.g. for training from scratch you want to start at a higher LR), and training for longer. | ||||||||||||||||||||||||||
|