Remix.run Logo
qntty 3 hours ago

Pre-training mean exposing an already-trained model to more raw text like PDF extracts etc (aka continued pre-training). You wouldn't be starting from scratch, but it's still pre-training because the objective is just next token prediction of the text you expose it to.

Post-training means everything else: SFT, DPO, RL, etc. Anything that involves things like prompt/response pairs, reward models, or benefits from human feedback of any kind.

losvedir 3 hours ago | parent [-]

Er, then what is the "already trained" model? I thought pre-training was the gradient descent through the internet part of building foundational models.