| ▲ | datsci_est_2015 an hour ago | ||||||||||||||||
So why then do we stop training LLMs and keep them stored at a specific state? Is it perhaps because the results become terrible and LLMs have a delicate optimal state for general use? This sounds like an even worse case for a model of intelligence. | |||||||||||||||||
| ▲ | stavros an hour ago | parent [-] | ||||||||||||||||
Nope, it's not that, but it's nice of you to offer a straw man. Makes the argument flow better. | |||||||||||||||||
| |||||||||||||||||