▲ | ijk 3 days ago | |||||||
I'm frustrated by the number of times I encounter people assuming that the current model behavior is inevitable. There's been hundreds of billions of dollars spent on training LLMs to do specific things. What exactly they've been trained on matters; they could have been trained to do something else. Interacting with a base model versus an instruction tuned model will quickly show you the difference between the innate language faculties and the post-trained behavior. | ||||||||
▲ | Workaccount2 3 days ago | parent [-] | |||||||
Some of the Anthropic guys have said that the core thing holding the models back is training, and they're confident the gains will keep coming as they figure out how to onboard more and more training data. So yeah, Claude might suck at reading and writing plumbing diagrams, but they claim the barrier is simply a function of training, not any kind of architectural limitation. | ||||||||
|