| ▲ | __alexs 4 hours ago | |||||||
I think OpenAI and Anthropic probably have a lot of that lying around by now. | ||||||||
| ▲ | jhrmnn 4 hours ago | parent | next [-] | |||||||
So most training data would be grey and a little bit coloured? Ok, that sounds plausible. But then maybe they tried and the current models get it already right 99.99% of the time, so observing any improvement is very hard. | ||||||||
| ▲ | nairboon 4 hours ago | parent | prev [-] | |||||||
They have a lot of data in the form: user input, LLM output. Then the model learns what the previous LLM models produced, with all their flaws. The core LLM premise is that it learns from all available human text. | ||||||||
| ||||||||