| ▲ | cpldcpu 5 hours ago | |||||||
Yes, marks of AI all over the place. Also the SVGs. >No solution written, 100% score. Its weird. Turns out that hardest problem for LLMs to really tackle is long-form text. | ||||||||
| ▲ | basch 5 hours ago | parent | next [-] | |||||||
Maybe in one shot. In theory I would expect them to be able to ingest the corpus of the new yorker and turn it into a template with sub-templates, and then be able to rehydrate those templates. The harder part seems to be synthesizing new connection from two adjacent ideas. They like to take x and y and create x+y instead of x+y+z. | ||||||||
| ▲ | sidpatil 5 hours ago | parent | prev [-] | |||||||
Someone here mentioned a whole ago that the labs deliberately haven't tried to train these characteristics out of their models, because leaving them in makes it easier to identify, and therefore exclude, LLM-generated text from their training corpus. | ||||||||
| ||||||||