| ▲ | throwaway27448 5 hours ago | ||||||||||||||||||||||
Even at orders of magnitude greater speed, we've still hit diminishing returns for quality of output. We simply haven't found anything like superhuman reasoning ability, just superhuman (potentially) reasoning speed. | |||||||||||||||||||||||
| ▲ | LarsDu88 3 hours ago | parent | next [-] | ||||||||||||||||||||||
I disagree with this. Reinforcement learning with verifiable rewards training is actually the secret sauce that is leading Claude and GPT to automating software engineering tasks. All the easily verifiable domains such as mathematics, coding, and things that can be run inside a reasonable simulation are falling very very fast. By next year if not sooner, mathematicians will be wildly outpaced by LLMs for reasoning. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | energy123 4 hours ago | parent | prev | next [-] | ||||||||||||||||||||||
It's not that easy to assess diminishing returns with saturated benchmarks where asymptoting to 100% is mathematically baked in. I could point to the number of Erdos proofs being solved by AI going from 0 to many very recently as evidence for acceleration. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | horsawlarway 4 hours ago | parent | prev [-] | ||||||||||||||||||||||
Possibly - but we've also seen that spending more tokens on a task can improve the quality of the output (reasoning, CoT, etc). So it's not impossible to have things that seem orthogonal, like generation speed or context length, have an impact on quality of result. | |||||||||||||||||||||||