| ▲ | tovej 6 hours ago | ||||||||||||||||
It generated a proof that was close enough to something in its training data to be generated. | |||||||||||||||||
| ▲ | keeda 4 hours ago | parent | next [-] | ||||||||||||||||
That may be, and we can debate the level of novelty, but it is novel, because this exact proof didn't exist before, something which many claim was not possible with AI. In fact, just a few years ago, based on some dabbling in NLP a decade ago, I myself would not have believed any of this was remotely possible within the next 3 - 5 decades at least. I'm curious though, how many novel Math proofs are not close enough to something in the prior art? My understanding is that all new proofs are compositions and/or extensions of existing proofs, and based on reading pop-sci articles, the big breakthroughs come from combining techniques that are counter-intuitive and/or others did not think of. So roughly how often is the contribution of a proof considered "incremental" vs "significant"? | |||||||||||||||||
| |||||||||||||||||
| ▲ | qnleigh 6 hours ago | parent | prev [-] | ||||||||||||||||
Do you know that from reading the proof, or are you just assuming this based on what you think LLMs should be capable of? If the latter, what evidence would be required for you to change your mind? - Edit: I can't reply, probably because the comment thread isn't allowed to go too deep, but this is a good argument. In my mind the argument isn't that coding is harder than math, but that the problems had resisted solution by human researchers. | |||||||||||||||||
| |||||||||||||||||