| ▲ | energy123 15 hours ago | ||||||||||||||||||||||||||||||||||
A surprising % of these LLM proofs are coming from amateurs. One wonders if some professional mathematicians are instead choosing to publish LLM proofs without attribution for career purposes. | |||||||||||||||||||||||||||||||||||
| ▲ | kristopolous 15 hours ago | parent | next [-] | ||||||||||||||||||||||||||||||||||
It's probably from the perennial observation "This LLM is kinda dumb in the thing I'm an expert in" | |||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||
| ▲ | Workaccount2 4 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||
>One wonders if some professional mathematicians are instead choosing to publish LLM proofs without attribution for career purposes. This will just become the norm as these models improve, if it isn't largely already the case. It's like sports where everyone is trying to use steroids, because the only way to keep up is to use steroids. Except there aren't any AI-detectors and it's not breaking any rules (except perhaps some kind of self moral code) to use AI. | |||||||||||||||||||||||||||||||||||
| ▲ | 11 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||
| [deleted] | |||||||||||||||||||||||||||||||||||
| ▲ | mlpoknbji 6 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||
I think a more realistic answer is that professional mathematicians have tried to get LLMs to solve their problems and the LLMs have not been able to make any progress. | |||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||
| ▲ | Davidzheng 14 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||
I'm actually not sure what the right attribution method would be. I'd lean towards single line on acknowledgements? Because you can use it for example @ every lemma during brainstorming but it's unclear the right convention is to thank it at every lemma... Anecdotally, I, as a math postdoc, think that GPT 5.2 is much stronger qualitatively than anything else I've used. Its rate of hallucinations is low enough that I don't feel like the default assumption of any solution is that it is trying to hide a mistake somewhere. Compared with Gemini 3 whose failure mode when it can't solve something is always to pretend it has a solution by "lying"/ omitting steps/making up theorems etc... GPT 5.2 usually fails gracefully and when it makes a mistake it more often than not can admit it when pointed out. | |||||||||||||||||||||||||||||||||||