Remix.run Logo
jmyeet 4 hours ago

This is a known problem and an active area of research [1][2][3][4].

[1]: https://arxiv.org/html/2505.15623v1

[2]: https://medium.com/@adnanmasood/why-large-language-models-st...

[3]: https://www.reachcapital.com/resources/thought-leadership/wh...

[4]: https://mathoverflow.net/questions/502120/examples-for-the-u...

simianwords 3 hours ago | parent [-]

the research does't capture the fact that LLM's can easily multiply these results. I mean it literally won gold in IMO, Putnam.

Take 10,000 such multiplications. I'm sure not even a single one would be incorrect with GPT 5.2 (thinking). Want a wager?