▲ | JohnKemeny 5 days ago | |
I prefer not to due to privacy concerns. Perhaps you can try yourself? I will say that after checking, I see that the model is set to "Auto", and as mentioned, used almost 8 minutes. The prompt I used was:
It did a lot of thinking, including
And I can see that it visited 13 webpages, including icpc, codeforces, geeksforgeeks, github, tehrantimes, arxiv, facebook, stackoverflow, etc. | ||
▲ | jsnell 5 days ago | parent [-] | |
A terse prompt and expecting a one-shot answer is really not how you'd get an LLM to solve complex problems. I don't know what Deepmind and OpenAI did in this case, but to get an idea of the kind of scaffolding and prompting strategy that one might want, have a look at this paper where some floks used the normal generally available Gemini Pro 2.5 to solve 5/6 of the 2025 IMO problems: https://arxiv.org/pdf/2507.15855 |