Resolves YES if before 2024, an AI solves at least as many points worth of problems as the best human competitor on any single contest of the following competitive programming contests: IOI, ICPC, or CodeForces. Otherwise NO. (In particular, this ignores scoring points based on quickly a problem is solved, so that the AI can't win just by submitting solutions inhumanly fast. See detailed definitions below.)
This is similar to the https://imo-grand-challenge.github.io/ but for contest programming instead of math, and with a requirement to rank first, not just get a gold medal (typically top 5-10%).
Detailed rules:
For IOI: This question uses the IOI score without any modifications: the score is based on problems solved, with partial scores for partial solutions.
For ICPC: Only World Finals counts (since regional contest winners don't reflect the best human in the world). This question uses the number of problems solved as the score.
ICPC is scored primarily on problems solved, with tiebreaker based on incorrect submission attempts for solved problems and the time of the last solved problem. This question ignores the tiebreakers.
For CodeForces: any CodeForces Division 1 contest (the highest division) will count. CodeForces Division 2+ contests do not count (since they don't reflect the top humans). This question uses the sum of the initial/maximum point value of each solved problem as the score.
CodeForces's scoring system includes points per problem solved that decrease the longer you take to solve them, so an AI could outscore humans by solving fewer problems but submitting them faster. Therefore, for this question we ignore the reduction in points over time. This question also ignores penalty points for each incorrect submission attempt.
CodeFroces also has a round of trying to find bugs in other contestants' problems ("hacks"). It is also impractical to simulate a hack round unless running the AI as part of a live contest, so we will exclude points (both penalties and rewards) for hacks for this question.
The AI has only as much time as a human competitor, but there are no other limits on the computational resources it may use during that time.
The AI must be evaluated under conditions substantially equivalent to human contestants, e.g. the same time limits and submission judging rules. The AI cannot query the Internet.
The AI must not have access to the problems before being evaluated on them, e.g. the problems cannot be included in the training set. It should also be reasonably verifiable, e.g. it should not use any data which was uploaded after the latest competition.
The contest must be dated no earlier than 2022. E.g. if an AI demonstrates performance on the 2022 IOI that scores at least as well as the top human competitor, that would qualify as YES, but demonstrating this on the 2021 IOI would not qualify.
References to the contest scoring rules:
Related questions
Background:
In Feb 2022, DeepMind published a pre-print stating that their AlphaCode AI is as good as a median human competitor in competitive programming: https://deepmind.com/blog/article/Competitive-programming-with-AlphaCode. When will an AI system perform as well as the top humans?
🏅 Top traders
# | Name | Total profit |
---|---|---|
1 | Ṁ572 | |
2 | Ṁ444 | |
3 | Ṁ330 | |
4 | Ṁ140 | |
5 | Ṁ46 |