The SWE-bench is a benchmark developed to evaluate if language models can resolve real-world GitHub issues. The leaderboard showcases various models and their performances in terms of the percentage of SWE-bench instances they resolved. Each instance in the SWE-bench represents a GitHub issue. The leaderboard is categorized into two main sections: Unassisted and Assisted.
Unassisted: In this category, models are evaluated without any assistance. This means that they don't have the advantage of the "oracle" retrieval setting where the correct files to edit are directly given to them.
This question is only about the Unassisted category of this benchmark.
http://www.swebench.com/#
Current SOTA is <2%
The prediction market will resolve based on the SWE-bench leaderboard standings as of 11th October 2024.
In the extremely unlikely case that the number would fit in two intervals, the lowest will be chosen.
@MikhailDoroshenko why does that count given that it wasn't submitted to the official leaderboard or verified?
@Fay42 It counts because all their solution are available on GitHub and nobody objected their claim.
I will be using full dataset, because the question was about full dataset.
@Sss19971997 I am a bit in a pickle. I am happy that you bet on my market, but I am not sure I understand your bets. Like which resource would you name as a reputable source if you believe swebench.com is fake?
@EliLifland @vluzko Wanna bet against me? I put a huge limit order for YES on 0-15%
I dont believe the leaderboard. Those were fake