
The model has a Elo greater than 1190 on ChatbotArena (or if ChatbotArena is no longer available/updating, achieves GPT 4 (03.14) equivalent or greater performance on both MMLU and MT-Bench)
When running inference in a geographically distributed fashion (the computational hardware is not colocated, and is networked over typical consumer equipment)
on heterogeneous hardware (the computational hardware is varied in type, e.g. different GPU models)
without the act of distributed inference causing the model to require 2 OOM more energy usage (e.g. if doing so Is incredibly lossy and inefficient, it does not count. The burden of proof lies on anyone claiming this clause should be activated)
Note: if (or, when) an edge case is presented, it's applicability to this question will be evaluated in mine + Robert's understanding of the spirit of the question.
🏅 Top traders
# | Name | Total profit |
---|---|---|
1 | Ṁ162 | |
2 | Ṁ134 | |
3 | Ṁ106 | |
4 | Ṁ66 | |
5 | Ṁ65 |