Will a Mamba 7b model trained on 2 trillion tokens outperform Llama2-13B
19
342
Ṁ638Ṁ1k
Jul 1
69%
chance
1D
1W
1M
ALL
Question will resolve positive if someone trains a Mamba (https://twitter.com/tri_dao/status/1731728602230890895) language model with <=7.5billion parameters on <=2 trillion tokens that outperforms Llama2-13B on the huggingface open llm leaderboard (https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Get Ṁ600 play money
Related questions
Related questions
Will the 400B+ open source Llama 3 model rank higher than GPT-4-Turbo-2024-04-09 on the lmsys leaderboard?
60% chance
Is it true that "Llama2-70B's Elo rating will be ~3000"?
2% chance
Will Llama 3 400B be better than GPT-4?
59% chance
Will a Mamba-based LLM of GPT 3.5 quality or greater be open sourced in 2024?
51% chance
Will Mamba be the de-facto paradigm for LLMs over transformers by 2025?
25% chance
Will LLaMA-3 be on par with or better than GPT-4?
81% chance
Will a Language Model under 10B parameters play chess at Grandmaster level by 2050?
39% chance
Will the best LLM in 2027 have <1 trillion parameters?
29% chance
Will any LLM outrank GPT-4 by 150 Elo in LMSYS chatbot arena before 2025?
21% chance
Any Gemini LLM will achieve a higher rating than all OpenAI's GPT-4 models on Chatbot Arena Leaderboard by Jan 1st 2025?
44% chance