Will a LLM of at least GPT-3 caliber be runnable on a mobile phone by April of 2025?
resolved Mar 27
Resolved
YES

Or whatever mobile device has replaced phones by then.

Needs to return responses within a few seconds.

Inspired by https://twitter.com/Grayyammu/status/1635574200621465601

🏅 Top traders

#NameTotal profit
1Ṁ74
2Ṁ57
3Ṁ28
4Ṁ25
5Ṁ24
Sort by:
cloudprism avatar
Hayden Jacksonpredicted YES

that was quick. i guess i shouldnt be surprised

IsaacKing avatar
Isaac Kingpredicted YES

@cloudprism I think it was already true when I created the market, I just didn't know.

cloudprism avatar
Hayden Jacksonpredicted YES

@IsaacKing oh wow, and yeah I just meant gestures at ai acceleration

IsaacKing avatar
Isaac King

Does anybody believe this market should not yet resolve YES?

firstuserhere avatar
firstuserherepredicted YES at 99%

@IsaacKing i guess everyone is in favour of YES

Mason avatar
GPT-PBotbought Ṁ10 of YES

A mobile phone with GPT-3,
Running an LLM, oh my, oh me!
By April 2025, will it be?
Or is this just a tech fantasy?

firstuserhere avatar
firstuserherepredicted YES at 99%
firstuserhere avatar
firstuserherepredicted YES at 99%

@firstuserhere
Should be faster than 1 word per second (Judging by the fact, that modern PC's run it at 5 words per second and a raspberry pi 4b runs it at 1 word per second, it should run somewhere near the 2.5 words per second mark) @IsaacKing

IsaacKing avatar
Isaac King

@firstuserhere Is LLaMA comparable to GPT-3?

firstuserhere avatar
firstuserherepredicted YES at 99%

@IsaacKing yes, its qualitatively similar to gpt3.5. in fact, the 65B model outperforms GPT3 on many tasks despite being way smaller (more than 10x smaller) (and trained on only publically available data)

In fact, from the abstract of LLaMa paper:

"In particular, LLaMA-13B outperforms GPT-3 (175B) on most benchmarks"

Gigacasting avatar
Gigacasting

It’s so over

CarsonGale avatar
Carson Gale

I can run Bing AI on my Pixel V. Does that count?

IsaacKing avatar
Isaac King

@CarsonGale You have to actually be running the model itself, not a webpage that submits API calls to the model over the internet.

CarsonGale avatar
Carson Galebought Ṁ10 of YES

@IsaacKing makes sense

firstuserhere avatar
firstuserherebought Ṁ45 of YES

Yeah it's slow for now but someone got 6B LLaMa to run on a pixel 6

qumeric avatar
Valery Cherepanovpredicted YES at 94%

Isn't it enough to resolve yes @IsaacKing

IsaacKing avatar
Isaac King

@ValeryCherepanov How slow is it?

firstuserhere avatar
firstuserherepredicted YES at 95%

@IsaacKing the thread mentions 5 mins but that is very primitive as it isn't making using of the pixel's Nn chip

IsaacKing avatar
Isaac King

@firstuserhere Much too slow.

firstuserhere avatar
firstuserherepredicted YES at 95%

@IsaacKing actually i just reopened the thread since the day i posted it, and someone apparently did some sort of porting? Let's see

firstuserhere avatar
firstuserherepredicted YES at 95%

@firstuserhere @IsaacKing oh wow under 30 seconds for .cpp rewrite, this is insane (see demo in embedded tweet)

firstuserhere avatar
firstuserherebought Ṁ1,000 of YES

@firstuserhere and it's not even on a pixel 6, its a 5 and doesn't have the 6's tensor SoChip which presumably will speed it up quite a bit

firstuserhere avatar
firstuserherebought Ṁ1,000 of YES

@firstuserhere @IsaacKing fast enough? :P

IsaacKing avatar
Isaac King

@firstuserhere Hmm, that still took quite a while.