Will any LLM have a context window of at least 1 million characters by the end of 2028?
14
closes 2029
83%
chance

Using characters instead of tokens because token size can be changed, and characters are what humans actually care about. If they advertise a context window in tokens, I'll convert it to characters at the average rate of that tokenizer on representative human text.

Sort by:
PatrickDelaney avatar
Patrick Delaneyis predicting YES at 81%

Let's keep in mind that LLM's are not purely generative in nature and don't have to be based upon a GPT or any pre-determined architecture, they are merely a probability distribution over sequences of words. So as written, this question has a wide interpretation. I would almost advocate for narrowing down the definition further to make it more interesting.

Mira avatar
Mirais predicting YES at 81%

@PatrickDelaney It should probably be "equivalent to GPT-3 on some benchmarks", otherwise a random tree search or markov chain would qualify. (Well, a "large" markov chain)

Maybe the "Evals" repo that was introduced with GPT-4 would be a good one? openai/evals: Evals is a framework for evaluating OpenAI models and an open-source registry of benchmarks. (github.com)

PatrickDelaney avatar
Patrick Delaneybought Ṁ10 of YES

I'm advancing the idea which of course helps my YES bet that this must include private LLM's not publicly disclosed LLM's, so if there is a leak or a news report about any kind of LLM with said context window, it qualifies.

Mira avatar
Mirabought Ṁ100 of YES

Not with transformers since it scales quadratically but I'm sure somebody will train a test model using e.g. hyena operators just to test its limits.

firstuserhere avatar
firstuserhere

@Mira the standard transformer* with dot product attention* scales quadratically

jonsimon avatar
Jon Simonbought Ṁ10 of NO

No, because it would be totally superfluous... No reason to waste compute like that

IsaacKing avatar
Isaac

@jonsimon Humans have an unlimited context window. An AGI probably would too.

jonsimon avatar
Jon Simonis predicting NO at 66%

@IsaacKing would be much easier to use some kind of external knowledge/state store rather than a massive context window

PatrickDelaney avatar
Patrick Delaneyis predicting YES at 81%

@jonsimon No reason? How do you know that a-priori, do you know every industry? Do you know every possible architecture that might come out in 6 years? 1 million characters is 600 to 800 pages. I could imagine existing some odd esoteric application for that. If it was 100,000 or 1 million pages, increasingly less likely. But what about...government/intelligence summarization?

PatrickDelaney avatar
Patrick Delaneyis predicting YES at 81%

@IsaacKing I don't want to argue too harshly because this is all metaphors we're talking about but...humans do not have an unlimited context window analogous to an autoregressive GPT's context window...right? You would have to be able to, "remember," e.g. "tokenize," every conversation you ever had in detail, e.g. including every word to fulfill the condition, "*unlimited* context window." I think you might mean...humans can have a context window that stretches back selectively for years if not decades...something like that? To me, "unlimited," means just a massive billions or trillions of words long corpus including everything one ever heard, read, wrote or spoke. I hardly can remember what I did 10 minutes ago.

firstuserhere avatar
firstuserherebought Ṁ10 of YES

Relevant: "GPT-4 is capable of handling over 25,000 words of text, allowing for use cases like long form content creation, extended conversations, and document search and analysis."

Related markets

By 2027, will it be generally agreed upon that LLM produced text > human text for training LLMs?45%
Will LLMs become a ubiquitous part of everyday life by June 2026?90%
By 2029 end, will it be generally agreed upon that LLM produced text/code > human text/code for training LLMs?70%
By 2025 end, will it be generally agreed upon that LLM produced text/code > human text/code for training LLMs?40%
Will an LLM have an epic name by 2024?86%
Will an LLM improve its own ability along some important metric well beyond the best trained LLMs before 2026?57%
Will LLMs be better than typical white-collar workers on all computer tasks before 2026?42%
Will any LLM have roughly GPT-3-level losses with a context window of at least 50,000 tokens before April of 2024?36%
Will LLM training costs fall 1,000x by 2028?75%
Will we have a popular LLM fine-tuned on people's personal texts by June 1, 2024?33%
Will we have a popular LLM fine-tuned on people's personal texts by June 1, 2026?76%
Will we have a popular LLM fine-tuned on people's personal texts by June 1, 2025?65%
Will LLMs' non-language capabilities be used commercially by the end of 2023?90%
Will there be an LLM which can do fluent conlang translations by EOY 2024?30%
Will a LLM considerably more powerful than GPT-4 come out in 2023?23%
Will LLMs (or similar AI systems) be meaningfully integrated into US public school education by 2025?7%
Will an LLM that someone is trying to shut down stop or avoid that in some way before 2026?26%
Will at least 1 in 8 teams at a FAANG company routinely deploy LLM-written production code by the end of 2023?43%
Will LLM training costs fall 100x by 2028?84%
Will LLM training costs fall 100x by 2028?90%