Will DPO or an Explicitly DPO-based Technique be Used to Train a Public Frontier Lab LLM Before Jan 1 2025?
28
1kṀ2166Jan 2
84%
chance
1H
6H
1D
1W
1M
ALL
An explicitly DPO-based technique is one that cites DPO as seed material for its creation.
Frontier labs currently include: OpenAI, DeepMind, Anthropic, Google. I will modify this description if this changes (e.g. if Meta releases a SOTA LLM.)
Public simply means that it has been announced or otherwise discovered that this DPO LLM has been trained.
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
People are also trading
Related questions
By the end of June 2025, will closed-source LLMs increase access to pandemic agents?
47% chance
Will a LLM trained with FP4 have frontier-level performance before 2028?
31% chance
Which next-gen frontier LLMs will be released before GPT-5? (2025)
Will a lab train a >=1e26 FLOP state space model before the end of 2025?
15% chance
Will any widely used LLM be pre-trained with abstract synthetic data before 2030?
72% chance
Will researchers extract a novel program from the weights of an LLM into a Procedural/OO programming language by 2026?
27% chance
Will one of the major LLMs be capable of continual lifelong learning (learning from inference runs) by EOY 2025?
26% chance
Will relaxed adversarial training be used in practice for LLM alignment or auditing before 2028?
79% chance
Will there be a state-of-the-art LLM that is NOT based on next raw token prediction before 2029?
50% chance
Will we get a new LLM paradigm by EOY?
32% chance