Is it possible to align scaffolded LLMs with human values?
5
110Ṁ712100
70%
chance
1H
6H
1D
1W
1M
ALL
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
People are also trading
Related questions
are LLMs easy to align because unsupervised learning imbues them with an ontology where human values are easy to express
33% chance
By 2025 end, will it be generally agreed upon that LLM produced text/code > human text/code for training LLMs?
11% chance
Will "LLMs for Alignment Research: a safety priority?" make the top fifty posts in LessWrong's 2024 Annual Review?
13% chance
Will an LLM improve its own ability along some important metric well beyond the best trained LLMs before 2026?
50% chance
Will relaxed adversarial training be used in practice for LLM alignment or auditing before 2028?
79% chance
EOY 2025: Will open LLMs perform at least as well as 50 Elo below closed-source LLMs on coding?
30% chance
Will LLMs be the best reasoning models on these dates?
Will there be an LLM which scores above what a human can do in 2 hours on METR's eval suite before 2026?
67% chance
By 2027, will it be generally agreed upon that LLM produced text > human text for training LLMs?
62% chance
Will LLMs be used for academic peer review by 2030?
71% chance