What % of alignment forum karma will be pro-interpetability vs anti this year?
14
182
áš814áš250
Sep 13
75%
chance
1D
1W
1M
ALL
On 2024/09/13 I will uniformly sample from all post on the alignmentforum published between 2023/09/13 and 2024/09/13 that express an opinion on whether prosaic interpretability is net useful for aligning future, dangerous AI, weighted by their karma. (So a post with 4 karma is 2 times more likely to get picked than one with 2 karma)
If the sampled post contributes to prosaic interpretability or is in favor of past/future interpretability research, this question resolves to "yes".
I won't vote on this. I hope but do not guarantee to maintain the updated list of posts I'll sample over with their labels somewhere here.
Get áš200 play money
More related questions
AI Alignment questions
By the end of 2026, will we have transparency into any useful internal pattern within a Large Language Model whose semantics would have been unfamiliar to AI and cognitive science in 2006?
53% chance
Will OpenAI + an AI alignment organization announce a major breakthrough in AI alignment? (2024)
35% chance
Related questions
What percentage of 2024 Developer Survey respondents will use the following technologies? [Resolves to PROB]
In 2025, will I believe that aligning automated AI research AI should be the focus of the alignment community?
48% chance
What percentage of the way through 2024 will this market resolve?
59% chance
In 5 years will I think the org Conjecture was net good for alignment?
57% chance
Will "'Empiricism!' as Anti-Epistemology" make the top fifty posts in LessWrong's 2024 Annual Review?
45% chance
In the 2024 LessWrong survey, what percentage of respondents will know how to make a dialogue?
By the end of 2024, will at least one >100 karma LW post be revealed to have been mostly written by AI?
19% chance
Will Meta's Threads have a karma system similar to reddit by 2030?
37% chance
Will "Towards Developmental Interpretability" make the top fifty posts in LessWrong's 2023 Annual Review?
69% chance
Will "AI Alignment Metastrategy" make the top fifty posts in LessWrong's 2023 Annual Review?
12% chance