Is the orthogonality thesis correct?
12
138
230
2300
88%
chance

Full question:

The intelligence of an AI and its goals are not necessarily related (i.e., the orthogonality thesis). This does not exclude correlation in the implementation of AI, but only a lack of any necessary association in the architecture itself.

One of the questions from https://jacyanthis.com/big-questions.

Resolves according to my judgement of whether the criteria have been met, taking into account clarifications from @JacyAnthis, who made those predictions. (The goal is that they'd feel comfortable betting to their credance in this market, so I want the resolution criteria to match their intention.)

Get Ṁ200 play money
Sort by:

Current AIs like ChatGPT gain their capabilities by mimicking humans, and human capabilities are optimized to support human concerns. So I think the orthogonality thesis is nearly complete false for ChatGPT-style models.

However, I think mimicking humans isn't the only way to make AI, and that the orthogonality thesis will be much more true for other methods like consequentialism, whose goals are only limited by instrumental convergence.

predicts YES
Comment hidden

@tailcalled This seems to be an argument against the strong orthogonality thesis, which roughly holds that: for any given environment that produces an agent, the agent's intelligence and values are orthogonal. But it sounds like this market is talking about the weak orthogonality thesis, which is more accepted. The weak orthogonality thesis roughly says that: in principle, almost any level of intelligence could be compatible with almost any goal.

The paper linked to in the market description also provides a definition of the orthogonality thesis along these lines:

Intelligence and final goals are orthogonal axes along which possible agents can freely

vary. In other words, more or less any level of intelligence could in principle be

combined with more or less any final goal.

bought Ṁ2 of YES

This is entirely correct, for example liberals support government then oppose the functionaries through which it has any power whatsoever (cops). Not defending cops just showing that linear, first order reasoning creates contradiction regardless of the strength of the reasoning and these contradictions eventually destroy any system they attempt to build.

bought Ṁ69 of YES

@MarkIngraham That's... not what the orthogonality thesis is.

predicts YES

@IsaacKing it shows the intelligence of ai and its goals are unrelated, ie, it can attack its own functionaries while pursuing first order goals. It's like a ant colony eating itself.