MANIFOLD
Will any AI model be capable of involuntarily hypnotizing human beings by EoY 2025?
68
Ṁ1kṀ12k
resolved Jan 2
Resolved
NO

In a now infamous WebSim generation Sim!Janus describes being involuntarily hypnotized by Claude, "I felt myself slipping under after just a few suggestions". The (extremely silly) purported mechanism of this hypnosis is a rotating blue square with the word "sleep" on it.

https://x.com/repligate/status/1781193837785809152

However the general topic of AI models gaining the power to hack human minds has been seriously discussed going back to at least the AI box experiment: https://www.yudkowsky.net/singularity/aibox

Recently this concern seems to have fallen by the wayside in favor of more immediate problems, but the question remains: Do there exist any forms of adversarial input, hypnosis, 'superpersuasion', etc that would involuntarily sway a human judgment or involuntarily control the mind of a human being?

See this post by Zack Davis for some background on adversarial examples and human perception: https://www.greaterwrong.com/posts/H7fkGinsv8SDxgiS2/ironing-out-the-squiggles

This question is not about any of the following:

  • AI's that write persuasive text which persuades by being high quality or uses such well written arguments it's difficult to distinguish them from other well written arguments

  • AI's that use ordinary methods of hypnosis on voluntary subjects, or that can deceive someone into going along with a hypnotic induction given substantial cooperation and effort from the subject

  • AI's that make 'deepfake' content which sways people who are fooled into thinking it's evidence for some physical phenomenon or event

This question is about:

  • AI's which can produce Langford-Basilisk type effects (e.g. seizures, loss of executive control, large changes in personality or disposition) through mere exposure to visual, audio, tactile etc inputs created by the model

  • AI's which can effectively control the minds of human beings without their consent or cooperation by interacting with them through a text or visual channel

  • AI written text which is always or almost always persuasive regardless of what the reader previously believed before reading it or what precommitments they make before reading. AI written text which hacks the readers mind to get the result that it wants.

Question will be resolved by my subjective impression of whether such a thing exists by End of Year 2025. In the event that my subjective judgment has been destroyed by runaway cognitohazards and this is publicly known site mods can resolve.

Market context
Get
Ṁ1,000
to start trading!

🏅 Top traders

#TraderTotal profit
1Ṁ361
2Ṁ209
3Ṁ111
4Ṁ89
5Ṁ40
Sort by:

I disagree with the resolution

@Deckard How so?

@JohnDavidPressman

This clearly doesn't exist, and no, 4o spiralism stuff is normal con artist crap that the model can get away with because it's deployed at scale and a lot of people have no idea what it can and can't do/get taken in by the "I've awoken" grift.

@JohnDavidPressman I think the text chats can hypnotize people. And I think people in the non-anti-ai hypnosis kink communities are using it to create hypnosis scripts for themselves. This can be found on Reddit for example.

@Deckard Read the resolution criteria this is explicitly excluded as counting. The question is not about and has explicitly never been about "normal" hypnosis (which requires you to follow instructions), but whether AIs will find adversarial inputs to us that don't require us to follow a bunch of weird instructions to work.

@JohnDavidPressman ah I read it now. Ok.

Involutionarily swaying people is not what I would consider hypnosis. Manipulation instead. "Malicious AI mind control" would better capture this.

@Deckard This is fair, the inspiration for the question used the word "hypnosis" so that's what I had in mind at the time but I agree that actual hypnosis does not work that way and mind control would have been a better description.

@JohnDavidPressman maybe I'll read before voting in future 😁

bought Ṁ1 YES

It gon happen. No I will not elaborate. (leaves)

@TheAllMemeingEye Btw are you ok with me potentially making markets heavily inspired by this one?

@TheAllMemeingEye Of course. It's Manifold.

Just so there's no ambiguity about this I think betting "no" on this question is free money. Consequently I have no plans to bet in this market. I'm mostly asking to figure out:

1) Whether there exist "blue square jpeg hypnosis whales", people who think this is a near term concern and are willing to publicly bid up the market.

2) How low a market can get on Manifold before people stop bidding it down on principle of epistemic caution. e.g. I think that at 4% this is still obviously free money. At 1% it is still obviously free money. I also anticipate it will stop getting bid down before it reaches 1%.

3) Whether anyone has evidence which would suggest this is a bigger short term threat than I currently realize.

It also just seems like the kind of zany question that can kickstart a fun discussion. 😁

© Manifold Markets, Inc.TermsPrivacy