I will try to find estimates of how many people are doing this in order to resolve it, but no guarantees.

I think a lot of people are talking with ChatGPT. And many talk a lot. I do. And when you get into this even if you use it strictly for work, I think at some point you will realize it has a lot of other options and if you will need therapy it's only natural to try talking with chatbot because it is free and super simple.
And a lot of people kinda want/need therapy (something like 10%)? So I think that it is very likely that it would happen if things will go as they do now.
But it's possible that therapy or coach chatbots are not going to be popular and instead, we will have ultra-general chatbots which can act however you want. In this case, it may be very hard to estimate numbers.
Also of course we have x-risk and other serious catastrophic scenarios. I think they are not very likely but still worth to consider because 5 years is a lot of time. Being somewhat conservative, I think it has a 5%+ probability.

@ValeryCherepanov yes most of my allocation for how this 0.1% claim fails to come true are related to legislation, war, etc. With normal development it seems like a lock. I just wish it could be resolved instantaneously.



I expect this would count: https://marginalrevolution.com/marginalrevolution/2023/03/how-will-ai-transform-childhood.html
Does "at least monthly" imply "on two consecutive months" or just perhaps "said they talked to an AI therapist in the last 30 days?"
also how widely do we interpret "coach"? does a careers advisor count? a literal sports coach?

I made a market on whether Scott will say his 5% prediction was a mistake:


@Conflux No, I think he is saying he will take it as evidence if like 5 percent of people he knows or reads are talking about it. Since there aren’t reliable numbers published.

@Conflux Though I except that will change before 2028. These companies are going to be a lot more transparent if they hope to avoid crippling regulation.
@Conflux I think Scott's estimate here is so blatantly wrong there is a chance it is a mistake and his actual prediction is much higher than 5%.

@AlexPower I don’t think that number represents his base rate. It definitely does not.


No, I think he is saying he will take it as evidence if like 5 percent of people he knows or reads are talking about it.
If he meant this, he definitely could have worded it in a less confusing way.
@bingeworthy He definitely didn't mean it as the base rate. All of the other questions also have a bolded percentage at the end that is a prediction for that question.

@Gabrielle Actually, "this number" is 350,000 and 5% is the base rate. Had to read all of them to pick up on it. This is way too low. I mean, if Weight Watchers launches a chatbot, which seems likely the second you think of it, then this is an immediate YES. Don't think of this question as being about something growing organically to 350,000, there are plenty of such "therapists or coaches" that could increase their productivity by having a chatbot supplement their personal care.
@BTE I suspect that Alexander is imagining an FDA approved therapist bot, or something like that, which would require regulatory approval. That would explain the relatively low rate. The fact that it just says "therapist or coach" complicates it, so I definitely agree that it could just be a mistake.
@Gabrielle I would believe 5% as a prediction for "FDA approval". Which is very much not what the post says.

@Gabrielle Chatbots are explicitly exempted from FDA review.

@Gabrielle This is what I do for a living actually. Even chatbots like Babylon Health have no regulatory oversight.























Related markets





Related markets




