Elon Musk has been very explicit in promising a robotaxi launch in Austin in June with unsupervised full self-driving (FSD). We'll give him some leeway on the timing and say this counts as a YES if it happens by the end of August.
As of April 2025, Tesla seems to be testing this with employees and with supervised FSD and doubling down on the public Austin launch.
PS: A big monkey wrench no one anticipated when we created this market is how to treat the passenger-seat safety monitors. See FAQ9 for how we're trying to handle that in a principled way. Tesla is very polarizing and I know it's "obvious" to one side that safety monitors = "supervised" and that it's equally obvious to the other side that the driver's seat being empty is what matters. I can't emphasize enough how not obvious any of this is. At least so far, speaking now in August 2025.
FAQ
1. Does it have to be a public launch?
Yes, but we won't quibble about waitlists. As long as even 10 non-handpicked members of the public have used the service by the end of August, that's a YES. Also if there's a waitlist, anyone has to be able to get on it and there has to be intent to scale up. In other words, Tesla robotaxis have to be actually becoming a thing, with summer 2025 as when it started.
If it's invite-only and Tesla is hand-picking people, that's not a public launch. If it's viral-style invites with exponential growth from the start, that's likely to be within the spirit of a public launch.
A potential litmus test is whether serious journalists and Tesla haters end up able to try the service.
UPDATE: We're deeming this to be satisfied.
2. What if there's a human backup driver in the driver's seat?
This importantly does not count. That's supervised FSD.
3. But what if the backup driver never actually intervenes?
Compare to Waymo, which goes millions of miles between [injury-causing] incidents. If there's a backup driver we're going to presume that it's because interventions are still needed, even if rarely.
4. What if it's only available for certain fixed routes?
That would resolve NO. It has to be available on unrestricted public roads [restrictions like no highways is ok] and you have to be able to choose an arbitrary destination. I.e., it has to count as a taxi service.
5. What if it's only available in a certain neighborhood?
This we'll allow. It just has to be a big enough neighborhood that it makes sense to use a taxi. Basically anything that isn't a drastic restriction of the environment.
6. What if they drop the robotaxi part but roll out unsupervised FSD to Tesla owners?
This is unlikely but if this were level 4+ autonomy where you could send your car by itself to pick up a friend, we'd call that a YES per the spirit of the question.
7. What about level 3 autonomy?
Level 3 means you don't have to actively supervise the driving (like you can read a book in the driver's seat) as long as you're available to immediately take over when the car beeps at you. This would be tantalizingly close and a very big deal but is ultimately a NO. My reason to be picky about this is that a big part of the spirit of the question is whether Tesla will catch up to Waymo, technologically if not in scale at first.
8. What about tele-operation?
The short answer is that that's not level 4 autonomy so that would resolve NO for this market. This is a common misconception about Waymo's phone-a-human feature. It's not remotely (ha) like a human with a VR headset steering and braking. If that ever happened it would count as a disengagement and have to be reported. See Waymo's blog post with examples and screencaps of the cars needing remote assistance.
To get technical about the boundary between a remote human giving guidance to the car vs remotely operating it, grep "remote assistance" in Waymo's advice letter filed with the California Public Utilities Commission last month. Excerpt:
The Waymo AV [autonomous vehicle] sometimes reaches out to Waymo Remote Assistance for additional information to contextualize its environment. The Waymo Remote Assistance team supports the Waymo AV with information and suggestions [...] Assistance is designed to be provided quickly - in a mater of seconds - to help get the Waymo AV on its way with minimal delay. For a majority of requests that the Waymo AV makes during everyday driving, the Waymo AV is able to proceed driving autonomously on its own. In very limited circumstances such as to facilitate movement of the AV out of a freeway lane onto an adjacent shoulder, if possible, our Event Response agents are able to remotely move the Waymo AV under strict parameters, including at a very low speed over a very short distance.
Tentatively, Tesla needs to meet the bar for autonomy that Waymo has set. But if there are edge cases where Tesla is close enough in spirit, we can debate that in the comments.
9. What about human safety monitors in the passenger seat?
Oh geez, it's like Elon Musk is trolling us to maximize the ambiguity of these market resolutions. Tentatively (we'll keep discussing in the comments) my verdict on this question depends on whether the human safety monitor has to be eyes-on-the-road the whole time with their finger on a kill switch or emergency brake. If so, I believe that's still level 2 autonomy. Or sub-4 in any case.
See also FAQ3 for why this matters even if a kill switch is never actually used. We need there not only to be no actual disengagements but no counterfactual disengagements. Like imagine that these robotaxis would totally mow down a kid who ran into the road. That would mean a safety monitor with an emergency brake is necessary, even if no kids happen to jump in front of any robotaxis before this market closes. Waymo, per the definition of level 4 autonomy, does not have that kind of supervised self-driving.
10. Will we ultimately trust Tesla if it reports it's genuinely level 4?
I want to avoid this since I don't think Tesla has exactly earned our trust on this. I believe the truth will come out if we wait long enough, so that's what I'll be inclined to do. If the truth seems impossible for us to ascertain, we can consider resolve-to-PROB.
11. Will we trust government certification that it's level 4?
Yes, I think this is the right standard. Elon Musk said on 2025-07-09 that Tesla was waiting on regulatory approval for robotaxis in California and expected to launch in the Bay Area "in a month or two". I'm not sure what such approval implies about autonomy level but I expect it to be evidence in favor. (And if it starts to look like Musk was bullshitting, that would be evidence against.)
12. What if it's still ambiguous on August 31?
Then we'll extend the market close. The deadline for Tesla to meet the criteria for a launch is August 31 regardless. We just may need more time to determine, in retrospect, whether it counted by then. I suspect that with enough hindsight the ambiguity will resolve. Note in particular FAQ1 which says that Tesla robotaxis have to be becoming a thing (what "a thing" is is TBD but something about ubiquity and availability) with summer 2025 as when it started. Basically, we may need to look back on summer 2025 and decide whether that was a controlled demo, done before they actually had level 4 autonomy, or whether they had it and just were scaling up slowing and cautiously at first.
13. If safety monitors are still present, say, a year later, is there any way for this to resolve YES?
No, that's well past the point of presuming that Tesla had not achieved level 4 autonomy in summer 2025.
14. What if they ditch the safety monitors after August 31st but tele-operation is still a question mark?
We'll also need transparency about tele-operation and disengagements. If that doesn't happen soon after August 31 (definition of "soon" to be determined) then that too is a presumed NO.
Ask more clarifying questions! I'll be super transparent about my thinking and will make sure the resolution is fair if I have a conflict of interest due to my position in this market.
[Ignore any auto-generated clarifications below this line. I'll add to the FAQ as needed.]
People are also trading
From the Q3 Investors call https://seekingalpha.com/article/4832142-tesla-inc-tsla-q3-2025-earnings-call-transcript
Musk: "we are expecting to have no safety drivers in at least large parts of Austin by the end of this year ... we expect to be operating in Nevada and Florida and Arizona by the end of the year."
and "Cybercab starts production in Q2 next year"
Ashok Elluswamy (AI lead): "We continue to operate our fleet in Austin without anyone in the driver seat, and we have covered more than 0.25 million miles with that. And then in the Bay Area, where we still have a person in the driver seat because of the regulations, we crossed more than 1 million miles."
@MarkosGiannopoulos Article by Ashok on how Tesla is solving autonomy https://x.com/aelluswamy/status/1981644831790379245
and his talk at the International Conference on Computer Vision (ICCV). https://x.com/SawyerMerritt/status/1981461127046258981
Improvements in Texas and California appear to have been minimal in the last month, but Tesla is now testing in the Vegas Loop tunnel. https://x.com/elonmusk/status/1979511562655133710

@MarkosGiannopoulos Ah, thanks! Am I being too cynical to think of this as, overall, maybe a negative update? Focusing on one specific tunnel seems consistent with the bear case, especially when Musk seems to be admitting they're not at level 4 there.
(Should we have yet another market for that "driverless in this tunnel in a month or two" claim?)
I also wrote about the non-update that was FSD version 14:
https://agifriday.substack.com/p/delusion
Definitely let me know if you're spotting anything unfair in these assessments or if there's anything else I've missed! I'm genuinely confused about how to square 5 months of driverlessness citywide in Austin with needing another month or two before being driverless in that Vegas tunnel.
@dreev a) In regard to "focusing on a tunnel": This is one of many areas where Tesla is doing Robotaxi testing. See this tracker https://x.com/jonathanwstokes/status/1978159082419708137 , where multiple cities are being mentioned with related jobs listingσ, and also sightings of test cars on the road
b) Musk's comment for "a month or two" is probably also related to regulatory approval. It's not just up to Tesla to feel safe about removing the safety driver. For the technical aspect, I'm unsure whether a tunnel with limited points of reference is easier to navigate than a road.
c) On FSD v14 being a "non-update": It is still in limited release (you will notice the low number of reports on the crowdsourced data), but people are reporting an improved experience. Some samples
https://x.com/hletes/status/1979745495233753420
https://x.com/SawyerMerritt/status/1979218249309888925
https://www.youtube.com/watch?v=nXcwDzcyTuY
And if one does not trust videos from Tesla fans, FSD is already performing at a good level, also internationally. See v13 reviews from Japan https://www.teslarati.com/tesla-fsd-gets-first-rave-reviews-media-outlets-japan/ and Australia https://www.news.com.au/technology/motoring/on-the-road/why-tesla-fsd-is-a-nuclear-moment/news-story/5b69240129b9f167b7c7fe70c19bd94f
If you're looking for a bear argument, consider the relative silence on the Tesla Robotaxi account, with no sign of adding more cars in Austin.
The Q3 earnings report is expected to come out this week, and hopefully, it will include some Robotaxi numbers.
@dreev On the overall topic of "is Tesla faking it?", I don't think they would be spending billions to develop a car without a steering wheel unless they were serious about FSD https://x.com/tslaming/status/1979775857452540262

@MarkosGiannopoulos That sounds like a positive update! To be clear, by cheating I just mean not at level 4 yet. My strongest prediction has been that they won't get there till they relent on lidar. But given enough time, I'm sure it can be done with just cameras.
(And of course the big question for this market is whether that happened already a couple months ago.)
@dreev Btw, the FSD tracker that you quoted on your article shows 0 (zero) critical disengagements (DriveNoCriticalDE=100%). Of course, this is with a small number of entries. Yet, you write "The crowdsourced data on miles between critical disengagements also shows little change."

@MarkosGiannopoulos Ah, thank you! I did try filtering for version 14 when I wrote that and I think I must've just been looking at the "miles to critical disengagement" number. That number actually goes down when you filter for version 14, but an asterisk seemingly indicating "no disengagements" appears, which would be consistent with 100% of drives with no critical disengagements.
But then what's up with this breakdown for version 14:

It looks like this is all based on exactly 2 people, one in Illinois and one in British Columbia. I guess version 14 is not rolled out very widely yet. And I can't tell what's up with that 0% and 100% respectively for fraction of drives without critical disengagements. I'm struggling to get a consistent picture of what this data is telling us exactly.
Update: Try jumping to page 9 of the teslafsdtracker.com page:

I guess I still can't parse this. For example, 40% of drives for version 14.1.2 have no disengagements at all, critical or otherwise, but 0% have no critical disengagements? What?
For a minute I thought this could all be explained if they screwed up their double negatives and "%DrivesNoCriticalDE" means "%DrivesWITHCriticalDE". But I think these numbers are self-contradictory regardless. :/
PS: I contacted the person who runs TeslaFSDTracker. Hopefully we can get to the bottom of this! (And hopefully they get more data soon.)
@dreev If you select 13.2.5 there is just 22 miles with no disengagements from one entry, the miles to critical disengagement event is 22. So if there are no disengagements are they reporting longest distance drive rather than total miles driven?
No ignore this
Selecting 14.1.1 I see 662 miles no critical DE but miles to critical DE as 49 with for no DE.
In the box by state province it shows
Illinois 49 miles
British Columbia 613 miles.
So the miles per DE is perhaps coming from Illinois/US and excluding British Columbia/Canada
Seems to work with all v14: 206 miles in Illinois and miles to critical DE 206*. The Star* means total miles driven with no critical DEs. Without the *no DE marker they divide total miles in US by number of critical DEs in US (but they cannot divide by zero).
Update! I talked to someone at TeslaFSDTracker.com and solved the mystery. It's true that there are exactly 2 people contributing data for version 14 so far (it's not widely deployed yet) and one of those people is using an older reporting system that doesn't distinguish critical vs noncritical disengagements.
To make this easier to talk about -- not confusing ourselves by failing to have a lack of double negatives -- let's think of "drives with no critical disengagements" as safe drives and "drives with no disengagements at all" as perfect drives.
So when a row of data says 40% of drives were perfect and 0% were safe, that's indeed impossible. Perfect is a subset of safe! But that "0% safe" is really saying that safe vs perfect wasn't distinguished. So all we know is is that 40% of drives were perfect and at least that many were safe.
So @MarkosGiannopoulos was correct from the start: We have no proof of any critical disengagements on version 14 so far. But the data is even more limited than we thought: a total of 1454 miles from a single person.
Still, that's at least weak evidence of an improvement in version 14!
PS: I may have spoken too soon. There's a column in the data right now with 45.2% perfect drives and 100% safe drives, which would imply that's from the person with the modern reporting system, distinguishing critical and noncritical disengagements. But on the same row we see 10 miles between disengagements and only 2 miles between critical disengagements. I don't know how to reconcile that with the above supposed solution to the mystery.
@dreev Re>"We have no proof of any critical disengagements on version 14 so far. But the data is even more limited than we thought: a total of 1454 miles from a single person."
I am saying the data is much more limited than that. While there is 1817 miles with v14, most of these miles are in Canada which are not used. There is only 228 miles in the USA with no critical or other disengagement event, hence the 'Miles to critical DE *No DE' showing as 228.
While no critical DE or any DE is good this is very limited data. v13.2.9 was running at 395 miles per critical DE v13.2.7 was at 644 miles per critical disengagement. At that level you wouldn't expect a critical disengagement event yet. The miles to DE on 13.2.7 was 84 so maybe you would expect at least one in 228 miles but if the testers are very limited with v14 and happen to be experienced tested who don't unnecessarily disengage much then maybe not and this means little.
@dreev Re "But on the same row we see 10 miles between disengagements and only 2 miles between critical disengagements. I don't know how to reconcile that with the above supposed solution to the mystery."
For me v14.1.2 row shows 253 miles, %drivesnocriticalDE as 100%, 10 miles to DE and 5 miles to critical DE.
My explanation is that some of the 253 were in Illinois and had no critical disengagements. If there were no miles in USA then %drivesnocriticalDE would show 0% like the v14.1.3 row. There were 25 DE all in Canada leading to calculation 253/25=10 miles to DE. One driver in Canada got new software showing critical DEs. If this person had done 2 miles and got a critical disengagement event when you saw the data but has now done 5 miles, this would explain the data wouldn't it?
btw I think that in Illinois there are 3 testers.
One with 157 miles
One with 67 miles
One with 4 miles
total 228 miles in Illinois all with no critical disengagements
In Canada, I would guess 5 testers:
There is one with 613 miles (over 400 mile category)
3 or 4 in the 200-400 miles range with 887 miles in total
1 in Ontario with 95 miles
(157+67+4=228 + 613+887+95 = 1823 rounding causes difference to 1824 total miles.
Now I am slightly confused again as there does not seem to be a tester in Canada with 5 or 10 miles. Perhaps there is a difference in how up to date the table page 9 is compared to the main page 1 display?
Alternately we could possibly imagine the Ontario tester tried some wild things got 19 critical DE in 38 miles to have 2 miles per DE then that tester settled down to some more normal testing and increased miles to 95 without further critical DEs so 95/19=5 miles per critical DE.
@dreev
"We'll give him some leeway on the timing and say this counts as a YES if it happens by the end of August."
Was close date of 2 Sep 2026 meant to be 2 Sept 2025? 2 days to assess latest situation might be a bit short but a year seems excessive.
Have there been any taxi rides without a safety monitor? There was, to my knowledge, only one car delivery to a customer but that doesn't seem like a robotaxi service.
Can this resolve no? If not, what are we waiting for and how long is that going to take?
@ChristopherRandles Yeah, I added a year as an upper bound. Hopefully won't take that long. But could take till April ish, I think, when new Texas laws go fully into effect and which may yield more transparency about disengagements.
@dreev
"this question depends on whether the human safety monitor has to be eyes-on-the-road the whole time with their finger on a kill switch or emergency brake. If so, I believe that's still level 2 autonomy. Or sub-4 in any case.
...
We need there not only to be no actual disengagements but no counterfactual disengagements."
We have video of at least one disengagement UPS truck trying to reverse into parking space.
3 accidents in July well they might be other drivers faults but it seems a lot of such incidents from small amount of driving, but we don't know for sure. Driving wrong way down street and excessive braking near police cars was also reported even if we do not know whether this led to disengagement(s).
Seems clearly sub level 4 from just one disengagement? Or, if there was a new software version (minor decimal or full version?) after this disengagement that didn't have any disengagements before or after 31 Aug and it went live 31 August or earlier perhaps that then has a chance to qualify? Do we have any version number with date deployed info to help clarify?
I am trying to see if we need to wait for disengagement data or if we already have enough to say "safety monitor has to be eyes-on-the-road the whole time" and/or can resolve because of actual disengagement.
@dreev The new law does not require reporting of disengagements, so I hope that's not what you're waiting for.
@WrongoPhD Hmm, yeah, I'm fuzzy on exactly what's going to be required in Texas by when. California does have such requirements, right? So if they're operating there without safety drivers then there could still be a narrow path to a YES resolution here. That will also mean answering at least all of following in the negative:
Does the touchscreen disengagement with the UPS truck count as supervision?
Was the passenger door button being used as a real-time kill switch (actually or counterfactually) as of August 31?
Was driving-speed tele-operation in use? (Human assistance after an autonomus MRM is ok.)
What's going to be agonizing is if Tesla keeps finding excuses to keep the safety monitors, not get licensed in California, and otherwise act in a way consistent with being sub-level-4 while still claiming to be level 4. See FAQ10 for more on this.
@dreev They are operating with safety driver in drivers seat in California and don't yet have a licence for fully driverless there. 31 Aug has past so I don't see any narrow path here.
Does the touchscreen disengagement with the UPS truck count as supervision?
How can you answer that as anything other than yes? The safety monitor saw a situation developing and reacted to it by pressing the stop in lane button on the main screen. If that is not both supervision and intervention, then I don't know what would count.
Is there a narrow path here? Maybe, if there is a major software update after that disengagement and before 31 Aug. Evidence of a disengagement/supervision intervention after 31 August might be useful to rule that out that small remaining path?
Maybe this helps:
https://teslafsdtracker.com/Main
%drivesnocriticalDE 96.8%
Drives no DE 72.5%
Miles to critical DE 465 (was 445 on v13.2.x)
Not sure of the accuracy of these numbers - seems to suggest >2k miles by 10+ testers.
Edit
This data is largely v13 and v12 data and I believe the robotaxis are using v14 so probably not much help sorry.
@ChristopherRandles @dreev Tesla is specifically not pursuing a California robotaxi license to avoid reporting requirements. Texas will eventually require reporting but only of accidents, which the NHTSS already requires. You are going to have to resolve this market without disengagement data.
I'm sure you mean well by hoping to wait for even the Elon himself to agree, but it's time for you to resolve this by trusting your own carefully crafted criteria.
"If humans, remote or in person, are monitoring in real-time and can intervene in real time (even if just hitting a button on the passenger door) then I'd interpret that as supervision."
"this question depends on whether the human safety monitor has to be eyes-on-the-road the whole time with their finger on a kill switch or emergency brake. If so, I believe that's still level 2 autonomy. Or sub-4 in any case."
"We need there not only to be no actual disengagements..."
That last statement alone is all you need. You said you need there to be no actual disengagements, and instead there were many.
Please resolve
@WrongoPhD A single disengagement by a safety monitor in June does not mean Tesla was not launching a Waymo-level-of-autonomy service on August 31st :)
@MarkosGiannopoulos But it is not level 4 if you have a safety monitor supervising.
They simply didn't launch it (lvl 4) before September with no safety monitor / driver.
@ChristopherRandles This has already been discussed extensively before. In short, in my view, just the presence of the safety monitor does not end the Level 4 discussion. Launching an AV service includes testing with safety precautions. Waymo did it too.
@MarkosGiannopoulos That might be an arguable position if the FAQ questions didn't say things like
"9. What about human safety monitors in the passenger seat?
my verdict on this question depends on whether the human safety monitor has to be eyes-on-the-road the whole time with their finger on a kill switch or emergency brake. If so, I believe that's still level 2 autonomy. Or sub-4 in any case.
We need there not only to be no actual disengagements but no counterfactual disengagements. Like imagine that these robotaxis would totally mow down a kid who ran into the road. That would mean a safety monitor with an emergency brake is necessary, even if no kids happen to jump in front of any robotaxis before this market closes. Waymo, per the definition of level 4 autonomy, does not have that kind of supervised self-driving."
If I were to take your view of launching (which I don't) and there is supervision from Aug 2025 with V14 and safety monitors are removed in early 2026 with v16, Would you say lvl 4 launched August 2025 or early 2026? Presumably there would be some point at which you would say ok they didn't really launch level 4 in Aug 2025 there was more development work to do. Where is that line defined? It seems very subjective.
Instead taking the line that we look at things like "We need there not only to be no actual disengagements" as FAQ 9 does, then this gives a more objective time determination of when you can say it has or has not happened. This more objective approach is much to be preferred over a very subjective opinion of when you say it is launched.
@ChristopherRandles Great point about avoiding subjectivity.
Part of what I'm kinda hoping, in terms of being able to fairly resolve this market, is that we'll see a clear case of a kill switch in use after August 31, or a similar smoking gun with tele-operation. Something that makes this an unambiguous NO. At the other extreme, imagine Tesla drops all the safety monitors and shoots past Waymo and complies with California reporting requirements, etc etc, all before the year is out. Even then there'd still be some debate to be had about whether Tesla had truly hit level 4 by August 31 but it might feel like quibbling and that it was a YES in spirit. I'd love to see more markets about this. My guess is that the second extreme is unlikely, and we're on track for a presumed NO if things don't change fairly drastically by early 2026, in addition to no smoking guns in terms of supervision from on or after Aug 31.
What do you think? I can add something like above to the FAQ if it's sounding fair to people. The fundamental difficulty is how in-the-dark we are about what Tesla actually has here. But the longer we wait, the less in the dark we'll be, if for no other reason than the scaling up or lack thereof that we're about to see will itself be evidence. Unfortunately that leaves a large potential gray area where it will be hard to avoid a subjective resolution. I figure we'll cross that bridge when we come to it, but I'm mindful of my conflict of interest and won't just resolve per my own judgment call, should we land in that gray area. Again, my hope is we end up landing in a non-gray area. I do stand by everything in the FAQ so far and will wait for more opinions before augmenting it with anything in this comment.
This is likely mainly FSD supervised for tesla owned cars rather than Taxis
https://www.bbc.co.uk/news/articles/cvg02rdxxz7o
Tesla is being investigated by the US government after reports the firm's self-driving cars had broken traffic laws, including driving on the wrong side of the road and not stopping for red lights.
It said it was aware of 58 reports where the electric cars had committed such violations, according to a filing, external from the National Highway Traffic Safety Administration (NHTSA).
https://static.nhtsa.gov/odi/inv/2025/INOA-PE25012-19171.pdf
58 incidents
Date Opened:10/07/2025
@ChristopherRandles "The traffic authority said Tesla had taken action "to address the issue" of cars going through red lights at a particular intersection in Maryland, where the problem repeatedly occurred." - Yes, this is not about the Robotaxi service but about Supervised FSD in customer cars.


