Tesla's decision not to use the LIDAR as a safety feature (i.e. having reliable high-resolution data about things the car can collide with) is so incredibly indefensible, since solving the last 1% of this using only vision likely requires a general artificial intelligence
Prediction: Tesla will be the last of all major auto manufacturers to get to L5 autonomy. Time interval between when Tesla L5 FSD is finally available and when humanity is destroyed by the general AI it runs on will be very awesome and also very short
You get sparse point cloud from LIDAR sensors, not accurate 3D maps. This is the main reason why some people think LIDAR may not work well (mostly, only comma.ai and Tesla folks).
Vision can also get you 3D maps, either in active manner (IR floodlight or structured lighting) or not. I will reserve my judgement until see more from either side.
This framing is a common error in the debate. It's not cameras or lidar, it's cameras or cameras + lidar + radar. Nobody is driving on lidar alone. Many others actually have more cameras and are doing substantially more vision than Tesla is, they're just fusing lidar and radar perception with their vision pipeline. It gives you a more robust view of the world than using a single sensor modality.
If you have one piece of rotten meat in a perfect stew, you still have a disgusting dish. Good sensor fused with garbage in is still garbage in. That was one of the major points of the talk - the vision-only system is more accurate than the one with other modalities thrown in, even though the latter has more data. We intuit that the fusion network should just learn to ignore the bad sensor when it's unreliable, but this rarely happens in practice.
If anything, knowing when to reliably ignore a sensor modality is the kind of intuition more associated with general AI.
A similar paradox occurs when trying to fuse multispectral imagery. You'd think early fusion of RGB and IR would be better since it gives the higher-resolution filters access to more data, but it does worse than late fusion. My understanding is that late fusion forces the network to "work harder" to solve object detection using IR only, and then once you've wrung what you can out, then you fuse with RGB detections.
Since radar is "one pixel" there's essentially only one object detector possible: object or nothing. If yes-object, fusion tries really hard to make sense of the RGB filters to figure out what partial detection looks like an object, which is almost always a false positive.
You have fallen victim to the trap that this video so perfectly laid out for you. Tesla used the mmWave radar that has been in cars forever since it's a good way to do emergency braking and things like adaptive cruise control, particularly when you are a new company and you need these capabilities on your luxury sedan from day one. Now that they are much further along in their FSD efforts, they realize this mmWave radar isn't very helpful anymore. Cool, but nobody else was using it to begin with. LIDAR is totally and utterly different sensor technology.
They are using Continental 2D radars from 2014. The rest of the SDC industry uses what is called 4D high resolution imaging radars. That's why Teslas have poor performance like phantom braking under bridges.
https://arberobotics.com/product/ is one. It was rumored Tesla themselves wanted to use Arbe radars, but evidently they have ditched the plan and gone radar less. I think AutoX from China is planning to use Arbe as well.
Waymo has its own custom developed 4D imaging radar and I imagine most of the other SDC players have their own versions.
None what? I just gave you an example of Waymo using 4D imaging radar [1]. Aurora also uses imaging radar [2], Mobileye as well [3]. AutoX is going to start using Arbe radars. All of them are leading players in SDC industry.
You said “everyone else” besides Tesla uses 4D radar, and “that's why Teslas have poor performance like phantom braking”. But there are zero production vehicles on the road using this tech right now.
I said “The rest of the SDC industry uses 4D radars” in a follow up comment, which clarifies I don’t mean car manufacturers.
Given this post is about radar usage in self driving and how Tesla/Karpathy touts their capabilities, I compare them to other SDC players, not other car manufacturers who are not in the game. And their SDC competitors are running away because they have the newer 4D radars.
Well, until these are actually adopted by a car company and out on the road, any claims about performance are meaningless. Your original comment implied Tesla chose to use inferior tech compare to the market which is not true in the slightest.
> That was one of the major points of the talk - the vision-only system is more accurate than the one with other modalities thrown in, even though the latter has more data. We intuit that the fusion network should just learn to ignore the bad sensor when it's unreliable, but this rarely happens in practice.
This makes some important assumptions, namely that Tesla built a lidar and radar perception pipelines and sensor fusion of equivalent quality to their competitors, and then decided they were unnecessary.
Given that their competitors have shown substantially better perception than Tesla, and that Tesla has a significant economic incentive to deliver autonomous driving on a sensor suite that already shipped years ago, I find that difficult to believe. Did Tesla build good enough perception to dismiss lidar and radar purely on their merits. Unlikely I think. Did an intern build a student-quality lidar pipeline that "proved" Elon's camera-first approach is the right one? More likely.
Sensor fusion is hard, especially with data at various qualities (different generations of LIDAR, radar, probably images are more portable actually).
I am actually on sensor-fusion side, and think a transformer can merge everything and generate a coherent world-view. But this is a hard problem and people side-step them after the evaluation shouldn't be dismissed blatantly.
For one:
> It gives you a more robust view of the world than using a single sensor modality.
One of the arguments made in this talk is which sensor to trust when there is a disagreement. I'd love to see the scenarios where this happens with three types of sensors and the logic they use to choose the "right" one.
You can't avoid this problem by only having a single sensor modality, because you still have to decide if you trust your sensor data or not. In effect, you actually made the problem harder because you have no other information from any other sensor modalities to help understand the world around you.
In effect, they solved the disagreement problem by sticking their head in the sand and pretending their sensors are never wrong.
You’re intent on oversimplifying the issue to support your view, but this comment makes no sense. Of course you can avoid sensor fusion issues by not doing sensor fusion. Regardless of how many sources you have, you still have to decide if you trust the data, having less sensors doesn’t change that. And having more sensors means you have to decide which input to trust.
That's a problem for deciding which system drives the thing (eg. do you use cameras for detecting oncoming cars and objects or lidar/radar), but if there are major disagreements between them it would make most sense to disengage the autonomous part and ask the human driver to take over, much like current TSLA cars do. Obviously not possible if you want level 4 self driving, but it can be good enough for level 3.
> Many others actually have more cameras and are doing substantially more vision than Tesla
This is actually my main criticism of Tesla's approach. They don't seem to have enough cameras to do the job well, and its showing in a lot of actual system limitations.
LIDAR is valuable as a safety feature, i.e. it can (unlike radars or cameras) reliably see (at least in clear weather) if there's anything in the car's path warranting evasion/braking maneuvers. In particular it's important that the LIDAR is dumb, i.e. its failure mores are predictable
LIDAR operates by timing how a photon reflects from a surface, it doesn't guarantee see everything. As you stated already, it cannot see in snow or rain.
I am actually on sensor-fusion side, but I don't think we should jump to the conclusion LIDAR is the best 3D mapping method.
For one example, a truck has a breaking distance of 600ft, either Velodyne or OS1 LIDAR has range less than that.
I agree about LIDARs not being the best general 3D mapping method, my point was mostly about using it as a dumb physics-based safety system. For autonomous trucks, limited LIDAR range could be mitigated by reducing speed accordingly and/or employing more powerful (or SWIR) LIDARs since trucks are more expensive and there could be a bigger budget for sensors
They’re betting that they can use a massive feedback loop to train a set of neural networks to the point where they are as accurate as LiDAR without actually firing any lasers.
Even if you believe this goal is possible to achieve at some point in the future, I think the argument falls apart when you consider that it will take years, probably decades, for a pure vision approach to catch up to where Waymo is today in terms of safety. (They have cameras too.)
That Tesla can’t afford to fit expensive LiDAR sensors to all of the cars it sells is Tesla’s problem. Regulators won’t give a shit that pure vision is “better” in theory. They will simply compare Tesla’s crash rate in autonomous mode with that of Waymo and other AV operators, and act accordingly.
I understand why they made the "no-LIDAR" bet early when the LIDARs were completely unpractical for a production consumer car
However, nowadays it starts to look that 100% reliable depth estimation from cameras might actually require a human-level AI to work and also solid-state LIDAR technology is becoming cheap enough and integrateable into normal cars, but Tesla can't really change their stance on this without admitting that FSD options they already sold would not actually become FSD within the lifetimes of these vehicles. I suspect this might also be the reason why Karpathy looks more and more nervous with each new talk
>100% reliable depth estimation from cameras might actually require a human-level AI to work
you don't need 100%, and even humans are far from 100% (500Mpx resolution of our eyes allows to basically sheer brute force through it in many cases). The stereo setup provides great and fast estimation with several megapixel resolution with good fps (way better than lidar) for majority of situations. It is some share of the [part of the] scenes, and you really know it right then and there, where you need AI and/or very sophisticated compute heavy algorithms. So instead of throwing AI and the compute power at those parts, you just pull the points from the lidar (and even radar if the things are that bad) covering that segment. And that way, given a couple more iterations of sensors (from current 20Mpx+ to the hundreds Mpx) and compute, it will be doing even better than humans. Anybody not doing sensor fusion would be a loser though - just like going into a fist fight with one hand intentionally disabled.
One of the gains from using lidar is also that it's a different sensor altogether from cameras, with different failure modes.
For example, cameras are sensitive to glare from reflections (sun near sunset or reflecting on metallic objects) and oncoming traffic at night. Lidars operate on a different narrower wavelength and are unlikely to be affected by that, although they might struggle with objects that have low reflectivity at a long distance.
The fact that these sensors are different means that the intersection where a dangerous situation would not be detected by either sensor is much smaller than any sensor individually.
In any case, once AVs are deployed at scale, if it becomes apparent that some sensors can be removed or replaced by something else, then they will be if there's a case for it.
> I think the argument falls apart when you consider that it will take years, probably decades, for a pure vision approach to catch up to where Waymo is today in terms of safety. (They have cameras too.)
On what set of metrics do you think Waymo is safer? IMO it's too early to compare and cherry-picked proofs both from Waymo and Tesla are not really representative.
For starters, Waymo reported 1 disengagement per 29,944 miles driven in 2020 to the California DMV [1], while in the talk, Karpathy implies that a Tesla being able to drive around the SF area for 2 hours without a disengagement is unusual. Note that Tesla didn't file a disengagement report because they didn't do any autonomous testing on public roads in California in 2020.
There are issues with reading too much into disengagements, but there certainly seems to be a large difference here.
Isn’t Waymo only driving in small HD mapped areas? Previously the were only driving within a 50 mile area of Arizona where it’s clear weather all the time.
Arizona roads are also mapped to extreme precision, have very wide lanes, and are optimized for cars. Waymo has prioritized low intervention by being overly cautious and avoiding hard maneuvers (like many left turns).
That doesn't work when they scale up to any other set of normal roads, especially as density and complexity increases.
They don't avoid left turns. There are plenty of videos from Chandler, AZ of Waymo performing unprotected left turns perfectly fine.
They will always map roads to precision, whether it's Arizona or San Francisco. Why is that a problem? You should either look at their CA disengagement reports over the years or wait until they roll out a service in SF (where they've been testing heavily). That will show how safe they are in dense environments.
From what I gather, they manually mark sections as hard when the cars get stuck there, e.g. due to road work, and then their routing system chooses another route, e.g. one that avoids the left turn.
The video with the Waymo car getting stuck and taking off from the rescue team had an example of this.
I guess it makes perfect sense from a engineering perspective.
I predict the opposite. Tesla sold half a million cars last year and will sell nearly one million this year. The data they have access to is increasing by orders of magnitude. I bet there is a point, let's say 20 million cars total, where they can pull so much high quality data that they will be able to surpass lidar capabilities for the purposes of self driving.
The lidar/no lidar discussion is a fun one because people have different ideas about how the world works. Personally I think LiDAR is the modern version of expert systems. It appeals to a logical/geometric intuition but the approach is brittle to real world contact, especially when paired with HD maps which are a great way to drive yourself into a local maximum.
The fallacy here is that the scale of the neural network used by Tesla is sufficient to capture the problem of driving given enough training. There is no guarantee that a reasonably priced neural network can encompass the task of driving.
Having training data beyond a certain point is overrated, and Tesla's advantage in gathering it is overstated. Other companies are capturing this data as well. Is there any indication that the data Tesla is collecting is of a higher value, or is it just more bytes?
It seems as if the people gobbling up the "Tesla has the data! Autopilot will keep getting better!" line have never trained a neural network in their life. Models converge. Loss stops decreasing, regardless of more incoming data. Extreme manual data cleaning effort becomes required to prevent overfitting. Model architecture has to change and hyper parameters have to be tweaked. Then you're back at square one as far as testing goes if you change any of those things.
The notion that Tesla's model HAS to keep improving simply because they will be able to pile on more (unlabeled!) data is laughably false. And, in fact, quite insulting to the intelligence of even the most casual ML engineers.
> And, in fact, quite insulting to the intelligence of even the most casual ML engineers.
Exactly, casual ML engineers. The issue of plateauing tends to occur because there is no more novelty to be had in the data. What mega-experiments like GPT and similar have shown us is that actually you can keep adding novel data and keep improving the model. Kinda inelegant, yet effective. The problem is, most institutions can't add more novelty beyond a certain scale, since that usually means shoveling more money at data storage and compute, on top of the novelty collection.
Tesla merely has to open the money tap to get more of both compute and storage, and let the real-time data flow in.
> Tesla merely has to open the money tap to get more of both compute and storage, and let the real-time data flow in.
And if you watch the other parts of the presentation, you'll see the bits about them buying clusters with 5k+ A100 GPUs. Presumably they intend to do something with those. Probably not streaming Fortnite concerts.
I would agree if their increase in data was linear, but it is increasing by orders of magnitude, which should have qualitative consequences for what they're able to accomplish as they claw their way through 9s. I don't see how it's possible to get progressively more 9s without scaling in both data and compute.
The point of the higher scale isn't just more data, it also makes it easier to solve the unbalanced data problem, because rarer and rarer scenarios will appear in large enough numbers to work with.
You make it sound extremely manual and sequential when reality is anything but.
A team with funds like Tesla, Google, FAIR is going to be using NAS and have a continuous testing pipeline. Tesla has arguably the best environment for continuous testing which is the most difficult part of improving a model. Andrej even said in his talk that their supercomputer is in the top 5 for FLOPs.
SOTA on ImageNet for the past few years has been driven by pre-training on massive datasets. Vision transformers are increasingly more common and are extremely data-hungry.
I'd say the data that Tesla collects is of lower value, because it doesn't have sensor info from a different modality. Other companies are getting a good reference to ground truth for both camera to lidar and lidar to camera. I don't know how much more valuable accurate distance sensing over a 3d field is compared to not having it, but I so know it's more valuable.
It may be valuable enough to require a few petaflops less computing power.
We are chasing the 9s... diminishing returns are still returns. If 10x the data improves from 99.999 to 99.9995, that is still progress. Maybe 100x data gets you to 99.9999.
Are we even at 80% yet, compared to non-drunk/drugged/sleep-deprived regular human drivers? All of the videos I've seen are well below the driving capacity of a human taking their first driving lesson.
> Prediction: Tesla will be the last of all major auto manufacturers to get to L5 autonomy.
Tesla is also the only company to claim to target L5 autonomy. Everyone else, including Waymo, is strictly targeting L4 and say L5 autonomy is not possible or realistic. L5 is a pipe dream.
Well yes, but I think they'll be singing a different tune in 5 years, and even more so when more Tesla cars actually start driving in weather conditions less favorable than Palo Alto and Austin.
At some point when LIDAR is cheap enough there is no reason for Elon Musk to not give in and use them. Right now he's constraining the problem to the cost of the car.
LiDAR is not better in worse weather conditions. LiDAR performance degrades in rain and snow. That’s where something like radar is better. The Tesla get is that humans drive with vision so they should be able to as well. Also every other self driving solution must solve the vision problem also in order to be successful. LiDAR doesn’t tell you that it’s a bag on the road vs a raccoon. So the question is that once you solve the vision problem, do you still need LiDAR for any meaningful impact?
> LiDAR doesn’t tell you that it’s a bag on the road vs a raccoon
I think that's largely an issue with the early LIDAR devices today, but not necessarily what may be to come.
There's something to be said about measuring actual data with solid physics vs. inferring distances with billions of operations on RGB data. If you were landing a commercial aircraft in fog, you most certainly don't rely on your eyes to do most of it, but it is in fact possible to do safely precisely because we do have good sensors on them.
I fully agree with leveraging the scale and maturity of RGB sensors for cars today, the talk is spot on about that, but that's (a) circling back to the fact that Tesla needs to sell cars now not next year and (b) not a good case against use of LIDAR in the future.
It’s easy to add a few hundred thousand dollars of sensors to a $100 million plane. But $20k on a $40k car is too much and makes it cost prohibitive for most people. If vision alone can get you to say 1 crash in 10 million miles, that’s more than good enough to replace human drivers. If someone decides they want to go the extra mile and will pay for something rated for 1 in 20 million miles for 2x the cost, then they are free to do so.
As mentioned in this talk Tesla is taking an iterative approach and trying to make things safer for people today, not in 5 yrs from now. Maybe in a few years Tesla will see that to go from 10 million miles to 20 million miles they need 4k cameras at 60 fps, but the work they had already done would still have had a big impact. You don’t need to do things in one shot and get to the finish line.
> As mentioned in this talk Tesla is taking an iterative approach
> You don’t need to do things in one shot and get to the finish line.
I think we agree on these things. I wasn't trying to dispute their approach, rather explain their rationale for their current method, which is focused at actually selling product now.
> But $20k on a $40k car is too much
Those sensors don't need to be 20K-40K, they just aren't mass produced yet. That's why I think Elon Musk will change his tune in some years when those sensors are 1/100 the cost and 10X more accurate than vision.
I strongly disagree. By all measures I've seen (including a couple of slides in the OP's video), Tesla's self-driving is far safer than human driving: the number of accidents and deaths per mile driven are something like an order of magnitude lower (i.e., around 10x safer). I mean, the machine never gets distracted, tired, sleepy, emotional, drunk, etc., so it is a LOT LESS likely to crash on boring, monotonous road segments than most people -- who do get distracted, tired, sleepy, etc. Not only that, but people make really scary mistakes in routine circumstances. The video shows several examples of human drivers hitting the accelerator when they actually meant to hit the brake!
The criticism of autopilot is really about it getting tripped-up in response to statistically rare, unusual circumstances, i.e., edge cases. Karpathy et al are working on getting better at those, bringing the rate of situations that surprise autopilot closer and closer to 0%, even if it can never be achieved -- there will be always be surprises. Personally, I would rather take a tiny risk of crash on rare, once-in-a-million-miles events with autopilot driving than a ~1% risk of crash per 1000 to 2000 miles with everyday human driving.
Prediction: Tesla will be the first of all major automakers to get to level 4 and 5 autonomy.
>Tesla's self-driving is far safer than human driving: the number of accidents and deaths per mile driven are something like an order of magnitude lower (i.e., around 10x safer).
Lies, damn lies, and statistics.
Tesla here, is, again, being funny with the numbers. They LOVE to cite autopilot ON death statistics as being "10x safer than normal driving". What they fail to note is that Autopilot can ONLY be on while driving on a limited access highway. Highways are much safer to drive on than a mix of ALL ROADS, which is where the baseline figure comes from.
Another confounding factor is the price of the vehicle. The average CONFIGURED Tesla with the FSD package today costs what? $65k? More? Those X's and S's are $100k+. Nobody is buying that base Model 3. The point is that Tesla drivers are 1) Older and 2) Wealthy. Wealthy, older people get in far fewer car crashes than the average driver. In fact, car crash fatalities are really driven by two groups: drunks (or pill addicts), and young (teenage) men. Not saying it's IMPOSSIBLE to have a substance abuse problem and own a Tesla, but the average Tesla owner is less likely to have these issues. It's also less likely to own a Tesla while young.
So, Tesla autopilot stats should be compared to other comparably priced vehicles while driving on the highway ONLY. That would actually be a fair, honest comparison. I believe a recent outgoing BMW 5 series chassis finished its entire life without a single fatality in the US. That's right -- 4-5 years of service in the US without a single death. Turns out, wealthy people who drive expensive family sedans don't get in a lot of fatal highway crashes.
Here's a Forbes article (sorry) doing some of the back-of-the-napkin math. They estimated that in Q3 2019, autopilot really wasn't any safer than manual driving.
> So, Tesla autopilot stats should be compared to other comparably priced vehicles while driving on the highway ONLY.
I disagree. Autopilot driving on the highway should be compared to all human drivers driving on the highway. Otherwise you wouldn't be comparing against human performance per mile driven apples-to-apples.
Modern vehicles have much safer crash characteristics than older cars. The average vehicle on the road in the US is 11 years old. Do you know how much crash characteristics of cars have improved in the last decade? The comparison needs to stay in the modern, $65k+ vehicle realm for it to be apples-to-apples. Otherwise, you're comparing a bunch of decade old rust buckets with heat-cycled rubber, no blind spot monitoring, and Takata airbags to modern vehicles and claiming victory. Come on.
The BMW F10 535i had ZERO FATALITIES over its entire life in the US. Zero. It had no "self-driving" capabilities. Just lane-departure warning, BLIS, and ACC.
> The BMW F10 535i had ZERO FATALITIES over its entire life in the US. Zero. It had no "self-driving" capabilities. Just lane-departure warning, BLIS, and ACC.
Yes, all evidence I've seen indicates that cars partially driven by computers (adaptive cruise control, lane-departure warning, blind spot information, etc.) are safer than cars entirely driven by human beings. The BMW you mention is safer precisely because it is partially driven by machines. The more we automate driving, as machines get better and better at it, the safer we will all be on the road.
> Yes, all evidence I've seen indicates that cars partially driven by computers
Emphasis on PARTIALLY. Anyone who has read recent takeover scenario studies is rightfully horrified at the notion of a completely “hands off” driving experience, where the driver is expected to remain alert and vigilant but they’re not inputting any steering, throttle, or braking. Unsurprisingly, it takes people about 2 seconds to re-engage as active drivers. 2 seconds is way too long, which is why it may be safer to NOT use a system that does steering input in addition to throttle and brake. You need to keep the drivers actively engaged. And no, “touch the steering wheel every 30 seconds” is not active engagement. And if a car has “self driving” but also active driver monitoring, what’s the point? The driver doesn’t get to relax at all. The stress of driving doesn’t come from the input unless you’re racing. The stress of driving comes from having to stay alert. If I have to stay alert, I’d rather just drive myself instead of trusting an experimental system that drives like an indecisive, half-blind grandmother.
The BMW was safe PRIMARILY because it’s a well-designed, modern car, driven by an older and wealthy (safe) demographic. The assistance systems are probably secondary. They weren’t even standard on all vehicles and they were very primitive in that first generation.
If you actually read the Forbes article above, the back of the napkin math actually DOESN’T indicate that Teslas in autopilot are safer than normal driving. That’s the entire contention. I do not think these full-takeover systems are safer at the present time than active human drivers in comparable vehicles with safety assist systems. Tesla is very clearly fudging the numbers to make it appear as if autopilot is safer, but the claim doesn’t stand up to some really basic analysis.
Tesla does not have self driving. Production Autopilot is not self driving. It's advanced cruise control with lane keeping, nothing that any other manufacturer doesn't offer on their cars. Full self driving doesn't even work, and is acknowledged by Tesla themselves, calling FSD a beta. And FSD can barely even do basic tasks like making an unprotected left turn.
>Tesla's self-driving is far safer than human driving: the number of accidents and deaths per mile driven are something like an order of magnitude lower
Still not close to good enough for people to accept:
"Participants from both countries required Self Driving Vehicles to be 4-5 times as safe as Human Driven Vehicles" [0]
It’s already about 10x. But it is a bit tricky since they don’t break the numbers out by highway and local driving. Also their active safety features outside of AP also improve safety. So does it need to be 4-5x better than an already improved system that has AEB and lane departure avoidance and other safety features?
> In the 1st quarter, we registered one accident for every 4.19 million miles driven in which drivers had Autopilot engaged. For those driving without Autopilot but with our active safety features, we registered one accident for every 2.05 million miles driven. For those driving without Autopilot and without our active safety features, we registered one accident for every 978 thousand miles driven. By comparison, NHTSA’s most recent data shows that in the United States there is an automobile crash every 484,000 miles.
There are way too many confounding variables in those numbers for you to directly compare them like you are doing. The biggest being that Autopilot is predominately engaged in situations that are already safer than average driving.
Of course autopilot and FSD are safer than unaided human driving. That's because a human is still required to be in ultimate control. They are aids to the human, not replacements for the human.
Illusory superiority [1] will make us think the bad drivers are only those below average drivers. It will take a while for people to truly trust FSD just by accident stats.
So all of that 'research' and several FSD crashes later and they have added a new driver monitoring tool in the refreshed line of Tesla Model S Plaid vehicles because even with FSD turned on, you must have your eyes on the road at ALL times whilst driving. [0] Looks like someone made a correct prediction on this technology and it was none other than Comma.ai [1]
I have to give it to Elon that he was able to keep his fans believing a big lie of their FSD system to achieve Level 4 / 5 autonomy last year, when this year it was admittedly Level 2 [2].
The fans will continue to believe his lies and keep on saying "It's coming soon, you'll see..." when they have just been sold a Fools Self Driving System™.
No one is going to get to a true L5 for a long time. That is totally irrelevant. It's a war of attrition. Whoever can monetize L3/L4 and can scale without any vehicle upgrade cost is going to win. It's pretty obvious lidar is very very silly since it doesn't scale.
It is also pretty easy to see that Tesla doesn't have to hit L5 to have won autonomy. It just has to successfully monetize L3/L4.
It's an anti-fragile approach, a term you'll recognize if you know of Nassim Taleb and his work. I think it will win in the long run, because not requiring HD Maps or specialized sensors is an advantage, and even if it takes more resources to make it work initially, it will save billions in the future, assuming, of course, it ever ships.
Obviously, pure vision is a viable system, as it's what we as humans use. The question remains as to whether or not it will be comparable to more precise LiDAR based systems in the near future.
This is not a good argument overall for several reasons. First, we should aim to greatly exceed human performance on safety. Second, whatever work goes into making the algorithms work well using vision can just as easily still inform a system that fuses that data with LIDAR for enhanced situational awareness and safety.
Waymo etc do not use LIDAR for object sensing, only for positioning. LIDAR sucks for object sensing because it gives you no information about whether it's a plastic bag or a person — you still need vision for that. Even if you just err on the safe side and brake, that itself can cause an accident unecessarily.
Next-gen LIDAR has great density, and I bet it would do a decent job differentiating between a plastic bag or a rock in the middle of the road. In addition to depth LIDAR also returns intensity and several other metrics, which can be used as input to an ML model. It's why you can read the lettering on the side of the semi truck in this video of Waymo's next-gen LIDAR.
How can you make such a strong statement when you simply don't know how to achieve full autonomy? This reminds of the teapot orbiting the Sun argument [0]. You and people defending Lidar by their teeth don't sound too different from religious zealots who've "seen the light".
Ultimately the proof is in the pudding. With Tesla FSD you can drive the highways from New York to Boston without any issues. I am sure there are many more routes across the country that can also be driven like that. Definitely not L5, but it works. Yet to see that from any other automaker, LIDAR or not. As far as I am concerned, great job Tesla! Keep it up, I am sure you will work through more tougher problems.
Generally this is true, but in the safety critical domain you don't gamble. You do your homework and make sure you aren't exposing your users to unnecessary levels of risk.
If Tesla was developing their system with trained safety drivers or on closed courses, I think they would have higher moral ground to gamble here. But placing the untrained public behind the wheel of alpha quality software is unethical IMHO. There ways to develop autonomous software that are significantly less risky, and the only reason Tesla is doing it this way is for marketing/PR purposes as far as I can tell.
Why do you think the last 1% is dependent on LIDAR versus any of the other multitude of gaps between today’s autonomy and L5?
If the only way it becomes practical to achieve L5 is to use LIDAR, Tesla can obviously add it. But if they waited until LIDAR was cheap and practical, they still wouldn’t be shipping any hardware doing autonomy today, and not collecting the data needed to train their models and delivering value today.
Also, with vision based systems, it operates in somewhat an intuitive fashion given we have eyes too.
Prediction: Tesla will be the last of all major auto manufacturers to get to L5 autonomy. Time interval between when Tesla L5 FSD is finally available and when humanity is destroyed by the general AI it runs on will be very awesome and also very short