The reason I ask, is I’ve been seeing alot of news and cases of Tesla’s self driving acting up and being a point of contention. But back in 2016-17 my ex’s uncle and aunt got a Model X when they first dropped and they “auto-drove” us like 50 miles without any noticeable issue.

Was i just gambling my life or has the tech somehow gotten worse?

  • fine_sandy_bottom@lemmy.federate.cc
    link
    fedilink
    arrow-up
    6
    ·
    3 hours ago

    The tech hasn’t regressed, it just hasn’t progressed while the marketing has.

    Look up the automation levels: https://au.pcmag.com/cars-auto/94559/is-your-car-autonomous-the-6-levels-of-self-driving-explained

    My wife’s car is 6 years old, and is level 2. Nothing amazing now, but kinda cool in 2018.

    Since then expectations have increased dramatically, and the problems you’re hearing about are cars expected to have the higher levels of automation but failing to achieve that.

    It seems like this is one of those technical problems that gets exponentially more difficult to solve, the closer we get to solving it. What I mean is, suppose a human averages 100,000km per “incident”. It was easy to make a car do 90,000km per incident, less so to have it do 95,000km per incident, but we’re finding it very very difficult to get that last 5% performance.

    • FuglyDuck@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 hour ago

      It seems like this is one of those technical problems that gets exponentially more difficult to solve, the closer we get to solving it. What I mean is, suppose a human averages 100,000km per “incident”. It was easy to make a car do 90,000km per incident, less so to have it do 95,000km per incident, but we’re finding it very very difficult to get that last 5% performance.

      to this, most cars with relatively high degrees of automation bounce out of it when there’s something that it feels it can’t solve adequately. Most of the time that works, but it would be like if a human were somehow able to only drive when everything was routine. (90% of our driving is routine. it’s that last ten percent that isn’t- and that’s when we crash. well. barring things like drunk driving or distracted driving. Humans are dumb. autonomous driving was designed by humans and isn’t any smarter)

  • FuglyDuck@lemmy.world
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    5 hours ago

    Tesla’s marketing has consistently lied to you about the state of their “full self driving”.

    It’s not. It never was. This is a perfect example of “fake it ‘till you make it”

  • Sanctus@lemmy.world
    link
    fedilink
    English
    arrow-up
    24
    arrow-down
    1
    ·
    6 hours ago

    The tech never “evolved” in the first place. It was a deformity from the start. LIDAR was developed for this reason. But Tesla uses cheap ass cameras that try to interpret what it is seeing through the visual data. I’m guessing with my layman’s knowledge this is why they veer at semi trucks. Because the technology itself is based on a shitty premise.

    • BertramDitore@lemm.ee
      link
      fedilink
      English
      arrow-up
      13
      ·
      5 hours ago

      In general I think you’re right about the tech just being shitty, but a slight correction: LiDAR was not developed for self-driving, it’s just a relevant application of the technology. LiDAR has been around for quite a while, and was initially best known as a remote sensing technology. It is effective at remote sensing because it can penetrate certain solid materials, most importantly foliage. So when an aerial LiDAR dataset is collected for a forested area, since the light can penetrate through most of the foliage, one can essentially ‘delete’ the vegetation from the resulting point cloud, leaving a bare earth model, which is a very close approximation of the landscape’s actual topography if there had been no trees. This can be especially valuable for archaeological research, as foliage is often a significant obstacle for accurately mapping large sites, or even finding them in the first place.

      All of that to say, yeah, self-driving buzz made LiDAR well known as tech, but it wasn’t developed for that purpose.

      • Sanctus@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 hours ago

        Thanks, I did not know that and it is honestly a way cooler application. Foliage X Ray vision.

    • RaoulDook@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 hours ago

      In a non-Tesla car I’ve driven, there was an autopilot cruise control mode that just used cameras. In practice it only works out well if you’re driving long distance on a highway with low traffic. It’s still nice to have (much better than having no autopilot cruise mode) but I don’t trust it around multiple lanes of other cars doing unpredictable shit. Also quits working in the rain when the cameras are obscured.

  • hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    4 hours ago

    Probably the gambling thing. And it’s not like they wreck themselves every 50 miles, so the anecdotal evidence doesn’t really apply. I mean unles they can’t do that anymore. But I think the negative press is about some more rare incidents.

  • NeoNachtwaechter@lemmy.world
    link
    fedilink
    arrow-up
    10
    arrow-down
    1
    ·
    6 hours ago

    Was i just gambling my life or has the tech somehow gotten worse?

    We can safely assume that tech has evolved to the better during this time.

    But it is still too dangerous. It should not be allowed on public roads yet.

  • thesohoriots@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    5 hours ago

    There are also many more Teslas on the road, and the “full self driving” incidents are more widely reported on since the new ownership likes to overpromise and vastly underdeliver. Other commenters have already addressed the tech side, but a few years ago, the Tesla-specific FSD was found to be active right up until a split second before some prolific collisions with emergency vehicles, leading to speculation on liability. Tesla aside, I think it’s just laziness on the part of drivers used to FSD doing the menial tasks of driving.

  • WastedJobe@feddit.org
    link
    fedilink
    arrow-up
    3
    ·
    5 hours ago

    A guy on YT used to (or still does idk) upload entire FSD trips after every update. I haven’t watched any lately but the trend was that the software was improving with some regressions every now and then, which were fixed in the next updates.

  • ch00f@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    5 hours ago

    The first Model X has Autopilot 1 which was a system designed by Mobileye. Tesla’s relationship with Mobileye fell apart and they replaced it with an Nvidia based system in 2017(?). It was really really bad at the start as they were essentially starting from scratch. This system also used 8 cameras instead of the original 1.

    Then Tesla released AP hardware 3 which was a custom-built silicon chip designed specifically for self-driving which also enabled proper navigation of surface streets in addition to the just highway lanekeeping offered in AP1. This broadened scope of actually dealing with turns and traffic from multiple angles is probably where the reputation of it being dangerous has come from.

    My HW3 enabled Model 3 does make mistakes, though it’s rarely anything like hitting a pedestrian or running off the road. Most of my issues are with navigational errors. If the GPS gets messed up in the tunnel, it’ll suddenly decide to take an exit that it isn’t supposed to, or it’ll get in the left lane to pass someone 1/4 mile from a right-exit.

      • ch00f@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        4 hours ago

        Heh, I guess I should have phrased that differently.

        But yeah, it’s actually really courteous. Sometimes a little too much. It’ll move over to the left side of the lane if it sees a cyclist or pedestrian on the shoulder to the right. Unfortunately, it doesn’t understand when there’s a 3 ft concrete barrier between me and the pedestrian and will do it anyway. Makes some narrow bridge crossings a little scarier than necessary.

  • NevermindNoMind@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    6 hours ago

    Just a guess, but it’s probably a combination of two things. First, if we say a self driving car is going to hit an edge case it can’t resolve once in every, say, 100,000 miles, the number of Tesla’s and other self driving cars on the roads now means more miles driven more frequently which means those edge cases are going to occur more frequently. Second, people are becoming over reliant on self driving - they are (incorrectly ) trusting it more and paying less attention, meaning less chance of human intervention when those edge cases occur. So probably the self driving is overall better, but the number of accidents overall is increasing.

  • ContrarianTrail@lemm.ee
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    8
    ·
    edit-2
    5 hours ago

    Tesla recently switched from human-written code to full neural networks with their FSD system, which resulted in a significant leap in how reliable and human-like it became. Nowadays, it’s safe to assume it will perform better than the average human driver on most trips, but it’s still not reliable enough to be blindly trusted. You still need to sit there, paying attention to make sure it doesn’t do anything stupid. It’s, by far the most advanced ‘self-driving’ system available on a vehicle you can buy.

    The reason it’s constantly in the news is that anything related to Elon, AI, or self-driving is guaranteed to get clicks in today’s mediascape. It’s not a flawless system by any means but it’s also not as bad as people make it to be.