Mark Rober just set up one of the most interesting self-driving tests of 2025, and he did it by imitating Looney Tunes. The former NASA engineer and current YouTube mad scientist recreated the classic gag where Wile E. Coyote paints a tunnel onto a wall to fool the Road Runner.

Only this time, the test subject wasn’t a cartoon bird… it was a self-driving Tesla Model Y.

The result? A full-speed, 40 MPH impact straight into the wall. Watch the video and tell us what you think!

  • FlashMobOfOne
    link
    fedilink
    English
    131 month ago

    I am very glad that Elon and Trump have overreached and now Tesla is suffering. I hope Starlink is the next domino to fall.

      • FlashMobOfOne
        link
        fedilink
        English
        111 month ago

        Good.

        I’m not rooting for America to fail, but I’m 100% rooting for Elon to fail.

    • Snot Flickerman
      link
      fedilink
      English
      241 month ago

      I’m so glad I wasn’t the only person who immediately thought “This is some Wile E. Coyote shit.”

      • @Mr_Dr_Oink@lemmy.world
        link
        fedilink
        English
        -291 month ago

        As much as i want to hate on tesla, seeing this, it hardly seems like a fair test.

        From the perspective of the car, it’s almost perfectly lined up with the background. it’s a very realistic painting, and any AI that is trained on image data would obviously struggle with this. AI doesn’t have that human component that allows us to infer information based on context. We can see the boarders and know that they dont fit. They shouldn’t be there, so even if the painting is perfectly lines up and looks photo realistic, we can know something is up because its got edges and a frame holding it up.

        This test, in the context of the title of this article, relies on a fairly dumb pretense that:

        1. Computers think like humans
        2. This is a realistic situation that a human driver would find themselves in (or that realistic paintings of very specific roads exist in nature)
        3. There is no chance this could be trained out of them. (If it mattered enough to do so)

        This doesnt just affect teslas. This affects any car that uses AI assistance for driving.

        Having said all that… fuck elon musk and fuck his stupid cars.

        • @teuniac_@lemmy.world
          link
          fedilink
          English
          691 month ago

          This doesnt just affect teslas. This affects any car that uses AI assistance for driving.

          Except for, you know… cars that don’t solely rely on optical input and have LiDAR for example

          • @Mr_Dr_Oink@lemmy.world
            link
            fedilink
            English
            -61 month ago

            Fair point. But it doesn’t address the other things i said, really.

            But i suppose,based on already getting downvoted, that I’ve got a bad take, either that or people who are downvoting me dont understand i can hate tesla and elon, think their cars are shit and still see that tests like this can be nuanced. The attitude that paints with a broad brush is the type of attitude that got trump elected…

            • @sugar_in_your_tea@sh.itjust.works
              link
              fedilink
              English
              41 month ago

              based on already getting downvoted

              In this case, yes, but in general, downvotes just mean your take is unpopular. The downvotes could be from people who don’t like Tesla and see any defense of Tesla as worthy of downvotes.

              So good on you for making the point that you believe in. It’s good to try to understand why something you wrote was downvoted instead of just knee-jerk assuming that it’s because it’s a “bad take.”

            • @Voroxpete@sh.itjust.works
              link
              fedilink
              English
              421 month ago

              No, it’s just a bad take. Every other manufacturer of self driving vehicles (even partial self driving, like automatic braking) uses LiDAR because it solves a whole host of problems like this. Only Tesla doesn’t, because Elon thinks he’s a big brain genius. There have been plenty of real world accidents with less cartoonish circumstances involving Teslas that also would have been avoided if they just had LiDAR sensors. Mark just chose an especially flashy way to illustrate the problem. Sometimes flashy is the best way to get a point across.

            • @Reyali@lemm.ee
              link
              fedilink
              English
              8
              edit-2
              1 month ago

              I agree the wall is convincing and that it’s not surprising that the Tesla didn’t detect it, but I think where your comment rubs the wrong way is that you seem to be letting Tesla off the hook for making a choice to use the wrong technology.

              I think you and the article/video agree on the point that any car based only on images will struggle with this but the conclusion you drew is that it’s an unfair test while the conclusion should be that NO car should rely only on images.

              Is this situation likely to happen in the real world? No. But that doesn’t make the test unfair to Tesla. This was an intentional choice they made and it’s absolutely fair to call them on dangers of that choice.

              • @Mr_Dr_Oink@lemmy.world
                link
                fedilink
                English
                51 month ago

                That’s fair.

                I didn’t intend to give tesla a pass. I hoped that qualifying what i said with a “fuck tesla and fuck elon” would show that.

                But i didn’t think about it that way.

                In my defense my point was more about saying “what did you expect” the car to do in a test designed to show how a system that is not designed to perform a specific function cant perform that specific function.

                We know that self driving is bullshit, especially the tesla brand of it. So what is Mark’s test and video really doing?

                But on reflection, i guess there are still a lot of people out there that dont know this stuff, so at the very least, a popular channel like his will go a longway to raising awareness of this sort of flaw.

        • @Daefsdeda@sh.itjust.works
          link
          fedilink
          English
          -51 month ago

          I agree that this just isn’t a realistic problem, and that there are way more problems with Tesla’s that are much more realistic.

        • KayLeadfootOP
          link
          fedilink
          11 month ago

          I am fairly dumb. Like, I am both dumb and I am fair-handed.

          But, I am not pretentious!

          So, let’s talk about your points and the title. You said I had fairly dumb pretenses, let’s talk through those.

          1. The title of the article… there is no obvious reason to think that I think computers think like humans, certainly not from that headline. Why do you think that?
          2. There are absolutely realistic situations exactly like this, not a pretense. Don’t think Loony Tunes. Think 18 wheeler with a realistic photo of a highway depicted on the side, or a billboard with the same. The academic article where 3 PhD holding engineering types discuss the issue at length, which is linked in my article. This is accepted by peer-reviewed science and has been for years.
          3. Yes, I agree. That’s not a pretense, that’s just… a factually correct observation. You can’t train an AI to avoid optical illusions if its only sensor input is optical. That’s why the Tesla choice to skip LiDAR and remove radar is a terminal case of the stupids. They’ve invested in a dead-end sensor suite, as evidenced by their earning the title of Most Lethal Car Brand on the Road.

          This does just impact Teslas, because they do not use LiDAR. To my knowledge, they are the only popular ADAS in the American market that would be fooled by a test like this.

          Near as I can tell, you’re basically wrong point by point here.

          • @Mr_Dr_Oink@lemmy.world
            link
            fedilink
            English
            2
            edit-2
            1 month ago

            Excuse me.

            1. Did you write the article? I genuinely wasn’t aiming my comment at you. It was merely commentary on the context that is inferred by the title. I just watched a clip of the car hitting the board. I didn’t read the article, so i specified that i was referring to the article title. Not the author, not the article itself. Because it’s the title that i was commenting on.

            2. That wasn’t an 18 wheeler, it was a ground level board with a photorealistic picture that matched the background it was set up against. It wasnt a mural on a wall, or some other illusion with completely different properties. So no, i think this extremely specific set up for this test is unrealistic and is not comparable to actual scientific research, which i dont dispute. I dont dispute the fact that the lack of LiDAR is why teslas have this issue and that an autonomous driving system with only one type of sensor is a bad one. Again. I said i hate elon and tesla. Always have.

            All i was saying is that this test, which is designed in a very specific way and produces a very specific result, is pointless. Its like me getting a bucket with a hole in and hypothesising that if i pour in waterz it will leak out of the hole, and then proving that and saying look! A bucket with a hole in leaks water…

            • KayLeadfootOP
              link
              fedilink
              11 month ago

              Y’all excused, don’t sweat it! I sure did write the article you did not read. No worries, reading bores me sometimes, too.

              Your take is one of the sillier opinions that I’ve come across in a minute. I won’t waste any more time explaining it to you than that. The test does not strike informed individuals as pointless.

              • @Mr_Dr_Oink@lemmy.world
                link
                fedilink
                English
                1
                edit-2
                1 month ago

                I dodnt not read it because “reading bores me.” i didn’t read it because i was busy. I have people round digging up my driveway, i have a 7 week old baby and a 5 year old son destroying the house :p i have prep for work and i just did a bit of browsing and saw the post. Felt compelled to comment for a brief break.

                Im not sure what you mean by “silly opinion.” Everyone who has been arguing with me has been stating that everyone knows that teslas dont use LiDAR, and thats why this test failed. If everyone knows this, then why did it need proving. It was a pointless test. Did you know: fire is hot and water is wet? Did you know we need to breathe air to live?

                No?

                Better make an elaborate test, film it, edit the video, make it last long enough to monetise, post it to youtube, and let people write articles about it to post to other websites. All to prove what everyone already knows about a dangerous self driving car that’s been around for 11 years…

                I am sorry, i just dont get it. I felt like I was pointing out the obvious in saying that a test that’s tailored to give a specific result, which we already know the result of, is a farcical test. It’s pointless.

      • @bstix@feddit.dk
        link
        fedilink
        English
        31 month ago

        A camera will show it as being more convincing than it is. It would be way more obvious in real life when seen with two eyes. These kinds of murals are only convincing from one specific point.

        • @ParadoxSeahorse@lemmy.world
          link
          fedilink
          English
          11 month ago

          …and clearly this photo wasn’t the point. In fact, it looks like a straight road from one of the camera angles he chooses later, not afaict from the pov of the car

        • @Buffalox@lemmy.world
          link
          fedilink
          English
          11 month ago

          That’s true, but it’s still way more understandable that a car without lidar would be fooled by it. And there is no way you would ever come into such a situation, whereas the image in the thumbnail, could actually happen. That’s why it’s so misleading, can people not see that?
          I absolutely hate Elon Musk and support boycott of Tesla and Starlink, but this is a bit too misleading even with that in mind.

      • @Gonzako@lemmy.world
        link
        fedilink
        English
        391 month ago

        still, this should be something the car ought to take into account. What if there’s a glass in the way?

          • Victoria
            link
            fedilink
            English
            231 month ago

            Yes, but Styrofoam probably damages the car less than shards of glass.

            • snooggums
              link
              fedilink
              English
              71 month ago

              Glass is far more likely to cause injuries to the driver or the people around the set, just from being heavier material than styrofoam.

        • @Buffalox@lemmy.world
          link
          fedilink
          English
          41 month ago

          Yes, I think a human driver who isn’t half asleep would notice that something is weird, and would at least slow down.

  • Magnus
    link
    fedilink
    English
    821 month ago

    I remember Elon foolishly saying his cars don’t need radar or lidar. Even software-disabling radar in cars that already had the hardware.

    • @gmtom@lemmy.world
      link
      fedilink
      English
      191 month ago

      Not even just his cars, he thinks the MILITARY, doesn’t need radar and can just use cameras to spot and track stealth fighters.

      He’s a fucking lunatic.

      • @towerful@programming.dev
        link
        fedilink
        English
        31 month ago

        As an augmentation, the ability to spot and track objects visually would be amazing.
        But then planes just have to fly above 10k ft, and pretty much guaranteed cloud cover.

  • @Polderviking@feddit.nl
    link
    fedilink
    English
    131 month ago

    I remember elon saying something along the lines of his camera system being just as good and they thusly don’t need to employ things like LIDAR.

  • @Itsamelemmy@lemmy.zip
    link
    fedilink
    English
    29
    edit-2
    1 month ago

    Entire video is worth watching. He also snuck a chest mounted lidar into Disney and mapped some rides.

  • Mayor Poopington
    link
    fedilink
    English
    1081 month ago

    I read something a while back from a guy while wearing a T-shirt with a stop sign on it, a couple robotaxies stopped in front of him. It got me thinking you could cause some chaos walking around with a speed limit 65 shirt.

    • @audaxdreik@pawb.social
      link
      fedilink
      English
      471 month ago

      I think one of my favorite examples was using simple salt to trap them within the confines of white lines that they didn’t think they could cross over. I really appreciate the imagery of using salt circles to entrap the robotic demons …

    • @SpaceNoodle@lemmy.world
      link
      fedilink
      English
      61 month ago

      They’re not reading speed limit signs; they’ll follow the speed limit noted on the reference maps, like what you see in the app on your phone.

      • @kibiz0r@midwest.social
        link
        fedilink
        English
        8
        edit-2
        1 month ago

        Yikes, there’s a 25 around here that shows up as a 55 in Google Maps.

        Also a 55 that goes down to I think 35 for just a moment when it joins up with a side road. I wonder what a Tesla would do if it was following that data.

      • MrScottyTay
        link
        fedilink
        English
        301 month ago

        There’s a lot of cars that check via camera too to double check, for missing/outdated information and for temporary speed limit signs.

        • @SomeoneSomewhere@lemmy.nz
          link
          fedilink
          English
          11
          edit-2
          1 month ago

          Lots of places also have variable limit signs that get updated based on traffic, accidents etc.

          Here in NZ those seem to all be marked on the speed limit maps as 100km/h even if in some places the signs never go above 80.

          Ngauranga Gorge is one such location and I believe has the country’s highest grossing speed camera.

        • @Giooschi@lemmy.world
          link
          fedilink
          English
          81 month ago

          Where I live there are a lot of “temporary” 30km/h speed limits that were never removed by the road workers after the work was completed.

  • @Billybob22@feddit.uk
    link
    fedilink
    English
    -141 month ago

    Don’t want to rock the boat but apart from being a you tube money earner this doesn’t prove or disprove anything. A lot of humans would be fooled by this also.

    I am suspicious of the way the polystyrene wall broke in cartoon like shagged edges, almost like they were precut.

    • Subverb
      link
      fedilink
      English
      191 month ago

      The point of the test is to demonstrate that vision-only, which Tesla has adopted is inadequate. A car with lidar or radar would have been able to “see” that the car was approaching an obstacle without being fooled by the imagary.

      So yes, it seems a bit silly, but the underlying point is legitimate. If the software is fooled by this, then can you ever fully trust it? Especially when sensor systems exist that don’t have this problem at all. Would you want to be a pedestrian in a crosswalk with this car bearing down on you in FSD?

    • @EpeeGnome@lemm.ee
      link
      fedilink
      English
      41 month ago

      It may not rise to the level of proof, but it is a memorable and easily understood demonstration of something already proven by car safety researchers, as mentioned in the article.

      Why shouldn’t they precut the wall into cartoony shapes? It adds entertainment and doesn’t compromise the demonstration.

      • @Billybob22@feddit.uk
        link
        fedilink
        English
        21 month ago

        Yep agreed. Having used Teslas adaptive cruise control I wouldn’t ever use self driving, not that I have it, unless I had a death wish. Quite honestly my previous Chinese MG was a lot less likely to kill me.

        • @JcbAzPx@lemmy.world
          link
          fedilink
          English
          51 month ago

          They were expecting this result to be possible. What were they supposed to do? Slam the car into the side if a building?

      • @Billybob22@feddit.uk
        link
        fedilink
        English
        -21 month ago

        Yes but the main point that has been shown is that putting a screen up with the exact copy of the road and surroundings behind the screen is a daft and dangerous idea. It would be a better test if they had put up a polystyrene tree in the middle of the road and then checked if the car stopped.

        I have never driven through a polystyrene wall with a picture of a road on it in 40 years because people just don’t put those things up, they don’t grow on roads etc etc.

        Great YT clip for entertainment though.

          • @Billybob22@feddit.uk
            link
            fedilink
            English
            11 month ago

            I have never seen a mural on a road depicting a road that is identical to the road that I am driving on. Hope that helps.

            • @privatizetwiddle@lemmy.sdf.org
              link
              fedilink
              English
              11 month ago

              Maybe someone should do a follow up experiment to see how different the “mural” would have to be for the car to recognise it. A human would obviously not fall for something like an artistic picture of a fantasy land, but would a Tesla?

              • @Billybob22@feddit.uk
                link
                fedilink
                English
                11 month ago

                Ha well done.

                Not sure people will be using self drive around a car park but if that is the plan then I guess you Americans will have to white wash that kind of thing.

        • @chaogomu@lemmy.world
          link
          fedilink
          English
          51 month ago

          The other two tests that the Tesla failed were more realistic. Heavy fog and heavy rain.

          It failed both.

          If your self driving car cannot handle weather, then it’s not self driving at all.

          Actual lidar isn’t fooled by weather. Shitty optical only cameras are.

    • KayLeadfootOP
      link
      fedilink
      331 month ago

      It’s dirt cheap, too. If this was a cost-cutting measure, it was a thoroughly idiotic one. Which feels like the mark… of a certain someone I can think of

    • @bier@feddit.nl
      link
      fedilink
      English
      241 month ago

      In short because Elon (wrongly) believes you only need cameras, he made the claim people also drive with just 2 eyes.

      The thing is, we recognize a truck with stickers of a stopsign, while AI vision gets confused.

      Waymo (Googles self driving side hussle) was build on lidar and other sensors and has been using robot taxis for many years now in geofenced specific areas.

      • @Yoga@lemmy.ca
        link
        fedilink
        English
        121 month ago

        The thing is, we recognize a truck with stickers of a stopsign, while AI vision gets confused.

        Lmao would it be illegal to put a stop sign on the back of your car?

        • @CmdrShepard42@lemm.ee
          link
          fedilink
          English
          71 month ago

          I was thinking the same thing. What would happen if you popped one out of the back of your car while driving in front of a self driving car on the freeway?

        • @kelseybcool@lemmy.world
          link
          fedilink
          English
          21 month ago

          Some school buses have a sticker / sign on the back that says “I stop for railroad crossings” and can have a stop sign on said sticker.

      • ferret
        link
        fedilink
        English
        11 month ago

        The funny thing is, apparently our depth perception, a product of our two eyes, is a feature beyond the reach of tesla. And it would have allowed to to complete this test.

  • @CosmoNova@lemmy.world
    link
    fedilink
    English
    21 month ago

    Show someone footage of 9/11 and they‘ll think of 2001. Show someone footage of a burning or crashing Tesla 20 years from now and they‘ll think of 2025.