Any minute now, Elon Musk will reveal details of the biggest update to Tesla’s Autopilot suite since its release in October of 2015. He was supposed to do this last Wednesday, then, when that didn't come to pass, over the weekend. I’m sure he had bigger fish to fry, which gave me more time to digest the lessons of my most recent Tesla cross-country drive, and contemplate what we can expect to see in Autopilot 8. Without further ado:

Improved Situational Awareness Display Might this not mean better object recognition, with a display drawn to scale? More accurate icons, including pedestrians? How about making the secondary nav and power displays smaller, and the central display bigger?

Previous Next Close

Mode Confusion/Alerts/Involuntary Disengagements I’ve lumped mode confusion, alerts, and involuntary disengagements together because they clearly go hand-in-hand-in-hand. Mode confusion is blamed for 90 percent of all plane crashes attributed to human error. As good as Autopilot 7 is, its state of engagement feedback—that is, the user’s awareness of whether or not it’s actually engaged—is mediocre at best. But that's better than everyone else’s, which is non-existent. After 7,000+ miles of Autopilot 7 usage, I’ve only ever had one mode confusion incident, and that was when I thought the system was on when it wasn’t. Again, even though Autopilot 7 is state of the art, its state of engagement notification is limited to two pairs of pleasant sounding chimes (ascending or descending based on engagement or disengagement) and a little grey or green steering wheel to the right of the speedometer. Not good enough. I’m not suggesting Christmas lights here, but man, anything between Christmas lights and what we currently have would be safer. I'm not exaggerating when I say the entire dash should pulse green when Autopilot is engaged, and the phrase "Autopilot Engaged" should scroll across it every 20 seconds. Better yet, install a heads-up display and have that also pulse in the driver’s line of sight. Do those things—do anything—before the government mandates something we all hate. Oh, and when the system disengages, make the entire dash flash red and sound an unpleasant horn. Over and over. Shocking? It should shock. It will also make it less likely people will try to engage the system in less-than-ideal conditions. Is it unsafe to shock? It’s safer to shock once and have people remember it than it is to crash and have people remember that. (See number 3: Conditional Prohibitions.)

Previous Next Close

Re-engagement Prohibitions We’ve already seen a leak about this, and it’s exactly what I hope Musk will give us. If a driver ignores the disengagement warnings in Autopilot 7, the car will eventually come to a stop. Fine. What’s allegedly coming in 8 is that the user won’t be able to re-engage Autopilot until the car has been placed in Park, then back into Drive. I have a strong suspicion that most of the accidents blamed on Autopilot occurred because drivers kept trying to re-engage in less than ideal conditions after numerous hands-off warnings. This should solve that. I hope. This is similar to rules baked into Airbus aviation autopilot systems, which I will be addressing in an upcoming article. We need a lot more aviation autopilot logic pulled into automotive, and this is a huge (if unsexy) step in the right direction.

AutoSteer Feel/Torquing The Autosteer aspect of Autopilot puts too much torque into the steering. Call it too much of a good thing. Everyone else puts too little. I want to be able to change lanes within Autopilot, but currently, any steering input while engaged disengages Autopilot altogether. Therefore, in order to keep it engaged I often have to remove one hand from the wheel where, at rest, it might disengage the system. Loosen this up, just a little. Say, 10 percent.

Previous Next Close

Geo-Capping/Conditional Prohibitions/Fleet Learning This one seems obvious. Weather, darkness, and direct sunlight all have negative effects on Autopilot, greatly increasing the likelihood of an involuntary disengagement; therefore, Autopilot should not be engage-able in these conditions. I’ve engaged Autopilot in moderate rain at 90 mph. I’ve also chosen to dial it back to 30 mph for my own safety. It appeared to work perfectly—emphasis on "appeared to." The average person is unable to judge whether such conditions are suitable for Autopilot, and shouldn’t be allowed to use such a technology until it matures, and both driver education and product training improves. The myriad scenarios in which such limitations should be imposed would fill volumes, but Tesla’s Fleet Learning must be capable of correlating historical weather, light levels, and sun placement with disengagements, driver inputs, and accidents. Expect to see a lot more Fleet Learning data integration—if not in Autopilot 8, then in interim updates. That’s the whole point of Fleet Learning, and where Tesla has an advantage. For now.

Previous Next Close

Geo-Fencing This one also seems obvious, probably because it’s kin with number three. The highway on which Joshua Brown’s accident occurred had perpendicular access roads lacking 4-way lights or stop signs. Would capping Autopilot’s speed on that road have saved Brown’s life? Maybe not, but it certainly wouldn’t have hurt his chances of survival, and if Brown had been forced to take the wheel because he personally felt the need to speed, he’d have been more likely to avoid the accident. Likewise, Autopilot isn’t intended to work in Manhattan traffic, so why should it be possible to engage it within our snarled grid system? I consider myself a bit of an expert on Autopilot. Really, I love it, but that just means I know more about its limitations than the usual critics, most of whom have somewhere between zero and one hour behind the wheel. Until Autopilot sees lights and signs, it shouldn’t be possible to use it in Manhattan unless you’re on the FDR Drive or the West Side Highway north of 57th street. For the record, I did the entire FDR Drive last week on Autopilot, in traffic, with only two disengagements. It was brilliant. (Video coming soon.) Anywhere there’s a happy Tesla owner, there’s an area they know Autopilot shouldn’t be used, at least right now. Let’s see some more Fleet Learning applied, maybe specifically toward new owners. Ooops, I think I just opened a can of worms. Well, here’s a bigger one:

Previous Next Close