When self-driving cars need help: Stanford study’s surprising finding

Share this:

FILE - This May 13, 2015 file photo shows the front of Google's new self-driving prototype car during a demonstration at the Google campus in Mountain View, Calif. New cars that can steer and brake themselves may lull drivers into a false sense of security. One way to keep people alert may be providing distractions that are now illegal, just one surprising finding from Stanford University research that studied the behavior of students in a self-driving car simulator.(AP Photo/Tony Avelar, File)

FILE - This Friday, Nov. 13, 2015 file photo provided by Virginia Tech shows Virginia Tech Center for Technology Development Program Administration Specialist Greg Brown behind the wheel of a driverless car during a test ride showing the alert system handing over automation to the driver while traveling street in Blacksburg, Va. New cars that can steer and brake themselves may lull drivers into a false sense of security. One way to keep people alert may be providing distractions that are now illegal, just one surprising finding from Stanford University research that studied the behavior of students in a self-driving car simulator.(Justin Fine/Virginia Tech via AP, File)

LOS ANGELES — New cars that can steer and brake themselves risk lulling people in the driver’s seat into a false sense of security — and even to sleep. One way to keep people alert may be providing distractions that are now illegal.

That was one surprising finding when researchers put Stanford University students in a simulated self-driving car to study how they reacted when their robo-chauffer needed help.

The experiment was one in a growing number that assesses how cars can safely hand control back to a person when their self-driving software and sensors are overwhelmed or overmatched. With some models already able to stay in their lane or keep a safe distance from other traffic, and automakers pushing for more automation, the car-to-driver handoff is a big open question.

The elimination of distracted driving is a major selling point for the technology. But in the Stanford experiment, reading or watching a movie helped keep participants awake.

Among the 48 students put in the driver’s seat, 13 who were instructed to monitor the car and road began to nod off. Only three did so when told to watch a video or read from a tablet.

Alertness mattered when students needed to grab the wheel because a simulated car or pedestrian got in the way.

There’s no consensus on the right car-to-driver handoff approach: the Stanford research suggests engaging people with media could help, while some automakers are marketing vehicles with limited self-driving features that will slow the car if they detect a person has stopped paying attention to the road.

Though research is ongoing, it appears that people need at least 5 seconds to take over — if they’re not totally checked out.

Self-driving car experts at Google, which is pursuing the technology more aggressively than any automaker, concluded that involving humans would make its cars less safe. Google’s solution is a prototype with no steering wheel or pedals — human control would be limited to go and stop buttons.

Meanwhile, traditional automakers are phasing in the technology. Mercedes and Toyota sell cars that can hit the brakes and stay in their lane. By adding new features each year, they might produce a truly self-driving car in about a decade.

One potential hazard of this gradualist approach became clear this fall, when Tesla Motors had to explain that its “auto pilot” feature did not mean drivers could stop paying attention. Several videos posted online showed people recording the novelty — then seizing the wheel when the car made a startling move.

Starting late next year, the Cadillac CTS will get a Super Cruise system, which will allow semi-autonomous highway driving. If the driver’s eyes are off the road, and they don’t respond to repeated prodding, the car slows.

“We are in no way selling this as a technology where the driver can check out,” General Motors spokesman Dan Flores said. “You can relax, glance away, but you still have to be aware because you know the technology’s not foolproof.”

One riddle automakers must solve: How to get owners to trust the technology so they’ll use it — but not trust it so much that they’ll be lulled into a false sense of security and therefore slower to react when the car needs them.

Trust was on the mind of researchers who in August published an extensive report on self-driving cars funded by the National Highway Traffic Safety Administration. “Although this trust is essential for widespread adoption, participants were also observed prioritizing non-driving activities over the operation of the vehicle,” the authors wrote.

Another wide-open question: How to alert the person in the driver’s seat of the need to take over.

It appears that the car should appeal to several senses. Visual warnings alone may not suffice. But combine a light with spoken instructions or physical stimulation such as a vibrating seat, and people react more quickly.

“If it is done courteously and subtle and not annoying, it could be missed by someone that is distracted,” said Greg Fitch, a research scientist at the Virginia Tech Transportation Institute. Then again, the way the car interacts with people will be one way automakers differentiate their product — and overbearing warnings may sour potential buyers.

Other issues Fitch cites include “mode confusion” (making sure the car clearly informs the person whether or not it is driving itself) and clear explanations to drivers of what the car can — and cannot — handle.

Cars with the right sensors are becoming really good at monitoring the outside world and have quicker response times than humans. People are much better at making decisions during uncertain circumstances.

One lesson from the Stanford study may be that master and machine are better viewed as collaborators.

“There’s really a relationship between drivers and cars,” said David Sirkin, who helped run the experiment at Stanford’s Center for Design Research, “and that relationship is becoming more a peer relationship.”

A data analysis firm employed by President Donald Trump's 2016 campaign tapped the Facebook profiles of more than 50 million users without their permission, allowing it to capitalize on the private social media activity of a large portion of the U.S. electorate, newspapers reported Saturday.