Self-driving cars are prone to hacks -- and automakers are barely talking about it

BI Graphics Autonomous World banner post size

Today’s self-driving cars rely on spinning sensors called lidar that can cost more than $10,000 each. But it took Jonathan Petit just $43 and a laser pointer to confuse and defeat them.

“Anybody can go online and get access to this, buy it really quickly, and just assemble it, and there you go, you have a device that can spoof lidar,” Petit, a cybersecurity expert, told Business Insider.

Google, Tesla, and major automakers are racing to build fully autonomous cars, creating a future where many won’t need to own a vehicle, the young, old and disabled can get around more easily, and transforming the way we live. One day they could dramatically reduce the roughly 30,000 annual deaths from crashes.

But until we get there, carmakers have to ensure clever hackers — and those less benevolent than Petit — can’t cause the cars to go haywire.

Tricking the sensors

Ford LiDARScreenshot via YouTubeTwo lidar systems on top of a Ford.

When Petit was growing up in France in the 1980s, cars were simpler machines, disconnected from the outside world.

Petit’s parents owned a restaurant in France, and whenever they had a good season, his dad would use the opportunity to buy a car. One was a Citroen DS that bounced on its hydraulic suspension, a technologically advanced feature at the time.

“When you think about the old times, there’s nostalgia about it,” he said. “You always think about, ‘aw yeah that was a cool time and you can feel the road.’ Yeah, I think that was nice. I really loved that car.”

Despite growing up around cars, it wasn’t until much later that Petit started devoting all his time to making cars resistant to malicious actors.

Petit began extensively studying automotive cybersecurity as a PhD student at Paul Sabatier University in France in 2007. However, it was during his post-doc research at UC Berkeley, working with its Partners for Advanced Transportation Technology, that he became more interested in the hacking risks for self-driving cars, specifically.

Google carBusiness InsiderGoogle’s self-driving car.

Then, in 2015, two hackers — Charlie Miller and Chris Valasek — took control of a Jeep Cherokee’s UConnect system — an Internet-connected computer feature that controls everything from your ability to make calls to the car’s navigation system.

From a couch 10 miles west of the highway, the two guys were able to toy with the car’s air conditioning, blast the radio, activate the windshield wipers, and ultimately cut its transmission. Fiat Chrysler Automotive recalled 1.4 million vehicles to install anti-hacking software following the demo.

As Petit puts it, the demonstration highlighted the importance of automotive security. Now, hackers could gain access without even leaving the couch. Prior hacking demonstrations required researchers to be connected directly to the car’s dashboard.

“When they did the hack remotely that was like, ‘wow that’s interesting.’ Now it’s not just looking at having physical access [to the car,]” he said. “It’s scary when you start to have remote attacks.”

When Petit performed the attack on the lidar, he became one of the first researchers to show how easy it is to hack self-driving cars’ sensors. He was able to trick a sensor into thinking objects were there when they weren’t, and vice versa.

“So here, you can think that the potential consequence of an attack like this could be, I tried to crash you into a vehicle ahead of you because I’m telling you there is no object here,” he said. “So I’m making you blind and now your system thinks it’s free.”

But that kind of hack can have other consequences too. The car could see an obstacle that isn’t there and change lanes to get away from it. That manoeuvre, designed to keep passengers safe, could disrupt traffic. It could also cause the car to go off course.

“So now you’ve changed the path of the vehicle by doing this, that’s also an impact, which means that then the risk could be I’m sending you to small street to stop you and rob you or steal the car,” he said.

Uber self-driving carUberA self-driving Uber used for the Pittsburgh pilot.

Petit not only tricked the lidar system self-driving cars use, but was able to blind the cameras they rely on by using different LED lights. If the car feels it can no longer operate safely because its cameras have been disabled, it could stop entirely, leading to those same kind of problems.

Now, it’s important to take these scenarios with a grain of salt. As Petit said, self-driving cars are built with redundant sensor systems, meaning they have multiple cameras and sensors in case one were to fail.

For example, the self-driving cars Uber is using for its Pittsburgh pilot have 20 cameras and several radar sensors to provide 360-degree coverage.

That means even if a hacker compromises one or even a few sensors on a self-driving car, the car may still be able to pull enough information from the ones that are operating effectively to continue driving safely.

But it also depends on when an attack occurs. For example, a self-driving car at night might be programmed to rely more on its lidar system since the cameras can’t see as well in the dark, Petit said. If a hacker were to then spoof the lidar, it doesn’t have as much data to fall back on, and it could put the car and its passengers in a dangerous situation.

“Even if you’re thinking with just my sensors, I’m secure. This is not true,” he said.

Exploiting communication channels

Petit has conducted other research highlighting how vulnerable self-driving cars are to hacks even beyond sensor vulnerabilities.

In 2011, when Petit was a senior researcher at University of Twente in the Netherlands, he set up equipment that could pick up the signals cars were sending each other and send them to a laptop. These “sniffing stations” were able to locate a security vehicle within a given residential or business zone on campus with 78% accuracy. Petit could then narrow that down to individual roads with 40% accuracy.

Those sniffing stations were able to track cars by taking advantage of vehicle-to-vehicle (V2V) communication.

V2V communication is something automakers are already starting to use using in cars today, like the 2017 Mercedes E-Class. The communication channel allows the cars to talk to other cars on the road to relay data on traffic flow, accidents ahead, or poor weather. It can then be used to send alerts to the driver so she can change her course if things look bad up ahead.

Some automakers are exploring using V2V for self-driving cars because the cars can use the data to navigate more safely without relying on their sensors exclusively to see obstacles, like a traffic jam, with their sensors.

The cars won’t send personally identifying information, but the data, like GPS locations, are sent to other vehicles unencrypted.

But as Petit showed by setting up sniffing stations, hackers could track the data being sent to other vehicles to see their whereabouts.

Petit said he spent roughly $500 on the sniffing stations he used to track the security vehicles around campus. But he said the price of the equipment is rapidly dropping.

For this experiment, Petit only set up two sniffing stations at two busy intersections where they were small enough to sit undetected. Naturally, adding more sniffing stations to different areas would improve accuracy. But even with limited information, Petit could track a security guard’s whereabouts in real-time.

Self-driving cars could rely on these types of communication channels, meaning if not secured properly, hackers could easily track the locations of passengers.

The government is actually proposing that all new cars and light trucks be built with V2V communication abilities.

“Privacy is also protected in V2V safety transmissions. V2V technology does not involve the exchange of information linked to or, as a practical matter, linkable to an individual, and the rule would require extensive privacy and security controls in any V2V devices,” the National Highway Traffic Safety Administration wrote in a press release about the proposal.

Audi also recently implemented the first vehicle-to-infrastructure (V2I) system, where some Audi cars can talk to traffic lights to see when the light will change.

Justin Cappos, a systems and security professor at NYU who specialises in automotive cybersecurity, also told Business Insider he worries about the risks associated with relying on V2V and V2I communication channels.

“Any time you open a new communication channel you raise the risk that bad guys could use that to get in. Both those scare me a lot,” Cappos told Business Insider.

Cappos is part of an NYU project that was awarded $1.4 million from the Department of Homeland Security “for the development of technology that can help defend government and privately owned vehicles from cyber-attacks.”

He said cars today already have “ten to a dozen different channels” the hackers can exploit. But the worrisome bit is that the number of channels are growing and are being connected to “fairly complicated parts of machinery in the car” over time.

“When you have a lot of complexity then it does certainly have the potential to affect the likelihood of being hacked,” he said. “As long as people design them well, which is often not the case, it’s easier to secure something.”

Finding solutions: ‘A tricky line’

San Francisco Uber self-drivingUberAn Uber self-driving vehicle climbs a hill in San Francisco.

Securing self-driving cars comes at a price, and it’s a matter of how much automakers are willing to pay.

From a hardware perspective, automakers can add more sensors so that if one were compromised, there are others to take over, Petit said.

But most automakers are looking to reduce trim down on redundant sensor systems to cut down on cost.

Uber is eventually looking to use Volvo XC90 cars integrated with self-driving tech instead of the Ford Fusions currently being used in Pittsburgh. (Uber launched a pilot program that uses the Volvo XC90s in San Francisco this week, but could face legal action from the DMV for not obtaining an autonomous car permit for them.)

Eric Meyhofer, the engineering lead at Uber ATC, said at the Pittsburgh pilot media event that the Volvo XC90 will come with fewer cameras.

“The system functionally is as good or better in every regard, but we make it smaller,” he said of the Volvo. “It has smaller lasers, it’s the next generation of laser. It has fewer cameras, but not less functionality.”

Uber declined to comment for this story.

Petit said that when it comes down to tough decisions, carmakers generally choose to save money.

“In the automotive space, just 10 cents off a dollar is kind of like a no go,” Petit said. “It’s a tricky line, here.”

Cappos said we’re already seeing the tension between cost and security play out today, as was the case when Miller and Valasek were able to exploit UConnect to control the car’s internal computer network.

The Jeep hack is far from the only example of vulnerabilities in connected cars, but Cappos’ point is that automakers are already adding more computers to control different features in cars without thinking critically about securing the network.

“Slowly over time we’ve added more and more computers with not enough attention made to security and what all that means,” Cappos said. “In many cases [automakers] made decisions that minimized the wiring and the cost this would incur.”

“I think what happened was they weren’t as worried about it. They were worried about other things like cost,” he continued.

Petit also said he fears security is being seen as an afterthought by automakers.

“I indeed have the unfortunate feeling that they look at security as an add on, which is a problem,” he said. “You should use this opportunity to have security by design and not doing it after thought.”

In addition to Uber, Lyft, Fiat Chrysler, Ford, Daimler, and BMW declined to comment on their cybersecurity measures for self-driving cars, specifically. Tesla did not reply to multiple requests for comments.

Some did provide details on how they’re addressing security in cars today that could have implications for self-driving cars of the future.

A BMW spokesperson wrote in a statement to Business Insider that communication between vehicles and the outside world “always takes place by means of a backend” operated by BMW to ensure privacy.

A Ford spokesperson wrote in a statement that its hardware has “built-in firewall and “white-listed” functions to separate its entertainment systems from the vehicle controls. That would reduce the likelihood of an attack similar to the Jeep Cherokee one.

An FCA spokesperson wrote in an email that it has a bug bounty program to provide financial rewards when hackers disclose vulnerabilities.

Tom Wilkinson, GM’s communications manager for cybersecurity and safety, said in a phone interview that GM is investing heavily in making its “connected ecosystem” safer to secure its future self-driving cars.

“I don’t think cybersecurity becomes a concern only when you have autonomous vehicles or automated driving controls,” he said. “We look at these as an extension of what we’re doing in the connected car.”

Wilkinson said the cybersecurity team for connected car products consist of 80 people ranging from mathematicians to cryptologist. GM also works with HackerOne to provide a bug bounty program for third-party hackers that find vulnerabilities.

Self-driving cars are still in their early days, meaning there’s still plenty of time for automakers to ensure the cars are secure by design. But players in the self-driving car space will need to consider these risks early to avoid a situation like the Jeep hack, where the cars are already on the road.

Some steps have already been taken.

The National Highway Traffic and Safety Administration released cybersecurity guidance to make automobiles more secure at the end of October.

Additionally, The Automotive Information Sharing and Analysis Center (Auto ISAC) was created in July 2015 so automakers can share cyber threats to address vulnerabilities more quickly.

But Petit said more needs to be done to make cars more secure in the design phase.

“It’s a pity that the security community has the feeling that you need to be the bad guy to force them to wake up,” Petit said. “I cannot say that I have cracked the formula to get them to do it.”

NOW WATCH: Tesla reveals how your self-driving car sees the road

NOW WATCH: Tech Insider videos

Want to read a more in-depth view on the trends influencing Australian business and the global economy? BI / Research is designed to help executives and industry leaders understand the major challenges and opportunities for industry, technology, strategy and the economy in the future. Sign up for free at