You'd think that self-driving cars would be most vulnerable to remote hacks
, but the biggest danger may come from someone nearby with a handful of cheap electronics. Security researcher Jonathan Petit has determined
that you can fool LIDAR
(the laser ranging common on autonomous vehicles
) by sending "echoes" of fake cars and other objects through laser pulses. All you need is a low-power laser, a basic computing device (an Arduino kit
or Raspberry Pi
is enough) and the right timing – you don't even need good aim. Petit managed to spoof objects from as far as 330 feet away in his proof-of-concept attack, and he notes that it's possible to present multiple copies of these imaginary objects or make them move. In other words, it'd only take one prankster to make a self-driving car
swerve or stop to avoid a non-existent threat.
There's no guarantee that this will be a major issue if and when self-driving cars become commonplace. Petit's technique only works so long as LIDAR units' pulses aren't encrypted or otherwise obscured. While that's true of many commercial systems at the moment, it's possible that production-ready vehicles will lock things down. Still, this is a not-so-friendly reminder that car makers have a lot of work ahead of them if they're going to secure their robotic rides.
This article by Jon Fingas originally ran on Engadget, the definitive guide to this connected life.