WASHINGTON (Reuters) - The U.S. National Transportation Safety Board said it was sending a team to Las Vegas to investigate a collision this week between a self-driving shuttle bus and a truck that was blamed on human error.
The board, which has the power to issue safety recommendations and determines probable causes of crashes, said four investigators were expected to arrive on Friday.
The Navya Arma, an autonomous and electric vehicle operated by Keolis North America, went into service on Wednesday. A few hours later, a delivery truck backed into the stopped shuttle, according to a reporter on the shuttle and one of its sponsor companies.
Las Vegas police issued the truck driver a ticket, the city government said in a blog post. The shuttle sustained minor damage.
"The shuttle did what it was supposed to do, in that its sensors registered the truck and the shuttle stopped," the city said.
The Automobile Association of America (AAA) of Southern Nevada, one of the shuttle's sponsors, said it would assist the safety board's investigation.
"Working together and sharing information will ensure this new technology is safely implemented for the public, and that's AAA's top priority," the organization said in a statement.
The shuttle also is sponsored by the city of Las Vegas, Keolis North America and the Regional Transportation Commission of Southern Nevada.
Reporter Jeff Zurschmeide, who was on the shuttle at the time of the crash, said the self-driving vehicle did what it was programmed to do but not everything a human driver might do.
"That's a critical point," Zurschmeide wrote on digitaltrends.com. "We had about 20 feet of empty street behind us (I looked), and most human drivers would have thrown the car into reverse and used some of that space to get away from the truck. Or at least leaned on the horn and made our presence harder to miss."
The crash follows a rising number of incidents involving human drivers behaving improperly or recklessly around self-driving cars. There have been 12 crashes in California alone since Sept. 8 involving General Motors Co's self-driving unit, Cruise Automation. All were the fault of human drivers in other vehicles, GM told regulators.
NTSB investigated a May 2016 crash of a Tesla Inc Model S that killed a driver using the vehicle's semi-autonomous "Autopilot" system. The board recommended in September auto safety regulators and automakers take steps to ensure that semi-autonomous systems were not misused.