By Tom Krisher

The U.S. government has opened a formal investigation into Tesla's Autopilot partially automated driving system after a series of collisions with parked emergency vehicles.

The investigation covers 765,000 vehicles, almost everything that Tesla has sold in the U.S. since the start of the 2014 model year. Of the crashes identified by the National Highway Traffic Safety Administration as part of the probe, 17 people were injured and one was killed.

NHTSA says it has identified 11 crashes since 2018 in which Teslas on Autopilot or Traffic Aware Cruise Control have hit vehicles at scenes where first responders have used flashing lights, flares, an illuminated arrow board or cones warning of hazards. The agency announced the action Monday in a posting on its website.

The probe is another sign that NHTSA under President Joe Biden is taking a tougher stance on on automated vehicle safety than under previous administrations. Previously the agency was reluctant to regulate the new technology for fear of hampering adoption of the potentially life-saving systems.

The investigation covers Tesla's entire current model lineup, the Models Y, X, S and 3 from the 2014 through 2021 model years.

The National Transportation Safety Board, which also has investigated some of the Tesla crashes dating to 2016, has recommended that NHTSA and Tesla limit Autopilot’s use to areas where it can safely operate. The NTSB also recommended that NHTSA require Tesla to have a better system to make sure drivers are paying attention. NHTSA has not taken action on any of the recommendations. The NTSB has no enforcement powers and can only make recommendations to other federal agencies.

Last year the NTSB blamed Tesla, drivers and lax regulation by NHTSA for two collisions in which Teslas crashed beneath crossing tractor-trailers. The NTSB took the unusual step of accusing NHTSA of contributing to the crash for failing to make sure automakers put safeguards in place to limit use of electronic driving systems.

The agency made the determinations after investigating a 2019 crash in Delray Beach, Florida, in which the 50-year-old driver of a Tesla Model 3 was killed. The car was driving on Autopilot when neither the driver nor the Autopilot system braked or tried to avoid a tractor-trailer crossing in its path.

“We are glad to see NHTSA finally acknowledge our long standing call to investigate Tesla for putting technology on the road that will be foreseeably misused in a way that is leading to crashes, injuries, and deaths,” said Jason Levine, executive director of the nonprofit Center for Auto Safety, an advocacy group. “If anything, this probe needs to go far beyond crashes involving first responder vehicles because the danger is to all drivers, passengers, and pedestrians when Autopilot is engaged.”

Autopilot has frequently been misused by Tesla drivers, who have been caught driving drunk or even riding in the back seat while a car rolled down a California highway.

A message was left early Monday seeking comment from Tesla, which has disbanded its media relations office.

NHTSA has sent investigative teams to 31 crashes involving partially automated driver assist systems since June of 2016. Such systems can keep a vehicle centered in its lane and a safe distance from vehicles in front of it. Of those crashes, 25 involved Tesla Autopilot in which 10 deaths were reported, according to data released by the agency.

Tesla and other manufacturers warn that drivers using the systems must be ready to intervene at all times. In addition to crossing semis, Teslas using Autopilot have crashed into stopped emergency vehicles and a roadway barrier.

The probe by NHTSA is long overdue, said Raj Rajkumar, an electrical and computer engineering professor at Carnegie Mellon University who studies automated vehicles.

Tesla's failure to effectively monitor drivers to make sure they're paying attention should be the top priority in the probe, Rajkumar said. Teslas detect pressure on the steering wheel to make sure drivers are engaged, but drivers often fool the system.

“It's very easy to bypass the steering pressure thing,” Rajkumar said. “It's been going on since 2014. We have been discussing this for a long time now.”

The crashes into emergency vehicles cited by NHTSA began on Jan. 22, 2018 in Culver City, California, near Los Angeles when a Tesla using Autopilot struck a parked firetruck that was partially in the travel lanes with its lights flashing. Crews were handling another crash at the time.

Since then, the agency said there were crashes in Laguna Beach, California; Norwalk, Connecticut; Cloverdale, Indiana; West Bridgewater, Massachusetts; Cochise County, Arizona; Charlotte, North Carolina; Montgomery County, Texas; Lansing, Michigan; and Miami, Florida.

“The investigation will assess the technologies and methods used to monitor, assist and enforce the driver's engagement with the dynamic driving task during Autopilot operation,” NHTSA said in its investigation documents.

In addition, the probe will cover object and event detection by the system, as well as where it is allowed to operate. NHTSA says it will examine “contributing circumstances” to the crashes, as well as similar crashes.

An investigation could lead to a recall or other enforcement action by NHTSA.

“NHTSA reminds the public that no commercially available motor vehicles today are capable of driving themselves,” the agency said in a statement. “Every available vehicle requires a human driver to be in control at all times, and all state laws hold human drivers responsible for operation of their vehicles.”

The agency said it has “robust enforcement tools” to protect the public and investigate potential safety issues, and it will act when it finds evidence “of noncompliance or an unreasonable risk to safety.”

In June NHTSA ordered all automakers to report any crashes involving fully autonomous vehicles or partially automated driver assist systems.

Shares of Tesla Inc., based in Palo Alto, California, fell 3.5% at the opening bell Monday.

Tesla uses a camera-based system, a lot of computing power, and sometimes radar to spot obstacles, determine what they are, and then decide what the vehicles should do. But Carnegie Mellon's Rajkumar said the company’s radar was plagued by “false positive” signals and would stop cars after determining overpasses were obstacles.

Now Tesla has eliminated radar in favor of cameras and thousands of images that the computer neural network uses to determine if there are objects in the way. The system, he said, does a very good job on most objects that would be seen in the real world. But it has had trouble with parked emergency vehicles and perpendicular trucks in its path.

“It can only find patterns that it has been ‘quote unquote’ trained on,” Rajkumar said. “Clearly the inputs that the neural network was trained on just do not contain enough images. They’re only as good as the inputs and training. Almost by definition, the training will never be good enough.”

Tesla also is allowing selected owners to test what it calls a “full self-driving” system. Rajkumar said that should be investigated as well.

Updated on August 16, 2021, at 11:44 a.m. ET with the latest details.

Share:
More In Business
New York Times, after Trump post, says it won’t be deterred from writing about his health
The New York Times and President Donald Trump are fighting again. The news outlet said Wednesday it won't be deterred by Trump's “false and inflammatory language” from writing about the 79-year-old president's health. The Times has done a handful of stories on that topic recently, including an opinion column that said Trump is “starting to give President Joe Biden vibes.” In a Truth Social post, Trump said it might be treasonous for outlets like the Times to do “FAKE” reports about his health and "we should do something about it.” The Republican president already has a pending lawsuit against the newspaper for its past reports on his finances.
OpenAI names Slack CEO Dresser as first chief of revenue
OpenAI has appointed Slack CEO Denise Dresser as its first chief of revenue. Dresser will oversee global revenue strategy and help businesses integrate AI into daily operations. OpenAI CEO Sam Altman recently emphasized improving ChatGPT, which now has over 800 million weekly users. Despite its success, OpenAI faces competition from companies like Google and concerns about profitability. The company earns money from premium ChatGPT subscriptions but hasn't ventured into advertising. Altman had recently announced delays in developing new products like AI agents and a personal assistant.
Trump approves sale of more advanced Nvidia computer chips used in AI to China
President Donald Trump says he will allow Nvidia to sell its H200 computer chip used in the development of artificial intelligence to “approved customers” in China. Trump said Monday on his social media site that he had informed China’s leader Xi Jinping and “President Xi responded positively!” There had been concerns about allowing advanced computer chips into China as it could help them to compete against the U.S. in building out AI capabilities. But there has also been a desire to develop the AI ecosystem with American companies such as chipmaker Nvidia.
Trump says Netflix deal to buy Warner Bros. ‘could be a problem’ because of size of market share
President Donald Trump says a deal struck by Netflix last week to buy Warner Bros. Discovery “could be a problem” because of the size of the combined market share. The Republican president says he will be involved in the decision about whether federal regulators should approve the deal. Trump commented Sunday when he was asked about the deal as he walked the red carpet at the Kennedy Center Honors. The $72 billion deal would bring together two of the biggest players in television and film and potentially reshape the entertainment industry.
What to know about changes to Disney parks’ disability policies
Disney's changes to a program for disabled visitors are facing challenges in federal court and through a shareholder proposal. The Disability Access Service program, which allows disabled visitors to skip long lines, was overhauled last year. Disney now mostly limits the program to those with developmental disabilities like autism who have difficulty waiting in lines. The changes have sparked criticism from some disability advocates. A shareholder proposal submitted by disability advocates calls for an independent review of Disney's disability policies. Disney plans to block this proposal, claiming it's misleading. It's the latest struggle by Disney to accommodate disabled visitors while stopping past abuses by some theme park guests.
Load More