#TeslaCrash: Three reasons for Tesla (and all of us) to be concerned

Tesla has just disclosed the first fatal crash of a driver using its "Autopilot" system. Tesla should be concerned about the question of who's liable, and we should all be concerned about the wider consequences of this tragic event.

5 minute read

July 2, 2013, 9:17 AM PDT

By Scott Le Vine @http://www.twitter.com/scottericlevine

Tesla Motors disclosed yesterday the first fatality of a customer using its 'Autopilot' system of partial self-driving automotive technology.  This is thought to be the first death attributable to a Highly-Automated Car, so is something of a watershed moment: the Age of Innocence is over.  The facts are murky at present and will emerge in due course.  Initial reports are that the driver may have been watching a (Harry Potter) video at the time of the crash, in flagrant violation of the agreement between him and Tesla.  Tesla maintains that "you need to maintain control and responsibility for your vehicle while using [Autopilot]", and has given no indication yet of whether or not it believes that it is liable in any way for the crash.

This crash is very sad news for the deceased's family, with potentially much wider ramifications as well.

Last week my team and I released a new (long: 20K words) Working Paper on the liability issues raised by Automated Cars, and how this will constrain how they drive.  I want to therefore share initial thoughts on this incident, starting with specific reasons for Tesla to be concerned, and concluding with broader reasons for the rest of us to also pay attention:

Three reasons that Tesla should be concerned:

  1. The reaction time of Automated Cars' control systems in pre-crash situations are in principle much faster than human drivers' reaction times.  This might mean that the system's designer (Tesla) could be vulnerable under the 'Last Clear Chance' legal doctrine, because Autopilot's superior reaction time might be interpreted to mean that it had the Last Clear Chance (i.e. later than the human drivers of both vehicles involved) to avoid the crash.  The 'Last Clear Chance' doctrine means that a human driver may bear some negligence in a crash even if they were not violating the Rules of the Road, if they had the 'Last Clear Chance' to avoid the crash.
  2. Even car passengers may have a 'Duty to Remonstrate' (i.e. raise objections) when a car driver is driving badly, otherwise they might be found partly negligent in the event of a crash.  Is the relationship between Tesla's Autopilot System and the driver using it not a stronger one than the relationship between a car driver and a passenger?  It would seem to be a stronger relationship in multiple, demonstrable ways.
  3. Courts have also held that car passengers may be partially liable under the 'Joint Enterprise' doctrine, if "there is an understanding or an agreement in advance between the driver and the passenger that the passenger has a right to tell the driver how to drive the automobile" (see Siruta v. Siruta, 348 P.3d 549 (Kan. 2015), which discusses "[shared] negligence as a result of joint driving decisions").  While it's clear that the human driver of the Tesla was negligent, these legal doctrines emanating originally from the traditional car-driver/car-passenger relationship may provide precedent for joint negligence between a driver and system-designer that are both concurrently making driving decisions.

Three reasons for the rest of us to be concerned:

  1. Tesla has been less than forthright here, taking nearly two months to inform its customers and the general public (and its shareholders) of an incident that it must have realized would be of wide interest.  It has not yet disclosed any remedy that it took in the interim to prevent a recurrence, such as, perhaps temporarily disabling or restricting the Autopilot system on its cars until the facts are established.  This period of delay cannot be an oversight, and violates the fundamental crisis-management principle of getting bad news out quickly and fully.  Designers of Automated Cars will not be able to maintain the public's trust if this behavior becomes habitual; openness and transparency are basic requirements for an orderly rollout of the technology.
  2. It was a fundamental sensing failure (apparently the failure to distinguish a left-turning tractor trailer from background scenery) that prevented the Autopilot system from intervening to prevent the crash (or at least to mitigate it: no braking was performed by either human or Autopilot).  This highlights that much basic Research & Development remains to be done; the technology is not "ready" yet.  It also makes plain that while Automation will eliminate many of the 94% of crashes that are directly attributable to human error, we will need to contend with new crash scenarios caused by system failures.
  3. Automated Cars will probably make our roads safer overall – we expect this, but can’t yet be 100% sure.  However, the tort system may well impose costs-per-crash on Automated Cars’ manufacturers that are much higher than costs-per-crashes of human drivers, which could reduce overall safety if this has a deterrence effect that inappropriately delays the rollout of the technology.  Automated Cars will rigorously follow the rules laid out by their designers – and saying ‘sorry I didn't mean it’ won't be a viable option, because the behavior is programmed in, not a reaction to a ‘sudden emergency’ as with humans’ crashes.  Designers of Automated Cars will have deep pockets, so will be also much more attractive targets for litigation than hapless human drivers, who have far fewer assets to pursue.  Finally, the corporate designers of Automated Cars making deliberative choices will probably be seen by juries as less sympathetic than us hapless human drivers who might ‘make honest mistakes’ in the 'heat of the moment', and therefore juries might feel emboldened to impose larger awards for actual and punitive damages. 

The risk is that we lose sight of the forest (overall safety benefits), owing to the trees (individually tragic incidents of which this is the first).

Scott Le Vine

Scott Le Vine, AICP is an Assistant Professor (Urban Planning) at the State University of New York (New Paltz), a Research Associate at Imperial College London, and a Visiting Professor at Southwest Jiaotong University (Chengdu, China).

Aerial view of snowy single-family homes in suburban Long Island, New York

New York Governor Advances Housing Plan Amid Stiff Suburban Opposition

Governor Kathy Hochul’s ambitious proposal to create more housing has once again run into a brick wall of opposition in New York’s enormous suburbs, especially on Long Island. This year, however, the wall may have some cracks.

March 20, 2023 - Mark H. McNulty

Empty parking garage at night with yellow lines marking spots and fluorescent lighting

Rethinking the Role of Parking in the American City

In cities big and small, the tide is turning against sprawling parking lots, car-centric development, and minimum parking mandates.

March 16, 2023 - The New York Times

A futuristic version of New York City, with plants growing neatly on top of modern skycrapers.

Friday Eye Candy: 20 AI-Generated Cityscapes

AI-generated images are creating new landscapes and cityscapes, capable of inspiring awe or fear.

March 17, 2023 - Chris Steins via Medium

A group of wetsuit-clad swimmers gathers to talk in shallow water near the shore of the San Francisco Bay.

Proposed Pool Would Make an Olympic-Sized Play Area in the San Francisco Bay

The San Francisco Bay is usually an undesirable place to swim, except for a hearty few. A development proposal seeking assistance at the state level would add a pool to the Bay’s waters to make the idea of going for a swim more appealing.

March 24 - The Mercury News

Chicago elevated train over busy city street surrounded by high-rise buildings

Chicagoland Transit Agencies Call for State Funding as Budget Shortfall Looms

Illinois transit agencies want to see changes to a law requiring them to collect half of their revenue from transit fares, arguing that low ridership and staffing shortages will lead to a massive budget gap without intervention.

March 24 - Crain's Chicago Business

Minneapolis Stone Arch Bridge

Panel: Minneapolis Zoning Updates Should Reflect Mixed-Use Future

A discussion of post-pandemic changes in work and commuting concluded that the city’s overhaul of its zoning code should be less restrictive with land uses.

March 24 - MinnPost

New Updates on PD&R Edge

HUD's Office of Policy Development and Research

HUD’s 2023 Innovative Housing Showcase

HUD's Office of Policy Development and Research

Write for Planetizen

Urban Design for Planners 1: Software Tools

This six-course series explores essential urban design concepts using open source software and equips planners with the tools they need to participate fully in the urban design process.

Planning for Universal Design

Learn the tools for implementing Universal Design in planning regulations.