December 20, 2012
JUSTIA.COM: As our world becomes more and more automated, with devices that do everything from recording video, to turning off appliances, to telling us where to go (and even taking us there, as in the case of Google’s self-driving cars), the ethical and legal question of “Who is responsible if something goes wrong?” becomes increasingly important.
Traditionally, in basic situations within the legal field of torts, or personal injury, the identity of a wrongdoer is often clear, or at least readily ascertainable. If Debbie trips Peter, then Debbie is liable to Peter for his injuries. If Acme Ladder Co. manufactures and sells a ladder that collapses when Pamela is on it, then Acme is legally responsible for Pamela’s resulting injuries. In these clear-cut cases, the ethical duty is essentially the same as the legal one.
However, there are now increasing numbers of situations in which the responsible party is less clearly defined. I will explore two devices that create such situations: autonomous cars and robots that assist with surgery. With the advent of self-driving cars and surgical robots, the answer to the question of who, exactly, would be responsible (in an ethical sense) and liable (in a legal sense) in the event of an error or accident becomes murky.
Evaluating a situation as to both ethical responsibility and legal liability
Because ethics is a philosophical discipline, a decision as to who is the ethical wrongdoer may not correlate with a decision as to who is the legally responsible party. Laws, which are codified and can change only through a deliberative (and often lengthy) process, tend to lag behind when it comes to adequately addressing novel situations created by rapidly advancing technologies.
In this column, I will explain the framework for evaluating a situation as to both ethical responsibility and legal liability, respectively. I will then apply that framework to two types of devices that blur the line between human action and machine action—self-driving vehicles and surgical robots.
I will consider two different situations that present questions of whose responsibility it is when something goes awry while these devices are operating, and I will discuss why the framework fails to produce acceptable outcomes. Finally, I will propose ways for the law to “keep up” with ethics, and argue for the importance of a flexible legal paradigm in this context.
Frameworks for evaluating ethical and legal responsibility
Let me start by defining the vocabulary I will be using in this discussion. I divide action into two levels: legal and ethical. In this framework, an ethical decision should always be legal, but an action that is legal may or may not be ethical. (I use “should” because this is a prescriptive model, but I acknowledge that there are situations where this ideal framework does not apply, and the law requires unethical action.)
The law provides only a baseline for acceptable behavior. Legislators and their laws and regulations—as well as other aspects of government—should be concerned with ensuring that people act in accordance with this minimum standard. The sources of ethical standards, on the other hand, can be the drives to maximize utility, contribute to social good, effect fairness or justice, and so on.
In the context of licensed professionals, ethical rules come from profession-specific rule-making bodies, such as a state bar, board, or licensing agency. These entities are uniquely able to consider the conventions of the profession, the limitations and capabilities of the individual professional, and other factors that are relevant to determining whether a particular action should be encouraged or discouraged.
The basic rules of personal injury law and how they are enforced
Put simply, the law of personal injury provides a remedy for a plaintiff who can prove that the defendant’s action or omission caused the plaintiff to suffer harm.
The law imposes on all individuals a duty of reasonable care to others, determined to be how a “reasonable person” in the same circumstances would act. If a person acts unreasonably, causing injury to another, then the law imposes liability on the unreasonable actor.
When the defendant is a licensed professional, the law will look to ethical rules as a guide. For example, in a legal malpractice lawsuit, a plaintiff may point to the attorney’s violation of the rules of professional ethics as evidence that her conduct fell below an expected standard of care.
Similarly, in a medical malpractice suit, a plaintiff would seek to establish that the physician’s actions were at odds with the standards accepted by the community of specialists in that discipline in order to prove that he breached his duty of care to the patient.
A second essential inquiry is one of causation. The law requires a showing of both “actual” and “proximate” causation before it imposes liability. To prove actual causation, a plaintiff in a lawsuit must show either that she would not have suffered injuries if it had not been for the defendant’s actions, or in some cases that the defendant’s actions were a “substantial factor” in bringing about the injury.
To prove proximate cause, also known as legal cause, a plaintiff must generally show that the defendant should have reasonably foreseen that his actions would cause the type of injury that the plaintiff suffered.
The basic rules of products liability and how they are enforced
Having thus established the framework for establishing legal liability in a personal injury case, I now turn to a second theory of liability: products liability.
The maker of a product owes a duty of care to the consumer, as well as to anyone who might foreseeably come into contact with the product. This duty of care encompasses the duty to design a safe product, to manufacture it free of defects, and to warrant that the product is fit for its ordinary purposes. If the manufacturer falls short on any of these duties, then it is legally liable for injuries that result from that failure.
If a consumer is using a product that malfunctions and injures a third party, the third party would likely sue the consumer as the operator of the malfunctioning device. In the interest of fairness and justice, the law permits the consumer to seek “contribution” from the manufacturer of the defective product (basically, it allows the consumer to transfer a portion of the damages he must pay based on the fault of the manufacturer), or indemnification—transfer of all the damages—if the consumer was entirely free of fault.
Autonomous vehicles: challenging the legal and ethical frameworks
As a result of lobbying by Google, Nevada and recently California have enacted legislation permitting self-driving cars to take to the roads. Both states require that a human be present in the car, sitting in the driver’s seat, and able to take control of the car at any time. Although there have reportedly been no accidents at all during the automatic operation of Google’s driverless cars, it would seem inevitable that an accident will someday occur as their use becomes more prevalent. Such an occurrence would present myriad issues for the existing framework of responsibility, as described above.
Consider, for instance, the following scenario: autonomous vehicle
Daniel is the backup driver in an autonomous vehicle. He sits behind the wheel, as required by law, and is attentive to his surroundings. It begins to rain lightly. The car advises that in inclement weather conditions, a driver must manually operate the vehicle. Because the rain is not heavy, Daniel believes it does not rise to the level of being “inclement weather,” so he allows the car to continue driving without assistance. The car suddenly makes a sharp turn and crashes into a tree, injuring Daniel.
The most salient question that arises out of this hypothetical scenario is “Who is at fault?” Should Daniel have taken the wheel when it began to rain, or was the car’s instruction too ambiguous to impose that responsibility on him? Daniel would likely sue the manufacturer of the vehicle under a theory of products liability. The manufacturer would argue that Daniel had a duty to operate the car manually when it began to rain. In this scenario, only Daniel himself was injured, but how would responsibility be distributed if a third party had been injured as a result?
There is no clearly applicable ethical standard to govern how Daniel should have acted, and the appropriate legal framework is even less clearly defined.
Google ostensibly intends for autonomous cars to help people who are unable to drive, thus presenting a challenge for the present requirement that a person be present and ready to drive the car manually at any given time. If a blind person is using an autonomous car and is faced with a situation requiring intervention, was that person ethically irresponsible for getting behind the wheel in the first place? Or was the manufacturer wrong to produce the vehicle with the intent of aiding drivers who require assistance?
As is illustrated by this and similar scenarios, the law of products liability and personal injury as currently formulated are not equipped to address these questions.
Never miss a press release, breaking news or special company announcement in the fast-paced domain of robotics.
Stay current with a premium membership to Robotics Business Review
Continue to browse more free previews