国产视频

In Short

The Moral Gray Space of AI Decisions

cyber-2377718_1920.jpg

This is part of The Ethical Machine: Big ideas for designing fairer AI and algorithms, an on-going series about AI and ethics, curated by Dipayan Ghosh, a former Public Interest Technology fellow. You can see the full series on the .

PATRICK LIN
PROFESSOR OF PHILOSOPHY AND DIRECTOR, ETHICS + EMERGING SCIENCES GROUP, CALIFORNIA POLYTECHNIC STATE UNIVERSITY, SAN LUIS OBISPO

Artificial intelligence in diverse applications鈥攆rom sex bots to war machines鈥攊s giving rise to equally diverse concerns: algorithmic bias, transparency, accountability, privacy, psychological impact, trust, and beyond. Of course, all of these issues don鈥檛 necessarily arise in all forms of AI; for instance, few people, if any, care about privacy with military robots. But one root ethical issue that does apply to the entire technology category is the general ability to make decisions. This is the linchpin issue to be examined here.

While there鈥檚 no consensus around how to define 鈥渁rtificial intelligence鈥 or even 鈥渋ntelligence,鈥 here鈥檚 a working definition: AI is a computational system designed to automate decisions, with the appearance of intelligence. A robot, then, is embodied AI that translates those decisions into physical tasks. Given that a core function of artificial intelligence is to automate decisions, it鈥檚 fair to ask whether AI fulfills this function 鈥渃orrectly.鈥

The question is more difficult than it may appear once we recognize that there are three kinds of AI decisions. First, there are 鈥渞ight鈥 ones, that is, uncontroversial decisions that are made as intended or expected (putting aside the question of whether they鈥檙e objectively right; the point is that no one has a real problem with them). Second, there are wrong decisions, such as when an undesired outcome is (e.g., programming errors) [1], or by emergent behavior (e.g., 鈥渇lash crashes鈥 that from the sheer complexity of the system or interacting programs) [2], or by 鈥溾 an AI system with adversarial examples (i.e., inputs that intentionally cause it to make a mistake) [3]. But, third, there are also decisions that are neither obviously right nor wrong, as they fall within the shadowy but sizable space of judgment calls.

Those judgment calls, crystallized as code, are the ones that demand serious ethical consideration, especially since they may raise challenges to risk and liability.

Those judgment calls, crystallized as code, are the ones that demand serious ethical consideration, especially since they may raise challenges to risk and liability. If not considered with due care, this moral gray space could both derail technological progress and harm things that matter, such as inappropriately denying a job to someone who does not fit the profile of what a programmer or system thinks is a good applicant.

Case Study 1: Autonomous Cars

Let鈥檚 look at as a case study that hold lessons for other types of AI [4], especially since they鈥檙e poised to be the first robotics integrated into society at scale, setting the tone for other autonomous systems that follow. In an unexpected driving emergency, a bad decision by a human driver鈥攕ay, swerving into another car鈥攃ould be forgiven as a panicked reflex without forethought or malice. But the same action by a self-driving car, no matter how reasonable, would be scripted or otherwise predetermined, even if the action was learned (e.g., via neural networks). Either way, the AI decision is less like an innocent accident and more like premeditated injury; and both ethics and law treat these very differently, the second much more harshly.

Admittedly, crash dilemmas trigger an allergic reaction in some people who complain they鈥檙e too fake. This is perhaps best represented by the, a dilemma that imagines a no-win choice between switching the tracks of a speeding train which would then kill one person instead of the five others who are initially in its way, or doing nothing and letting the five people die [5]. Though that popular criticism really [6], we can also find automated decisions in everyday scenarios, if you want more realism.

Consider a where a self-driving car is driving in the middle lane of a highway [7]. If it鈥檚 flanked by a large truck on one side but a small car on the other side, where should the robot car be positioned within its own lane? There鈥檚 no obviously right answer here. It鈥檇 be reasonable to instruct the car to stay exactly in the middle of its lane, if there鈥檚 no actual emergency; or to give wider berth to the truck, even if it means nudging closer to a smaller car; or to even make the opposite decision and give more room to the smaller object, as the most vulnerable in the scenario. (This last decision can be made clearer by imagining the smaller object were a motorcyclist, bicyclist, or pedestrian.)

Judgments about risk are often unstated assumptions and not proactively defended.

All of the above answers are defensible, but those judgments about risk are often unstated assumptions and not proactively defended. If we can鈥檛 see how the moral math is being worked out, then it鈥檚 unclear that AI developers have done their due diligence and taken proper care in designing their products. Giving preferential treatment to one class of objects, such as giving wider berth to trucks, transfers some amount of risk to other road users or even awareness [8]. This and other design choices, as reasonable as they seem, may hold legal implications to industry鈥檚 surprise.

Though it might sound ordinary, lane positioning is actually a safety-critical decision that can share the same kind of trade-offs found in more dramatic crash scenarios. Any programming decision that involves a trade-off鈥攕uch as striking object x instead of y, or increasing distance away from x and toward y鈥攔equires a judgment about the wisdom of the trade-off, that is, about the relative weights of x and y. AI decisions are generally opaque to most of us already, but safety-critical AI decisions, including risk-benefit calculations, demand special attention and transparency.

Case Study 2: Traffic-Routing Apps

As an AI technology deployed widely today and again with relevance for other forms of AI, let鈥檚 look at the traffic-navigation app Waze to draw out the hidden ethics in more everyday scenarios. Often, there鈥檚 more than one reasonable way to get to a destination: one route could be the shortest distance but involve heavier traffic, or another may be lengthier but faster, or the longer route may be more scenic, and so on. These might not seem ethically problematic, until we realize that route selection involves risk. For instance, the fastest route may be more dangerous statistically if it includes more intersections, left turns, pedestrians, and other risk factors.

Apps such as Waze will generally default to the fastest route even if it鈥檚 [9]. This creates possible liability for making that dangerous choice without the user鈥檚 consent or knowledge, especially if the decision leads to an accident. But there鈥檚 also liability in ignoring risk data that鈥檚 readily available, such as insurance and government statistics on where the most accidents occur in a certain town. Waze is also giving rise to complaints about: groups of cars are sent by algorithms through quiet neighborhoods not designed for heavy traffic [10]. This could increase risk to children playing on these streets, lower property values because of the added road noise, and create other externalities or unintended harms.

But let鈥檚 suppose Waze wants to account for that risk data: Should it avoid poorer neighborhoods if there鈥檚 a statistical risk for increased crime or accidents? Any answer will be controversial. Even if crime data identifies these areas as clear risks, it could still be discriminatory to route traffic around them. For instance, related to structural racism, can generate more incident reports and data, and this makes their situation look worse than it really is [11]. Routing around those neighborhoods could also harm local merchants who鈥檇 then be less visible to potential customers, further aggravating the economic depression that already tends to exist in such areas.

On the other hand, if certain data about neighborhoods is excluded in decision-making for the sake of equality鈥攕uch as median income鈥攖hen liability is created if there were a correlation between risk and that data. Again, the base ethical dilemma here of making one choice to the detriment of something else resembles the difficult trade-off required in the trolley problem and other 鈥渇ake鈥 crash dilemmas. Indeed, given their large-scale effects, these everyday scenarios suggest that technology developers are actually unwittingly crafting public policy鈥攁 serious activity that clearly demands all of your wits.

Relevance to Other AI Systems

The case studies offered above are versatile; their general lessons can be applied to other AI areas. To start with, there鈥檚 a natural link here to, which also need to navigate through social spaces [12]. This means making decisions related to human interaction, some of them safety-critical in nature. Social robots, such as 鈥渃are bots,鈥 also may鈥攄ecisions that are neither obviously right nor wrong鈥攖hat weigh a patient鈥檚 autonomy against their well-being or doctor鈥檚 orders (e.g., if they refuse to take their medication) [13].

In AI decision-making across the spectrum, we need to be made aware of the assumptions, biases, and background considerations that invisibly power our technologies.

In AI decision-making across the spectrum, we need to be made aware of the assumptions, biases, and background considerations that invisibly power our technologies. Without that, we cannot hope to understand the risks and therefore cannot make informed decisions. At the same time, technology developers must tread carefully and transparently in this moral gray space: liability and trust implications can be contained for only so long under the cover of intellectual property and trade secrets before they blow up. As AI takes over more of our jobs, it also takes on new responsibilities and duties, perhaps more than technology developers appreciate today. Thinking openly about ethics now is crucial to their survival鈥攁s well as to ours.

References

  1. Patrick Lin, 鈥淗ere鈥檚 How Tesla Solves a Self-Driving Crash Dilemma,鈥 Forbes, April 5, 2017, http://www.forbes.com/sites/patricklin/2017/04/05/heres-how-tesla-solves-a-self-driving-crash-dilemma/.
  2. Drew Harwell, 鈥淎 Down Day on the Markets? Analysts Say Blame the Machines,鈥 The Washington Post, February 6, 2018,
  3. OpenAI, 鈥淎ttacking Machine Learning with Adversarial Examples,鈥 OpenAI blog, February 24, 2017,
  4. Patrick Lin, 鈥淭he Ethical Dilemma of Self-Driving Cars,鈥 TED-Ed, December 8, 2015,
  5. Lauren Davis, 鈥淲ould You Pull the Trolley Switch? Does It Matter?鈥 The Atlantic, October 9, 2015,
  6. Patrick Lin, 鈥淩obot Cars and Fake Ethical Dilemmas,鈥 Forbes, April 3, 2017,
  7. Noah Goodall, 鈥淎way from Trolley Problems and Toward Risk Management,鈥 Applied Artificial Intelligence 30, no. 8 (2016): 810鈥821,
  8. Patrick Lin, 鈥淭he Robot Car of Tomorrow Might Just Be Programmed to Hit You,鈥 Wired, May 6, 2014, http://www.wired.com/2014/05/the-robot-car-of-tomorrow-might-just-be-programmed-to-hit-you/.
  9. Linda Poon, 鈥淲aze Puts Safety Over Speed by Minimizing Left Turns in LA,鈥 CityLab, June 20, 2016,
  10. John Battelle, 鈥淭he Waze Effect: AI and the Public Commons,鈥 Newco Shift, February 3, 2016,
  11. David Kennedy, 鈥淏lack Communities: Overpoliced for Petty Crimes, Ignored for Major Ones,鈥 Los Angeles Times, April 10, 2015,
  12. Sibil Nicholson, 鈥淟atest Amazon Patent Includes Gesture-Recognizing Drones,鈥 Interesting Engineering, March 25, 2018,
  13. Michael Anderson and Susan Anderson, 鈥淢achine Ethics: Creating an Ethical Intelligent Agent,鈥 AI Magazine 1, no. 28 (2007): 15鈥26,

More 国产视频 the Authors

Patrick Lin
The Moral Gray Space of AI Decisions