Moralities of Intelligent Machines is a project that investigates people’s attitudes towards moral choices made by artificial intelligence.
In the latest study completed under the project, study participants read short narratives where either a robot, a somewhat humanoid robot known as iRobot, a robot with a strong humanoid appearance called iClooney or a human being encounters a moral problem along the lines of the trolley dilemma, making a specific decision.
The participants were also shown images of these agents, after which they assessed the morality of their decisions. The study was funded by the Jane and Aatos Erkko Foundation and the Academy of Finland.
The trolley dilemma is a problem where a person sees a trolley careening on the tracks, without anyone in control, towards five people. The person can either do nothing or turn the trolley onto another track, saving the five people but killing another individual on the other track.
Attitudes more negative towards humanoid robots According to the study, people consider the choice made by the humanoid iRobot and iClooney less ethically sound than the same decision made by a human and a robot with a traditional robot-like appearance.
Michael Laakasuo, senior researcher in University of Helsinki, project lead and the principal investigator of the study, links the findings to the uncanny valley effect, which has been identified in prior research.
“Humanness in artificial intelligence is perceived as eerie or creepy, and attitudes towards such robots are more negative than towards more machine-like robots. This may be due to, for example, the difficulty of reacting to a humanoid being: is it an animal, a human or a tool?”
According to Laakasuo, the findings indicate that humans do not find robots making moral decisions a strange idea, since the decisions made by a human and a traditional robot were seen as equally acceptable. Instead, the appearance of the robot makes a difference to evaluating their morality.
Discussion guides the regulation of AI Laakasuo says that the number of intelligent machines making moral choices is growing in our society, with self-driving cars as an example.
“It’s important to know how people view intelligent machines and what kinds of factors affect related moral assessment. For instance, are traffic violations perpetrated by a stylish self-driving car perceived differently from those of a less classy model?”
This knowledge can influence the direction of AI and robotics development, as well as, among other things, product branding. Knowledge can also shape the political discussion relating to the regulation of artificial intelligence.
For example, self-driving cars can become test laboratories of sorts for private companies: in the case of accidents, the consequences can be dealt with using money, risking human health in the name of technological advancement with appeals to consequentialist morals.
“What kind of robots do we want to have among us: robots who save five people from being run over by a trolley, sacrificing one person, or robots who refuse to sacrifice anyone even if it would mean saving several lives? Should robots be designed to look like humans or not if their appearance affects the perceived morality of their actions?”
- 30Psychologists have long found that people behave differently than when they learn of peers' actions. A new study by computer scientists found that when individuals in an experiment about autonomous vehicles were informed that their peers were more likely to sacrifice their own safety to program their vehicle to hit a wall…