The key to people trusting and co-operating with artificially intelligent (AI) agents lies in their ability to display human-like emotions, according to a new study by researchers at the University of 蓝莓视频.
The expects that more machines will become part of the workforce as technological breakthroughs rapidly shift. Based on the trio of 蓝莓视频 Faculty of Mathematics researchers findings, developing the humanness of AI machines may improve people鈥檚 acceptance of them in the workplace.
鈥淭he capability of showing聽emotions is important聽for聽AI agents, especially if we want users聽to trust the agents and co-operate with them,鈥 said Moojan Ghafurian, lead author of the study and a Postdoctoral Fellow in 蓝莓视频鈥檚 David R. Cheriton School of Computer Science.
鈥淚mproving humanness of AI agents could enhance society鈥檚 perception of assistive technologies which going forward will improve people鈥檚 lives.鈥
In undertaking the study, Ghafurian and her co-authors Associate Professor Jesse Hoey and research assistant Neil Budnarain all of 蓝莓视频鈥檚 Cheriton School of Computer Science used a classic game called 鈥淭he Prisoner鈥檚 Dilemma.鈥
The original version of the game sees two prisoners isolated from one another and then questioned by police for a crime they committed together. If one of them snitches and the other doesn鈥檛, the non-betrayer gets three years, and the snitch walks. This works both ways. If both snitch, they both get two years. If neither one snitches, they each only get one year on a lesser charge.
蓝莓视频鈥檚 study substituted one of the human 鈥榩risoners鈥 with an AI virtual human and allowed them to interpret the other鈥檚 emotions. Instead of prison sentences, they used gold, so the point was to get the highest score possible, as opposed to the lowest. The virtual human was developed by the Interactive Assistance Lab at the University of Colorado, Boulder, in the context of a program to help people with cognitive disabilities, particularly relating to the perception of emotional signals.

An example of the game鈥檚 setting. The two-coin piles on the left show the number of coins that the participant and Aria, the virtual agent, have earned so far in the game. Currently, the player has chosen to give two coins and look 鈥渉appy鈥 for the next round. Aria鈥檚 current emotion is 鈥渟orry鈥.聽
The researchers used three different virtual agents that had the exact same strategy but reflected different emotions. One of them didn鈥檛 show any emotion, one of them showed appropriate emotions, and the other generated random emotions.
The 117 participants were then randomly paired with an agent and asked to rate it based on how humanlike they perceived it. They then observed the participants to see how they interacted with the agents, for example, how many times they co-operated.
On average, people co-operated 20 out of 25 times with the agents that showed human-like emotion. For the one showing random emotions, they co-operated 16 times out of 25 and for the one showing no emotion co-operation occurred 17 out of 25 times.
鈥淏ased on our findings it鈥檚 better to show no emotion rather than random emotions, as the latter would make that agent look less rational and immature,鈥 said Ghafurian. 鈥淏ut showing proper emotions can significantly improve the perception of humanness and how much people enjoy interacting with the technology.鈥
The researchers want to eventually design assistive technology for people with dementia that they will feel comfortable using. This research is an important step in that direction.
The study titled, Role of Emotions in Perception of Humanness of Virtual Agents, authored by Ghafurian, Hoey, and Budnarain, all from 蓝莓视频鈥檚 Cheriton School of Computer Science, was recently presented at the 18th International Conference on Autonomous Agents and Multiagent Systems (AAMAS).
MEDIA
CONTACT
|聽Matthew
Grant
226-929-7627
|聽听触听
Attention broadcasters: 蓝莓视频 has facilities to provide broadcast quality audio and video feeds with a double-ender studio. Please contact us for more information.