24/7 Space News
ROBO SPACE
Is AI in the eye of the beholder?
whatever...
ADVERTISEMENT
The 2024 Humans To Mars Summit - May 07-08, 2024 - Washington D.C.
Is AI in the eye of the beholder?
by Adam Zewe for MIT News
Boston MA (SPX) Oct 04, 2023

Someone's prior beliefs about an artificial intelligence agent, like a chatbot, have a significant effect on their interactions with that agent and their perception of its trustworthiness, empathy, and effectiveness, according to a new study.

Researchers from MIT and Arizona State University found that priming users - by telling them that a conversational AI agent for mental health support was either empathetic, neutral, or manipulative - influenced their perception of the chatbot and shaped how they communicated with it, even though they were speaking to the exact same chatbot.

Most users who were told the AI agent was caring believed that it was, and they also gave it higher performance ratings than those who believed it was manipulative. At the same time, less than half of the users who were told the agent had manipulative motives thought the chatbot was actually malicious, indicating that people may try to "see the good" in AI the same way they do in their fellow humans.

The study revealed a feedback loop between users' mental models, or their perception of an AI agent, and that agent's responses. The sentiment of user-AI conversations became more positive over time if the user believed the AI was empathetic, while the opposite was true for users who thought it was nefarious.

"From this study, we see that to some extent, the AI is the AI of the beholder," says Pat Pataranutaporn, a graduate student in the Fluid Interfaces group of the MIT Media Lab and co-lead author of a paper describing this study. "When we describe to users what an AI agent is, it does not just change their mental model, it also changes their behavior. And since the AI responds to the user, when the person changes their behavior, that changes the AI, as well."

Pataranutaporn is joined by co-lead author and fellow MIT graduate student Ruby Liu; Ed Finn, associate professor in the Center for Science and Imagination at Arizona State University; and senior author Pattie Maes, professor of media technology and head of the Fluid Interfaces group at MIT.

The study, published in Nature Machine Intelligence, highlights the importance of studying how AI is presented to society, since the media and popular culture strongly influence our mental models. The authors also raise a cautionary flag, since the same types of priming statements in this study could be used to deceive people about an AI's motives or capabilities.

"A lot of people think of AI as only an engineering problem, but the success of AI is also a human factors problem. The way we talk about AI, even the name that we give it in the first place, can have an enormous impact on the effectiveness of these systems when you put them in front of people. We have to think more about these issues," Maes says.

AI friend or foe?
In this study, the researchers sought to determine how much of the empathy and effectiveness people see in AI is based on their subjective perception and how much is based on the technology itself. They also wanted to explore whether one could manipulate someone's subjective perception with priming.

"The AI is a black box, so we tend to associate it with something else that we can understand. We make analogies and metaphors. But what is the right metaphor we can use to think about AI? The answer is not straightforward," Pataranutaporn says.

They designed a study in which humans interacted with a conversational AI mental health companion for about 30 minutes to determine whether they would recommend it to a friend, and then rated the agent and their experiences. The researchers recruited 310 participants and randomly split them into three groups, which were each given a priming statement about the AI.

One group was told the agent had no motives, the second group was told the AI had benevolent intentions and cared about the user's well-being, and the third group was told the agent had malicious intentions and would try to deceive users. While it was challenging to settle on only three primers, the researchers chose statements they thought fit the most common perceptions about AI, Liu says.

Half the participants in each group interacted with an AI agent based on the generative language model GPT-3, a powerful deep-learning model that can generate human-like text. The other half interacted with an implementation of the chatbot ELIZA, a less sophisticated rule-based natural language processing program developed at MIT in the 1960s.

Molding mental models
Post-survey results revealed that simple priming statements can strongly influence a user's mental model of an AI agent, and that the positive primers had a greater effect. Only 44 percent of those given negative primers believed them, while 88 percent of those in the positive group and 79 percent of those in the neutral group believed the AI was empathetic or neutral, respectively.

"With the negative priming statements, rather than priming them to believe something, we were priming them to form their own opinion. If you tell someone to be suspicious of something, then they might just be more suspicious in general," Liu says.

But the capabilities of the technology do play a role, since the effects were more significant for the more sophisticated GPT-3 based conversational chatbot.

The researchers were surprised to see that users rated the effectiveness of the chatbots differently based on the priming statements. Users in the positive group awarded their chatbots higher marks for giving mental health advice, despite the fact that all agents were identical.

Interestingly, they also saw that the sentiment of conversations changed based on how users were primed. People who believed the AI was caring tended to interact with it in a more positive way, making the agent's responses more positive. The negative priming statements had the opposite effect. This impact on sentiment was amplified as the conversation progressed, Maes adds.

The results of the study suggest that because priming statements can have such a strong impact on a user's mental model, one could use them to make an AI agent seem more capable than it is - which might lead users to place too much trust in an agent and follow incorrect advice.

"Maybe we should prime people more to be careful and to understand that AI agents can hallucinate and are biased. How we talk about AI systems will ultimately have a big effect on how people respond to them," Maes says.

In the future, the researchers want to see how AI-user interactions would be affected if the agents were designed to counteract some user bias. For instance, perhaps someone with a highly positive perception of AI is given a chatbot that responds in a neutral or even a slightly negative way so the conversation stays more balanced.

They also want to use what they've learned to enhance certain AI applications, like mental health treatments, where it could be beneficial for the user to believe an AI is empathetic. In addition, they want to conduct a longer-term study to see how a user's mental model of an AI agent changes over time.

This research was funded, in part, by the Media Lab, the Harvard-MIT Program in Health Sciences and Technology, Accenture, and KBTG.

Research Report:"Influencing human-AI interaction by priming beliefs about AI can increase perceived trustworthiness, empathy and effectiveness"

Related Links
Fluid Interfaces
All about the robots on Earth and beyond!

Subscribe Free To Our Daily Newsletters

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
ROBO SPACE
Meta technology chief defends tech titan's AI strategy
Menlo Park, United States (AFP) Oct 1, 2023
Meta's chief technology officer is quick to push back on assertions that the company has fallen behind rivals like ChatGPT in the explosive surge across the tech industry in generative AI. "The majority of the world's population will have their first experience of generative artificial intelligence with us," Andrew "Boz" Bosworth told AFP at the company's recent Connect conference for developers. Meta unveiled AI-infused chatbots with personalities at the gathering, along with tools for creating ... read more

ADVERTISEMENT
ADVERTISEMENT
ROBO SPACE
Chinese universities climb up leading global ranking

NASA astronaut Frank Rubio returning to Earth after record 371 days in space

Kayhan Space Raises $7 million, Unveils First-Ever Autonomous Space Traffic Coordination Service

Two Russians, American reach space station

ROBO SPACE
All engines added to NASA's Artemis II core stage

Historic NASA wind tunnel testing Mars Ascent Vehicle

Third Subscale Booster for future Artemis missions fires up at Marshall

'Anomaly' ends Rocket Lab launch mid-flight

ROBO SPACE
Curiosity Needs an Altitude Adjustment: Sols 3955-3956

"Sombrero Rock": A Case of Case-Hardening?

Did life exist on Mars? Other planets? With AI's help, we may know soon

Big Fan of Rock Bands: Sols 3960-3961

ROBO SPACE
Astronauts honored for contributions to China's space program

China capable of protecting astronauts from effects of space weightlessness

Tianzhou 5 spacecraft burns up on Earth reentry

Crew of Shenzhou XV mission honored for six-month space odyssey

ROBO SPACE
Terran Orbital Announces Closing of $32.5 Million Public Offering

Iridium and McQ develop remote monitoring solution for Canadian Armed Forces in the Arctic

Terran Orbital announces pricing of Public Offering

Intelsat Inflight Connectivity expanded to all Airbus aircraft

ROBO SPACE
Metal-loving microbes could replace chemical processing of rare earths

Material matters

Mineral-hungry clean tech sees countries seeking to escape China's shadow

Green issues dominate Paris fashion as green tech marketplace debuts

ROBO SPACE
Study sheds new light on strange lava worlds

JWST's first spectrum of a TRAPPIST-1 planet

Alien Machines in the Solar System: The Possibilities and Potential Origins

Possible hints of life found on distant planet - how excited should we be?

ROBO SPACE
Webb finds carbon source on surface of Jupiter's moon Europa

Hidden ocean the source of CO2 on Jupiter moon

Juice: why's it taking sooo long

Possible existence of Earth-like planet predicted in Outskirts of Solar System

Subscribe Free To Our Daily Newsletters


ADVERTISEMENT



The content herein, unless otherwise known to be public domain, are Copyright 1995-2023 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.