Are Messages From Robots Trustworthy?
From Nike and Google to Coca-Cola and McDonald’s, major brands are incorporating artificial intelligence (AI) into their advertising campaigns. But how do consumers feel about robots generating emotionally charged marketing content? That’s the question a New York Institute of Technology professor raises in a new Journal of Business Research study.
Whereas predictive AI allows marketers to forecast consumer behavior, generative AI enables them to produce novel content, including text, images, videos, or audio. For example, a recent AI-generated Toys“R”Us commercial featured a video of the company’s founder as a boy alongside its brand mascot Geoffrey the Giraffe. While many brands have trumpeted their AI-driven campaigns as a mark of innovation, others may fail to disclose AI use, leading to ethical concerns and calls for government regulation. However, even transparent brands receive backlash, as Google experienced when viewers were offended by its “Dear Sydney” ad, in which a father uses AI to help his daughter draft a fan letter to her favorite Olympic athlete.
“AI is a new territory for brand marketers, but what we do know is that consumers highly value authentic interactions with brands,” says the study’s lead author Colleen Kirk, D.P.S., professor of marketing and management. “Although more companies are now using AI-generated content to strengthen brand engagement and attachment, no study has explored how consumers view the authenticity of textual content that was created by a robot.”
Kirk and her study co-author, Julian Givi, Ph.D., a marketing researcher and faculty member at West Virginia University, completed various experiments to see how consumers react when emotional messages are written by AI. They hypothesized that consumers would view emotionally charged AI-generated content less favorably, impacting their perception of the brand and desire to interact with it.
In one scenario, participants imagined receiving a heartfelt message from a fitness salesperson who helped them buy a new set of weights. The message stated that he was inspired by the consumer’s purchase, with some participants believing it was AI-generated and others believing that the salesman drafted it himself (control group). While the members of the control group responded favorably, those in the AI group felt that the note violated their moral principles (moral disgust). As a result, this group was also unlikely to recommend the store to others and more likely to switch brands when making future purchases. Many even gave the store poor ratings on a simulated reviews site.
Other scenarios also revealed key findings in support of the researchers’ hypothesis:
- The negative effect diminishes when communications are factual, or AI is only used for editing.
- AI-generated messages in which the robot had self-autonomy (for example, an AI-generated memo signed by a chatbot) were viewed more favorably than AI-generated messages signed by a company representative.
- When participants believed that most emotional marketing communications were written by AI, they expressed disgust. The reverse was true when they believed most communications were written by a human. Therefore, brands may benefit from promoting the human origins of their products and communications.
- Human communicators (vs. AI) faced a greater “authenticity penalty” for copying emotional content.
In short, the findings suggest that companies must carefully consider whether and how to disclose AI-authored communications, always prioritizing authenticity in their interactions with consumers. As governments seek to increasingly regulate AI disclosure, making consumers more aware of how brands craft their messages, Kirk says marketers will want to pay close attention to the study’s findings.
“Consumers are becoming ever more skeptical of the human origin of marketing communications. Our research provides much-needed insight into how using AI to generate emotional content could negatively impact brands’ perceptions and, in turn, the consumer relationships that support their bottom lines. While AI tools offer marketers a new frontier, these professionals should bear in mind a time-tested principle: authenticity is always best,” she says.
More News
Teamwork Makes the Dream Work
M.B.A. students and roommates Krishna Jograna and Ritika Radadiya are inseparable and inspire each other to become better businesswomen.
Vancouver Campus Celebrates Its 24th Commencement Ceremony
On October 25, family, friends, faculty, and staff gathered at the Chan Centre for the Performing Arts at The University of British Columbia to celebrate New York Tech-Vancouver’s Class of 2024.
New Academic Year Brings New Faculty
New York Tech welcomes new and recently hired faculty to the university for the new academic year.
Intertwining Medicine and Management
Chiya Abramowitz (D.O./M.B.A. ’24) hopes to take his management and medical education to the next level as he pursues work in radiology and healthcare entrepreneurship.
School of Management Welcomes New Dean
On July 1, Jaishankar Ganesh, Ph.D., joined New York Tech as dean of the School of Management.
News Byte: Vancouver M.B.A. Students Win Prize in Global Competition
M.B.A. students Stella Luoou Li and Kumaran Ahilan received a second-place prize for their work creating solutions to corporate challenges in Southern Connecticut State University’s Global Business Students Competition.