ITHACA, N.Y. - Organizations using AI to monitor employees’ behavior and productivity can expect them to complain more, be less productive and want to quit more – unless the technology can be framed as supporting their development, Cornell University research finds.
Surveillance tools cause people to feel a greater loss of autonomy than oversight by humans, according to the research. Businesses and other organizations using the fast-changing technologies to evaluate employee behaviors should consider their unintended consequences, which may prompt resistance and hurt performance, the researchers say. They also suggest an opportunity to win buy-in, if the subjects of surveillance feel the tools are there to assist rather than judge their performance – assessments they fear will lack context and accuracy.
“When artificial intelligence and other advanced technologies are implemented for developmental purposes, people like that they can learn from it and improve their performance,” said Emily Zitek, associate professor of organizational behavior. “The problem occurs when they feel like an evaluation is happening automatically, straight from the data, and they’re not able to contextualize it in any way.”
Zitek is the co-author of “Algorithmic Versus Human Surveillance Leads to Lower Perceptions of Autonomy and Increased Resistance.” Rachel Schlund is first author.
The researchers conducted four experiments involving nearly 1,200 total participants. In a first study, when asked to recall and write about times when they were monitored and evaluated by either surveillance type, participants reported feeling less autonomy under AI and were more likely to engage in “resistance behaviors.”
A pair of studies asked participants to work as a group to brainstorm ideas for a theme park, then to individually generate ideas about one segment of the park. They were told their work would be monitored by a research assistant or AI—the latter represented in Zoom videoconferences as “AI Technology Feed.” After several minutes, either the human assistant or “AI” relayed messages that the participants weren’t coming up with enough ideas and should try harder. In surveys following one study, more than 30% of participants criticized the AI surveillance compared to about 7% who were critical of the human monitoring.
Beyond complaints and criticism, the researchers found that those who thought they were being monitored by AI generated fewer ideas – indicating worse performance.
“Even though the participants got the same message in both cases that they needed to generate more ideas, they perceived it differently when it came from AI rather than the research assistant,” Zitek said. “The AI surveillance caused them to perform worse in multiple studies.”
In a fourth study, participants imagining they worked in a call center were told that humans or AI would analyze a sample of their calls. For some, the analysis would be used to evaluate their performance; for others, to provide developmental feedback. In the developmental scenario, participants no longer perceived algorithmic surveillance as infringing more on their autonomy and did not report a greater intention to quit.
“Organizations trying to implement this kind of surveillance need to recognize the pros and cons,” Zitek said. “They should do what they can to make it either more developmental or ensure that people can add contextualization. If people feel like they don’t have autonomy, they’re not going to be happy.”
For additional information, read this Cornell Chronicle story.
Cornell University has dedicated television and audio studios available for media interviews.
-30-
Journal
Communications Psychology
Article Title
Algorithmic Versus Human Surveillance Leads to Lower Perceptions of Autonomy and Increased Resistance
Article Publication Date
6-Jun-2024