Humans less likely to return to an automated advisor once given bad advice

The ubiquitous Chat Bot popping up on websites asking if you need help has become standard on many sites. We dismiss, we engage, but do we trust the algorithm that is aiding our experience? Giving us answers and advice? A recent study by researchers at the University of Wisconsin found that participants were less likely to return to an automated advisor if given bad advice over a human advisor under the same circumstances.

Andrew Prahl and Lyn M. Van Swol (University of Wisconsin) will present their findings in June at the 66th Annual Conference of the International Communication Association in Fukuoka, Japan. The researchers' experiment asked to forecast scheduling for hospital operating rooms, a task they were unfamiliar with. In order to complete this task, participants were given help from either an "advanced computer system" or "a person experienced in operating room management." During the 7th trial of the 14 scheduling trials, participants were given bad advice from either the computer advisor or human advisor.

The researchers found that after participants received bad advice, they rapidly abandoned the computer advisor and did not use the advice on subsequent trials. This "punishment" for giving bad advice was not nearly as strong for the human advisor. It is almost as if people "forgave" the human advisor for making a mistake but did not extend the same feelings of forgiveness to the computer.

Past research has looked at trust of through simple tasks like warning systems. Prahl and Van Swol wanted to look at more sophisticated automation that actually makes predictions about future outcomes and compare that to advice. Unlike previous research, the researchers did not find that humans or algorithmic advisors were generally trusted more, the only significant differences arose after the advisor provided bad advice.

"This has very important implications because time and time again we are seeing humans being replaced by computers in the workplace," said Prahl. "This research suggests that any potential efficiency gains by moving towards automation might be offset because all the automation has to do is err once, and people will rapidly lose trust and stop using it - this is one of the few studies out there that really show the potential downsides of automation in the workplace."

More information: "The Computer Said I Should: How Does Receiving Advice From a Computer Differ From Receiving Advice From a Human," by Andrew Prahl and Lyn M. Van Swol; to be presented at the 66th Annual International Communication Association Conference, Fukuoka, Japan, 9-13 June 2016.

Citation: Humans less likely to return to an automated advisor once given bad advice (2016, May 25) retrieved 20 February 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Young people overly optimistic about finances


Feedback to editors