Understanding algorithm aversion: When is advice from automation discounted?

Andrew Prahl*, Lyn Van Swol

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

211 Citations (Scopus)

Abstract

Forecasting advice from human advisors is often utilized more than advice from automation. There is little understanding of why “algorithm aversion” occurs, or specific conditions that may exaggerate it. This paper first reviews literature from two fields—interpersonal advice and human–automation trust—that can inform our understanding of the underlying causes of the phenomenon. Then, an experiment is conducted to search for these underlying causes. We do not replicate the finding that human advice is generally utilized more than automated advice. However, after receiving bad advice, utilization of automated advice decreased significantly more than advice from humans. We also find that decision makers describe themselves as having much more in common with human than automated advisors despite there being no interpersonal relationship in our study. Results are discussed in relation to other findings from the forecasting and human–automation trust fields and provide a new perspective on what causes and exaggerates algorithm aversion.

Original languageEnglish
Pages (from-to)691-702
Number of pages12
JournalJournal of Forecasting
Volume36
Issue number6
DOIs
Publication statusPublished - Sept 2017
Externally publishedYes

Bibliographical note

Publisher Copyright:
Copyright © 2017 John Wiley & Sons, Ltd.

ASJC Scopus Subject Areas

  • Modelling and Simulation
  • Computer Science Applications
  • Strategy and Management
  • Statistics, Probability and Uncertainty
  • Management Science and Operations Research

Keywords

  • advice
  • algorithm aversion
  • automation
  • computers
  • trust

Fingerprint

Dive into the research topics of 'Understanding algorithm aversion: When is advice from automation discounted?'. Together they form a unique fingerprint.

Cite this