Underspecified Human Decision Experiments Considered Harmful
Abstract: Decision-making with information displays is a key focus of research in areas like human-AI collaboration and data visualization. However, what constitutes a decision problem, and what is required for an experiment to conclude that decisions are flawed, remain imprecise. We present a widely applicable definition of a decision problem synthesized from statistical decision theory and information economics. We claim that to attribute loss in human performance to bias, an experiment must provide the information that a rational agent would need to identify the normative decision. We evaluate whether recent empirical research on AI-assisted decisions achieves this standard. We find that only 10 (26%) of 39 studies that claim to identify biased behavior presented participants with sufficient information to make this claim in at least one treatment condition. We motivate the value of studying well-defined decision problems by describing a characterization of performance losses they allow to be conceived.
- Applying quadratic scoring rule transparently in multiple choice settings: A note. Technical Report. Jena Economic Research Papers.
- Does the whole exceed its parts? the effect of ai explanations on complementary team performance. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–16.
- Dirk Bergemann and Stephen Morris. 2019. Information design: A unified perspective. Journal of Economic Literature 57, 1 (2019), 44–95.
- To trust or to think: cognitive forcing functions can reduce overreliance on AI in AI-assisted decision-making. Proceedings of the ACM on Human-Computer Interaction 5, CSCW1 (2021), 1–21.
- The role of explanations on trust and reliance in clinical decision support systems. In 2015 international conference on healthcare informatics. IEEE, 160–169.
- Colin Camerer. 1995. Individual decision making. Handbook of experimental economics (1995).
- Value-suppressing uncertainty palettes. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1–11.
- Algorithm aversion: people erroneously avoid algorithms after seeing them err. Journal of Experimental Psychology: General 144, 1 (2015), 114.
- Evanthia Dimara and John Stasko. 2021. A critical reflection on visualization research: Where do decision making tasks hide? IEEE Transactions on Visualization and Computer Graphics 28, 1 (2021), 1128–1138.
- Uncertainty displays using quantile dotplots or cdfs improve transit decision-making. In Proceedings of the 2018 CHI conference on human factors in computing systems. 1–12.
- The impact of algorithmic risk assessments on human predictions and its analysis via crowdsourcing studies. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–24.
- Tilmann Gneiting. 2011. Making and evaluating point forecasts. J. Amer. Statist. Assoc. 106, 494 (2011), 746–762.
- Daniel G Goldstein and David Rothschild. 2014. Lay understanding of probability distributions. Judgment and Decision making 9, 1 (2014), 1–14.
- Human evaluation of spoken vs. visual explanations for open-domain qa. arXiv preprint arXiv:2012.15075 (2020).
- Ben Green and Yiling Chen. 2019. Disparate interactions: An algorithm-in-the-loop analysis of fairness in risk assessments. In Proceedings of the conference on fairness, accountability, and transparency. 90–99.
- Visual reasoning strategies for effect size judgments and decisions. IEEE transactions on visualization and computer graphics 27, 2 (2020), 272–282.
- Bayesian-assisted inference from visualized data. IEEE Transactions on Visualization and Computer Graphics 27, 2 (2020), 989–999.
- A bayesian cognition approach to improve data visualization. In Proceedings of the 2019 chi conference on human factors in computing systems. 1–14.
- Tell me more? The effects of mental model soundness on personalizing an intelligent agent. In Proceedings of the sigchi conference on human factors in computing systems. 1–10.
- Towards a science of human-ai decision making: a survey of empirical studies. arXiv preprint arXiv:2112.11471 (2021).
- Towards a Science of Human-AI Decision Making: An Overview of Design Space in Empirical Human-Subject Studies. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. 1369–1385.
- Nicolas S Lambert. 2011. Elicitation and evaluation of statistical forecasts. Preprint (2011).
- Optimization of scoring rules. In Proceedings of the 23rd ACM Conference on Economics and Computation. 988–989.
- Zhiyuan “Jerry” Lin, Jongbin Jung, Sharad Goel, and Jennifer Skeem. 2020. The limits of human predictions of recidivism. Science advances 6, 7 (2020), eaaz0652.
- Understanding the effect of out-of-distribution examples and interactive explanations on human-ai decision making. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–45.
- Failures in contingent reasoning: The role of uncertainty. American Economic Review 109, 10 (2019), 3437–3474.
- C. Thi Nguyen. 2023. Value Collapse. Presented as the Annual Cardiff Lecture of the Royal Institute of Philosophy. https://royalinstitutephilosophy.org/event/value-collapse/
- The influence of different graphical displays on nonexpert decision making under uncertainty. Journal of Experimental Psychology: Applied 21, 1 (2015), 37.
- Understanding Uncertainty: How Lay Decision-makers Perceive and Interpret Uncertainty in Human-AI Decision Making. In Proceedings of the 28th International Conference on Intelligent User Interfaces. 379–396.
- Daniel Read. 2005. Monetary incentives, what are they good for? Journal of Economic Methodology 12, 2 (2005), 265–276.
- ” Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. 1135–1144.
- Leonard J Savage. 1972. The foundations of statistics. Courier Corporation.
- Katie Steele and H Orri Stefánsson. 2015. Decision theory. (2015).
- John Von Neumann and Oskar Morgenstern. 2007. Theory of games and economic behavior (60th Anniversary Commemorative Edition). Princeton university press.
- Xinru Wang and Ming Yin. 2021. Are explanations helpful? a comparative study of the effects of explanations in ai-assisted decision-making. In 26th international conference on intelligent user interfaces. 318–328.
- The Rational Agent Benchmark for Data Visualization. IEEE Transactions on Visualization and Computer Graphics (forthcoming) (2023).
- How do visual explanations foster end users’ appropriate trust in machine learning?. In Proceedings of the 25th international conference on intelligent user interfaces. 189–201.
- Designing Shared Information Displays for Agents of Varying Strategic Sophistication. arXiv preprint arXiv:2310.10858 (2023).
- Sibyl: Understanding and addressing the usability challenges of machine learning in high-stakes decision making. IEEE Transactions on Visualization and Computer Graphics 28, 1 (2021), 1161–1171.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.