Empirical audit and review and an assessment of evidentiary value in research on the psychological consequences of scarcity
Edited by Elke U. Weber, Princeton University, Princeton, NJ, and approved October 5, 2021 (received for review February 21, 2021)
Letter
June 20, 2023
Letter
June 20, 2023
Abstract
Empirical audit and review is an approach to assessing the evidentiary value of a research area. It involves identifying a topic and selecting a cross-section of studies for replication. We apply the method to research on the psychological consequences of scarcity. Starting with the papers citing a seminal publication in the field, we conducted replications of 20 studies that evaluate the role of scarcity priming in pain sensitivity, resource allocation, materialism, and many other domains. There was considerable variability in the replicability, with some strong successes and other undeniable failures. Empirical audit and review does not attempt to assign an overall replication rate for a heterogeneous field, but rather facilitates researchers seeking to incorporate strength of evidence as they refine theories and plan new investigations in the research area. This method allows for an integration of qualitative and quantitative approaches to review and enables the growth of a cumulative science.
Sign up for PNAS alerts.
Get alerts for new articles, or get an alert when an article is cited.
Over the past decade, behavioral scientists have investigated how scarcity, or a lack of resources, changes how people think and behave, and how those changes can perpetuate poverty and other negative consequences such as reduced life expectancy (1), reduced cognitive ability (2), impulsiveness (3), and pain (4). Perhaps most strikingly, researchers have found that even otherwise prosperous people can fall into a scarcity mindset (5, 6). Because of its importance, researchers have worked tirelessly to understand scarcity’s origins and outcomes. What is the state of the evidence?
Assessing the Evidence of a Literature
During the same decade in which research has explored the psychology of scarcity, behavioral scientists have been revisiting commonplace scientific practices. Selective reporting of measures, treatments, and analyses, and the p-hacking they enable (7) has been identified as an existential threat to behavioral science (8, 9), a threat evident in large-scale replication efforts (10, 11). If researchers accept that not all published evidence is robust to replication, then they must wonder which evidence they should learn from.
Historically, there have been two broad categories of summary: the qualitative review and the quantitative meta-analysis. The former benefits from being broadly inclusive and flexibly articulated. It is a great tool for corralling the various questions researchers ask and the methods they use for finding answers. It is, however, unable to give a quantitative assessment of the underlying evidence. Meta-analysis seems promising in its attempts to use existing data to generate an overall estimate of effects. But meta-analysis risks reifying selectively reported findings (12), is vulnerable to biased selection (13), is indifferent to the quality of the original evidence (14), and answers a question that no one was asking, i.e., what the average effect of dissimilar studies is (15). These are both useful approaches, but they have limitations.
We develop and employ a third alternative: empirical review and audit. This approach defines the bounds of a topic area, followed by the identification and selection of studies that belong to that literature. Close replications of a broad cross-section of these studies allow for a systematic assessment of their robustness to replication.
This approach offers at least four notable advantages. First, the study identification process can limit bias through a clear inclusion rule and random selection within it or alternatively, carefully define the selection bias (e.g., by selecting the most-cited relevant paper from each year). Second, all methods and analyses can be preregistered, so the consequent empirical results will be free of selective reporting. Third, the reporting is not dependent on the results of the replications. We report all our results and consign none to the file drawer. This makes the results of our replications a better estimate of true effect sizes than any meta-analysis that is vulnerable to selective reporting and any publication that is usually dependent on the statistical significance of the results. Finally, the cumulative results offer a practical starting point for someone entering the discipline. The advantages of our approach allow researchers to get a clear sense of the quality of evidence for a specific topic area.
Method
We selected 20 studies for replication. We built a set of eligible papers and then drew from that set at random. The set included studies that 1) cited Shah et al.’s (2012) seminal paper on scarcity (5), 2) included scarcity as a factor in their design, and 3) could be replicated with an online sample. We did not decide on an operational definition of scarcity, but we accepted all measures and manipulations of scarcity that were proposed by the original authors. An initial set of 198 articles citing Shah et al. (2012) was evaluated by two reviewers from the research team, assigned at random (5). This set was narrowed to 32 articles that met all criteria, and at least one member of the research team believed contained studies which could be conducted online. This set was further narrowed to 23 that met all criteria and that both reviewers believed could be conducted online. If a paper contained multiple qualifying studies, we chose one at random. Although our selection rule might seem to allow a wide range of studies into the consideration set, we do not view this as a problem. Because the empirical audit and review is broad in its focus by definition, we opted not to substitute the authors’ definition of a concept with our own. Instead, an empirical audit and review can cover the entire scope of a research domain and does not rely on averaging or homogenizing the studies in a research area. As such, it can identify those strands of the topic that seem to have greater empirical merit, while allowing readers to consider and focus on the set of studies and characteristics that are relevant to their interests.
Each of the studies we selected used different manipulations and measures, but they had some commonalities. (The studies that are referenced directly in the text are cited in the main paper. The remaining studies are cited in SI Appendix as 1 through 13.) All included a conceptualization of scarcity as a factor in the study design, whether by asking people to consider a time they lacked resources, by restricting the supply of resources in the context of the experiment, or exposing people to images consistent with the absence of resources. The University of California Berkeley Institutional Review Board approved each study and participants provided their consent online at the start of each study.
Members of our research team had primary responsibility for replicating one study and secondary oversight for another study. We used original materials when possible and contacted the original authors if we encountered issues in recreating the study. We preregistered all methods, analyses, and sample sizes. To give us sufficient precision to comment on the statistical power of the original effects, our replications employed 2.5 times the sample size of the original paper (8). Because this approach would also allow us to detect smaller effects than in the original studies, it would have allowed us to detect significant effects even in the cases where the original findings were not significant.
Results
Fig. 1 shows our results. We report the results of 19 studies, due to discrepancies in the materials between the original and one of our replication studies (see SI Appendix for details). The Leftmost columns categorize commonalities among the studies. In the six studies featuring writing independent variables we reviewed the responses for nonsensical or careless responses and excluded them. Results including these responses are in SI Appendix. The Middle column shows that replication effect sizes were smaller than the original effect sizes for 80% of the studies, and directionally opposite for 30% of these 20 studies. Of the 18 studies that were significant in the original, four of our replication efforts yielded significant results. But significance is only one way to evaluate the results of a replication. The three Rightmost columns report estimates of the power in the original studies based on the replication effects. This analysis provides the upper bounds of the 95% CI for the estimated power of the original studies. Only 8 of the original studies included 33% power in these 95% CIs, indicating that most of the effects we attempted to replicate were too small to be detectably studied in the original investigations.
Fig. 1.
Empirical audit and review can look past aggregate conclusions to see either the specific cases that look empirically sound or categories of research manipulations or measures that seem stronger than others. While these studies are conceptually linked, there is no reason to assume that all hypotheses, designs, and tests are equally representative. Although most of our replications did not approach statistical significance, those that did (4, 16–18) shared some common features. Generally, they included scarcity manipulations related to economic scarcity and used dependent variables related to economic decisions. Three of the four successful replications involved priming participants with some type of financial constraint and then requiring them to engage in a financial or consumer decision task under these constraints. These commonalities suggest that a researcher who wants to build on scarcity priming literature should focus on how economic constraints affect financial decisions but shy away from attempting to build on the literature relating to other effects of scarcity.
General Discussion
Scarcity is a real and enduring societal problem, yet our results suggest that behavioral scientists have not fully identified the underlying psychology. Although this project has neither the goal nor the capacity to “accept the null” hypothesis for any of these tests, the replications of these studies indicate that within this set, scarcity primes have a minimal influence on cognitive ability, product attitudes, or well being. Nevertheless, feeling poor may influence financial and consumer decisions.
We used well-defined criteria for inclusion in our investigation. That is a strength in that it means there are conceptual links between the studies or methodological links in their execution. Nevertheless, it is also a weakness for further generalization. Perhaps studies conducted online are inherently smaller in effect size or more variable in replicability. Furthermore, some of the largest effects reported in the scarcity literature draw on considerably more-difficult-to-reach communities and pose a challenge to replication efforts (e.g., the field studies from ref. 2). Although many of our replications failed to find evidence for the psychological consequences of primed scarcity, real-life scarcity likely has many antecedents and consequences. Despite these limitations, our studies indicate that scarcity studies conducted online and using experimental manipulations do not reliably replicate.
Researchers strive to build a cumulative science. The idiosyncrasies of research questions, designs, and analyses can make that difficult, especially when publication bias and p-hacking distort findings. Qualitative reviews and meta-analyses offer imperfect solutions to the problem. The empirical audit and review corrals the disparate evidence and offers a tidier collection of findings. By identifying the strongest and weakest evidence, future researchers can build on the more solid parts of that foundation.
Data Availability
All original data, preregistration documents, and analysis code have been deposited in the Open Science Framework (https://osf.io/a2e96/) (19).
Acknowledgments
We thank the Dean’s office at the Haas School of Business for providing funding. NSF Graduate Research Fellowship Program Award DGE 1752814 supported S.A., M.E.E.-L., R.F., R.J., S.N.J., and J.M.O.
Supporting Information
Appendix 01 - Original (PDF)
- Download
- 430.19 KB
Appendix 01 - Updated (PDF)
- Download
- 459.80 KB
References
1
R. Chetty et al., The association between income and life expectancy in the United States, 2001-2014. JAMA 315, 1750–1766 (2016).
2
A. Mani, S. Mullainathan, E. Shafir, J. Zhao, Poverty impedes cognitive function. Science 341, 976–980 (2013).
3
D. Spears, Economic decision-making in poverty depletes behavioral control. B.E. J. Econ. Anal. Policy 11, 1–42 (2011).
4
E. Y. Chou, B. L. Parmar, A. D. Galinsky, Economic insecurity increases physical pain. Psychol. Sci. 27, 443–454 (2016).
5
A. K. Shah, S. Mullainathan, E. Shafir, Some consequences of having too little. Science 338, 682–685 (2012).
6
A. K. Shah, E. Shafir, S. Mullainathan, Scarcity frames value. Psychol. Sci. 26, 402–412 (2015).
7
J. P. Simmons, L. D. Nelson, U. Simonsohn, False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol. Sci. 22, 1359–1366 (2011).
8
U. Simonsohn, Small telescopes: Detectability and the evaluation of replication results. Psychol. Sci. 26, 559–569 (2015).
9
S. Vazire, Quality uncertainty erodes trust in science. Collabra Psychol. 3,. (2017).
10
Open Science Collaboration, Psychology. Estimating the reproducibility of psychological science. Science 349, aac4716 (2015).
11
C. F. Camerer et al., Evaluating the replicability of social science experiments in Nature and Science between 2010 and 2015. Nat. Hum. Behav. 2, 637–644 (2018).
12
I. Ahmed, A. J. Sutton, R. D. Riley, Assessment of publication bias, selection bias, and unavailable data in meta-analyses using individual participant data: A database survey. BMJ 344, d7762 (2012).
13
J. Vosgerau, U. Simonsohn, L. D. Nelson, J. P. Simmons, 99% impossible: A valid, or falsifiable, internal meta-analysis. J. Exp. Psychol. Gen. 148, 1628–1639 (2019).
14
J. P. Ioannidis, The mass production of redundant, misleading, and conflicted systematic reviews and meta‐analyses. Milbank Q. 94, 485–514 (2016).
15
L. Nelson, Data Colada, How Many Studies Have Not Been Run? Why We Still Think the Average Effect Does Not Exist, datacolada.org/70. Accessed 25 June 2021.
16
W. K. Bickel, A. G. Wilson, C. Chen, M. N. Koffarnus, C. T. Franck, Stuck in time: Negative income shock constricts the temporal window of valuation spanning the future and the past. PLoS One 11, e0163051 (2016).
17
L. A. Cook, R. Sadeghein, Effects of perceived scarcity on financial decision making. J. Public Policy Mark. 37, 68–87 (2018).
18
S. M. Tully, H. E. Hershfield, T. Meyvis, Seeking lasting enjoyment with limited money: Financial constraints increase preference for material goods over experiences. J. Consum. Res. 42, 59–75 (2015).
19
M. O’Donnell, et al., Empirical audit and review and an assessment of evidentiary value in research on the psychological consequences of scarcity. Open Science Framework (2022) https:/doi.org/10.17605/OSF.IO/A2E96.
Information & Authors
Information
Published in
Classifications
Copyright
Copyright © 2021 the Author(s). Published by PNAS. This open access article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND).
Data Availability
All original data, preregistration documents, and analysis code have been deposited in the Open Science Framework (https://osf.io/a2e96/) (19).
Submission history
Received: February 21, 2021
Accepted: September 22, 2021
Published online: October 28, 2021
Published in issue: November 2, 2021
Change history
November 15, 2022: Figure 1, the SI Appendix, and the main text of this article have been updated; please see accompanying Correction for details. Previous version (November 1, 2021)
Keywords
Acknowledgments
We thank the Dean’s office at the Haas School of Business for providing funding. NSF Graduate Research Fellowship Program Award DGE 1752814 supported S.A., M.E.E.-L., R.F., R.J., S.N.J., and J.M.O.
Authors
Competing Interests
The authors declare no competing interest.
Metrics & Citations
Metrics
Citation statements
Altmetrics
Citations
Cite this article
118 (44) e2103313118,
Export the article citation data by selecting a format from the list below and clicking Export.
Cited by
Loading...
View Options
View options
PDF format
Download this article as a PDF file
DOWNLOAD PDFLogin options
Check if you have access through your login credentials or your institution to get full access on this article.
Personal login Institutional LoginRecommend to a librarian
Recommend PNAS to a LibrarianPurchase options
Purchase this article to access the full text.