No Evidence that Experiencing Physical Warmth Promotes Interpersonal Warmth: Two Failures to Replicate Williams and Bargh (2008)

Author(s):  
Christopher Chabris ◽  
Patrick Ryan Heck ◽  
Jaclyn Mandart ◽  
Daniel Jacob Benjamin ◽  
Daniel J. Simons

Williams and Bargh (2008) reported that holding a hot cup of coffee caused participants to judge a person’s personality as warmer, and that holding a therapeutic heat pad caused participants to choose rewards for other people rather than for themselves. These experiments featured large effects (r = .28 and .31), small sample sizes (41 and 53 participants), and barely statistically significant results. We attempted to replicate both experiments in field settings with more than triple the sample sizes (128 and 177) and double-blind procedures, but found near-zero effects (r = –.03 and .02). In both cases, Bayesian analyses suggest there is substantially more evidence for the null hypothesis of no effect than for the original physical warmth priming hypothesis.

2019 ◽  
Vol 50 (2) ◽  
pp. 127-132 ◽  
Author(s):  
Christopher F. Chabris ◽  
Patrick R. Heck ◽  
Jaclyn Mandart ◽  
Daniel J. Benjamin ◽  
Daniel J. Simons

Abstract. Williams and Bargh (2008) reported that holding a hot cup of coffee caused participants to judge a person’s personality as warmer and that holding a therapeutic heat pad caused participants to choose rewards for other people rather than for themselves. These experiments featured large effects ( r = .28 and .31), small sample sizes (41 and 53 participants), and barely statistically significant results. We attempted to replicate both experiments in field settings with more than triple the sample sizes (128 and 177) and double-blind procedures, but found near-zero effects ( r = −.03 and .02). In both cases, Bayesian analyses suggest there is substantially more evidence for the null hypothesis of no effect than for the original physical warmth priming hypothesis.


2020 ◽  
Author(s):  
Taylor Winter ◽  
Benjamin Riordan ◽  
Anthony Surace ◽  
Damian Scarf ◽  
Paul Jose

Aims. Quantifying differences between minority and majority groups, such as sexual minorities (SM) and heterosexuals, is difficult due to small sample sizes. Bayesian analyses is one solution to addressing small sample sizes in minority group research, whereby previous research can be used to inform our models. In the present tutorial, we offered an overview of Bayesian statistics and described an approach to constructing informed priors using a large survey when estimating values in a smaller survey. In an applied example, we determined whether SMs in New Zealand reported more stress relative to heterosexuals and whether stress mediates the link between SM status and alcohol use.Design. Two cross-sectional, stratified, and nationally representative health surveys from the US (National Survey of Drug Use and Health (NSDUH)) and New Zealand (New Zealand Health Survey (NZHS)).Settings. United States, New ZealandParticipants. We used data from 83,661 (SMs = 5593) survey respondents in the US and 24,098 respondents in NZ (SMs = 619).Measurements. Demographic items (sex, age, ethnicity, sexual identity), the Kessler psychological distress scale, and the Alcohol Use Disorder Identification Test (AUDIT).Findings. Using a larger survey to inform priors reduced the uncertainty of estimates derived from small subgroups in a smaller survey relative to uninformed priors.Conclusion. Informed Bayesian analyses are an important tool for researchers studying minority groups and the application of informative priors allows for more reliable estimates of health disparities.


2019 ◽  
Vol 147 (2) ◽  
pp. 763-769 ◽  
Author(s):  
D. S. Wilks

Abstract Quantitative evaluation of the flatness of the verification rank histogram can be approached through formal hypothesis testing. Traditionally, the familiar χ2 test has been used for this purpose. Recently, two alternatives—the reliability index (RI) and an entropy statistic (Ω)—have been suggested in the literature. This paper presents approximations to the sampling distributions of these latter two rank histogram flatness metrics, and compares the statistical power of tests based on the three statistics, in a controlled setting. The χ2 test is generally most powerful (i.e., most sensitive to violations of the null hypothesis of rank uniformity), although for overdispersed ensembles and small sample sizes, the test based on the entropy statistic Ω is more powerful. The RI-based test is preferred only for unbiased forecasts with small ensembles and very small sample sizes.


2018 ◽  
Vol 13 (4) ◽  
pp. 403-408 ◽  
Author(s):  
Jeff Bodington ◽  
Manuel Malfeito-Ferreira

AbstractMuch research shows that women and men have different taste acuities and preferences. If female and male judges tend to assign different ratings to the same wines, then the gender balances of the judge panels will bias awards. Existing research supports the null hypothesis, however, that finding is based on small sample sizes. This article presents the results for a large sample; 260 wines and 1,736 wine-score observations. Subject to the strong qualification that non-gender-related variation is material, the results affirm that female and male judges do assign about the same ratings to the same wines. The expected value of the difference in their mean ratings is zero. (JEL Classifications: A10, C00, C10, C12, D12)


2021 ◽  
Vol 12 ◽  
Author(s):  
James W. B. Elsey ◽  
Merel Kindt

The idea that maladaptive memories may be rendered susceptible to interference after reactivation raises the possibility of reactivating and neutralizing clinically-relevant emotional memories. In this study, we sought to investigate the feasibility of such a “reconsolidation-based” intervention for arachnophobia, drawing upon previous research that successfully reduced fear of spiders in a subclinical sample. In Experiment 1, we piloted several reactivation procedures for conducting a reconsolidation-based treatment for arachnophobic individuals. All procedures involved some form of brief exposure to a fear-provoking spider, followed by the administration of 40 mg propranolol. In Experiment 2, we conducted a double-blind, placebo-controlled assessment of one procedure tested in Experiment 1. In Experiment 1, we found that most reactivation procedures produced drops in self-reported fear of spiders from pre- to post-treatment, including fear declines that were apparent up to 6- and even 14-months later. However, in Experiment 2, we found no evidence that the participants receiving propranolol were better off than those who received placebo. While our findings are limited by the small sample sizes used, they nevertheless show a different pattern of responses than was observed in a previous reconsolidation-based intervention for subclinical spider fearful participants. Alterations to the protocol made to accommodate the clinical participants may have led to greater opportunities for non-specific effects (e.g., exposure, placebo effects) to drive change in the participants. Our findings highlight both the challenges of translating reconsolidation-based procedures into clinical interventions, as well as the importance of controls for non-specific effects in reconsolidation-based research.


2020 ◽  
Vol 117 (32) ◽  
pp. 19151-19158 ◽  
Author(s):  
M.-A. C. Bind ◽  
D. B. Rubin

In randomized experiments, Fisher-exactPvalues are available and should be used to help evaluate results rather than the more commonly reported asymptoticPvalues. One reason is that using the latter can effectively alter the question being addressed by including irrelevant distributional assumptions. The Fisherian statistical framework, proposed in 1925, calculates aPvalue in a randomized experiment by using the actual randomization procedure that led to the observed data. Here, we illustrate this Fisherian framework in a crossover randomized experiment. First, we consider the first period of the experiment and analyze its data as a completely randomized experiment, ignoring the second period; then, we consider both periods. For each analysis, we focus on 10 outcomes that illustrate important differences between the asymptotic and Fisher tests for the null hypothesis of no ozone effect. For some outcomes, the traditionalPvalue based on the approximating asymptotic Student’stdistribution substantially subceeded the minimum attainable Fisher-exactPvalue. For the other outcomes, the Fisher-exact null randomization distribution substantially differed from the bell-shaped one assumed by the asymptoticttest. Our conclusions: When researchers choose to reportPvalues in randomized experiments, 1) Fisher-exactPvalues should be used, especially in studies with small sample sizes, and 2) the shape of the actual null randomization distribution should be examined for the recondite scientific insights it may reveal.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Florent Le Borgne ◽  
Arthur Chatton ◽  
Maxime Léger ◽  
Rémi Lenain ◽  
Yohann Foucher

AbstractIn clinical research, there is a growing interest in the use of propensity score-based methods to estimate causal effects. G-computation is an alternative because of its high statistical power. Machine learning is also increasingly used because of its possible robustness to model misspecification. In this paper, we aimed to propose an approach that combines machine learning and G-computation when both the outcome and the exposure status are binary and is able to deal with small samples. We evaluated the performances of several methods, including penalized logistic regressions, a neural network, a support vector machine, boosted classification and regression trees, and a super learner through simulations. We proposed six different scenarios characterised by various sample sizes, numbers of covariates and relationships between covariates, exposure statuses, and outcomes. We have also illustrated the application of these methods, in which they were used to estimate the efficacy of barbiturates prescribed during the first 24 h of an episode of intracranial hypertension. In the context of GC, for estimating the individual outcome probabilities in two counterfactual worlds, we reported that the super learner tended to outperform the other approaches in terms of both bias and variance, especially for small sample sizes. The support vector machine performed well, but its mean bias was slightly higher than that of the super learner. In the investigated scenarios, G-computation associated with the super learner was a performant method for drawing causal inferences, even from small sample sizes.


2013 ◽  
Vol 113 (1) ◽  
pp. 221-224 ◽  
Author(s):  
David R. Johnson ◽  
Lauren K. Bachan

In a recent article, Regan, Lakhanpal, and Anguiano (2012) highlighted the lack of evidence for different relationship outcomes between arranged and love-based marriages. Yet the sample size ( n = 58) used in the study is insufficient for making such inferences. This reply discusses and demonstrates how small sample sizes reduce the utility of this research.


Sign in / Sign up

Export Citation Format

Share Document