In this chapter, you are going to learn about The relationship between power, effect size and probability levels, The factors influencing power, Issues surrounding the use of significance levels
Pitfalls of NHST: Offers a rule-based frameworks for deciding whether to believe a hypothesis, Seems to provide an easy way to disentangle the 'correct' conclusion from the 'incorrect' one
Meehl: '"The almost universal reliance on merely refuting the null hypothesis is a terrible mistake, is basically unsound, poor scientific strategy, and one of the worst things that ever happened in the history of psychology."'
All the mean differences show a positive effect of antiSTATic; therefore, we have consistent evidence that antiSTATic works.
Four of the studies show a significant result (p < 0.05), but the other six do not. Therefore, the studies are inconclusive: some suggest that antiSTATic is better than placebo, but others suggest there's no difference. The fact that more than half of the studies showed no significant effect means that antiSTATic is not (on balance) more successful in reducing anxiety than the control
Looking at the confidence intervals rather than focusing on significance allows us to see the consistency in the data and not a bunch of apparently conflicting results
Researcher degrees of freedom - a scientist has many decisions to make when designing and analyzing a study, which could be misused to exclude cases to make the result significant
Practices that lead to the selective reporting of significant p-value, most commonly trying multiple analyses and reporting only the one that yields significant results
Two experiments with identical means and standard deviations yield identical conclusions when using an effect size to interpret them (both studies had d = −0.667)
Effect sizes are less affected than p-values by things like early or late termination of data collection, or sampling over a time period rather than until a set sample size is reached
There are still some researcher degrees of freedom (not related to sample size) that researchers could use to maximize (or minimize) effect sizes, but there is less incentive to do so because effect sizes are not tied to a decision rule in which effects either side of a certain threshold have qualitatively opposite interpretations