Tutorial

In Defense of 'Circling p-values'

You noticed that the p-values in a paper are close to .05. Does this make you a bad person? Can 'circling p-values' ever be a useful heuristic? My two cents on an (apparently?) controversial topic.

Today I F-ed Up: Two Errors I Recently Made In an Almost-Published Paper

Let's talk about errors in research, using myself as an example.

Designing More Efficient Studies with Group Sequential Designs

Creating studies that are powered to detect the smallest effect of interest... without collecting more data than you need to detect bigger effects

Four Lessons from Reproducing JCR Studies

What happens when two researchers attempt to reproduce all the pre-registered studies with open data published in JCR?

Increasing the Informativeness of Power Analysis

If power analysis cannot be based on the expected effect size, what should it be based on?

Power Analysis in Consumer Research: A Fool’s Errand?

Why power analysis, as traditionally performed, isn't a good tool for choosing sample size.

Large P-Values Cannot be Explained by Power Analysis

Can p-values be close to .05 because researchers ran careful power analysis, and collected 'just enough' participants to detect their effect? In this blog post, I evaluate the merits of this argument.

A Bad Pre-Registration is Better than No Pre-Registration

A short blog post showing the immense benefits of pre-registration on false-positive rates, even when the pre-registration is underspecified.

A Critical Perspective on Effect Sizes

What can we learn from effect sizes, and under which conditions?

Making Discussions of Statistical Evidence Less Awkward and More Constructive

How can researchers respectfully and constructively flag inadequate statistical evidence when we they see it in papers? This post offers some personal reflections on this complex question.