Effect Size

Examples and exercises for a tutorial on statistical inference.

Copyright 2016 Allen Downey

License: Creative Commons Attribution 4.0 International


In [ ]:
%matplotlib inline

import numpy
import scipy.stats

import matplotlib.pyplot as plt

from ipywidgets import interact, interactive, fixed
import ipywidgets as widgets

# seed the random number generator so we all get the same results
numpy.random.seed(17)

Part One

To explore statistics that quantify effect size, we'll look at the difference in height between men and women. I used data from the Behavioral Risk Factor Surveillance System (BRFSS) to estimate the mean and standard deviation of height in cm for adult women and men in the U.S.

I'll use scipy.stats.norm to represent the distributions. The result is an rv object (which stands for random variable).


In [ ]:
mu1, sig1 = 178, 7.7
male_height = scipy.stats.norm(mu1, sig1)

In [ ]:
mu2, sig2 = 163, 7.3
female_height = scipy.stats.norm(mu2, sig2)

The following function evaluates the normal (Gaussian) probability density function (PDF) within 4 standard deviations of the mean. It takes and rv object and returns a pair of NumPy arrays.


In [ ]:
def eval_pdf(rv, num=4):
    mean, std = rv.mean(), rv.std()
    xs = numpy.linspace(mean - num*std, mean + num*std, 100)
    ys = rv.pdf(xs)
    return xs, ys

Here's what the two distributions look like.


In [ ]:
xs, ys = eval_pdf(male_height)
plt.plot(xs, ys, label='male', linewidth=4, color='C0')

xs, ys = eval_pdf(female_height)
plt.plot(xs, ys, label='female', linewidth=4, color='C1')
plt.xlabel('height (cm)');

Let's assume for now that those are the true distributions for the population.

I'll use rvs to generate random samples from the population distributions. Note that these are totally random, totally representative samples, with no measurement error!


In [ ]:
male_sample = male_height.rvs(1000)

In [ ]:
female_sample = female_height.rvs(1000)

Both samples are NumPy arrays. Now we can compute sample statistics like the mean and standard deviation.


In [ ]:
mean1, std1 = male_sample.mean(), male_sample.std()
mean1, std1

The sample mean is close to the population mean, but not exact, as expected.


In [ ]:
mean2, std2 = female_sample.mean(), female_sample.std()
mean2, std2

And the results are similar for the female sample.

Now, there are many ways to describe the magnitude of the difference between these distributions. An obvious one is the difference in the means:


In [ ]:
difference_in_means = male_sample.mean() - female_sample.mean()
difference_in_means # in cm

On average, men are 14--15 centimeters taller. For some applications, that would be a good way to describe the difference, but there are a few problems:

  • Without knowing more about the distributions (like the standard deviations) it's hard to interpret whether a difference like 15 cm is a lot or not.

  • The magnitude of the difference depends on the units of measure, making it hard to compare across different studies.

There are a number of ways to quantify the difference between distributions. A simple option is to express the difference as a percentage of the mean.

Exercise 1: what is the relative difference in means, expressed as a percentage?


In [ ]:
# Solution goes here

relative_difference = difference_in_means / male_sample.mean()
print(relative_difference * 100)   # percent

# A problem with relative differences is that you have to choose 
# which mean to express them relative to.

relative_difference = difference_in_means / female_sample.mean()
print(relative_difference * 100)   # percent

STOP HERE: We'll regroup and discuss before you move on.

Part Two

An alternative way to express the difference between distributions is to see how much they overlap. To define overlap, we choose a threshold between the two means. The simple threshold is the midpoint between the means:


In [ ]:
simple_thresh = (mean1 + mean2) / 2
simple_thresh

A better, but slightly more complicated threshold is the place where the PDFs cross.


In [ ]:
thresh = (std1 * mean2 + std2 * mean1) / (std1 + std2)
thresh

In this example, there's not much difference between the two thresholds.

Now we can count how many men are below the threshold:


In [ ]:
male_below_thresh = sum(male_sample < thresh)
male_below_thresh

And how many women are above it:


In [ ]:
female_above_thresh = sum(female_sample > thresh)
female_above_thresh

The "overlap" is the area under the curves that ends up on the wrong side of the threshold.


In [ ]:
male_overlap = male_below_thresh / len(male_sample)
female_overlap = female_above_thresh / len(female_sample)
male_overlap, female_overlap

In practical terms, you might report the fraction of people who would be misclassified if you tried to use height to guess sex, which is the average of the male and female overlap rates:


In [ ]:
misclassification_rate = (male_overlap + female_overlap) / 2
misclassification_rate

Another way to quantify the difference between distributions is what's called "probability of superiority", which is a problematic term, but in this context it's the probability that a randomly-chosen man is taller than a randomly-chosen woman.

Exercise 2: Suppose I choose a man and a woman at random. What is the probability that the man is taller?

HINT: You can zip the two samples together and count the number of pairs where the male is taller, or use NumPy array operations.


In [ ]:
# Solution goes here

sum(x > y for x, y in zip(male_sample, female_sample)) / len(male_sample)

In [ ]:
# Solution goes here

(male_sample > female_sample).mean()

Overlap (or misclassification rate) and "probability of superiority" have two good properties:

  • As probabilities, they don't depend on units of measure, so they are comparable between studies.

  • They are expressed in operational terms, so a reader has a sense of what practical effect the difference makes.

Cohen's effect size

There is one other common way to express the difference between distributions. Cohen's $d$ is the difference in means, standardized by dividing by the standard deviation. Here's the math notation:

$ d = \frac{\bar{x}_1 - \bar{x}_2} s $

where $s$ is the pooled standard deviation:

$s = \sqrt{\frac{n_1 s^2_1 + n_2 s^2_2}{n_1+n_2}}$

Here's a function that computes it:


In [ ]:
def CohenEffectSize(group1, group2):
    """Compute Cohen's d.

    group1: Series or NumPy array
    group2: Series or NumPy array

    returns: float
    """
    diff = group1.mean() - group2.mean()

    n1, n2 = len(group1), len(group2)
    var1 = group1.var()
    var2 = group2.var()

    pooled_var = (n1 * var1 + n2 * var2) / (n1 + n2)
    d = diff / numpy.sqrt(pooled_var)
    return d

Computing the denominator is a little complicated; in fact, people have proposed several ways to do it. This implementation uses the "pooled standard deviation", which is a weighted average of the standard deviations of the two groups.

And here's the result for the difference in height between men and women.


In [ ]:
CohenEffectSize(male_sample, female_sample)

Most people don't have a good sense of how big $d=1.9$ is, so let's make a visualization to get calibrated.

Here's a function that encapsulates the code we already saw for computing overlap and probability of superiority.


In [ ]:
def overlap_superiority(control, treatment, n=1000):
    """Estimates overlap and superiority based on a sample.
    
    control: scipy.stats rv object
    treatment: scipy.stats rv object
    n: sample size
    """
    control_sample = control.rvs(n)
    treatment_sample = treatment.rvs(n)
    thresh = (control.mean() + treatment.mean()) / 2
    
    control_above = sum(control_sample > thresh)
    treatment_below = sum(treatment_sample < thresh)
    overlap = (control_above + treatment_below) / n
    
    superiority = (treatment_sample > control_sample).mean()
    return overlap, superiority

Here's the function that takes Cohen's $d$, plots normal distributions with the given effect size, and prints their overlap and superiority.


In [ ]:
def plot_pdfs(cohen_d=2):
    """Plot PDFs for distributions that differ by some number of stds.
    
    cohen_d: number of standard deviations between the means
    """
    control = scipy.stats.norm(0, 1)
    treatment = scipy.stats.norm(cohen_d, 1)
    xs, ys = eval_pdf(control)
    plt.fill_between(xs, ys, label='control', color='C1', alpha=0.5)

    xs, ys = eval_pdf(treatment)
    plt.fill_between(xs, ys, label='treatment', color='C0', alpha=0.5)
    
    o, s = overlap_superiority(control, treatment)
    plt.text(0, 0.05, 'overlap ' + str(o))
    plt.text(0, 0.15, 'superiority ' + str(s))
    plt.show()
    #print('overlap', o)
    #print('superiority', s)

Here's an example that demonstrates the function:


In [ ]:
plot_pdfs(2)

And an interactive widget you can use to visualize what different values of $d$ mean:


In [ ]:
slider = widgets.FloatSlider(min=0, max=4, value=2)
interact(plot_pdfs, cohen_d=slider);

Cohen's $d$ has a few nice properties:

  • Because mean and standard deviation have the same units, their ratio is dimensionless, so we can compare $d$ across different studies.

  • In fields that commonly use $d$, people are calibrated to know what values should be considered big, surprising, or important.

  • Given $d$ (and the assumption that the distributions are normal), you can compute overlap, superiority, and related statistics.

In summary, the best way to report effect size depends on the audience and your goals. There is often a tradeoff between summary statistics that have good technical properties and statistics that are meaningful to a general audience.


In [ ]: