Copyright 2016 Allen B. Downey

MIT License: https://opensource.org/licenses/MIT

```
In [1]:
```from __future__ import print_function, division
%matplotlib inline
import numpy as np
import nsfg
import first
import thinkstats2
import thinkplot

Again, I'll load the NSFG pregnancy file and select live births:

```
In [2]:
```preg = nsfg.ReadFemPreg()
live = preg[preg.outcome == 1]

Here's the histogram of birth weights:

```
In [3]:
```hist = thinkstats2.Hist(live.birthwgt_lb, label='birthwgt_lb')
thinkplot.Hist(hist)
thinkplot.Config(xlabel='Birth weight (pounds)', ylabel='Count')

```
```

To normalize the distribution, we could divide through by the total count:

```
In [4]:
```n = hist.Total()
pmf = hist.Copy()
for x, freq in hist.Items():
pmf[x] = freq / n

The result is a Probability Mass Function (PMF).

```
In [5]:
```thinkplot.Hist(pmf)
thinkplot.Config(xlabel='Birth weight (pounds)', ylabel='PMF')

```
```

More directly, we can create a Pmf object.

```
In [6]:
```pmf = thinkstats2.Pmf([1, 2, 2, 3, 5])
pmf

```
Out[6]:
```

`Pmf`

provides `Prob`

, which looks up a value and returns its probability:

```
In [7]:
```pmf.Prob(2)

```
Out[7]:
```

The bracket operator does the same thing.

```
In [8]:
```pmf[2]

```
Out[8]:
```

The `Incr`

method adds to the probability associated with a given values.

```
In [9]:
```pmf.Incr(2, 0.2)
pmf[2]

```
Out[9]:
```

The `Mult`

method multiplies the probability associated with a value.

```
In [10]:
```pmf.Mult(2, 0.5)
pmf[2]

```
Out[10]:
```

`Total`

returns the total probability (which is no longer 1, because we changed one of the probabilities).

```
In [11]:
```pmf.Total()

```
Out[11]:
```

`Normalize`

divides through by the total probability, making it 1 again.

```
In [12]:
```pmf.Normalize()
pmf.Total()

```
Out[12]:
```

Here's the PMF of pregnancy length for live births.

```
In [13]:
```pmf = thinkstats2.Pmf(live.prglngth, label='prglngth')

Here's what it looks like plotted with `Hist`

, which makes a bar graph.

```
In [14]:
```thinkplot.Hist(pmf)
thinkplot.Config(xlabel='Pregnancy length (weeks)', ylabel='Pmf')

```
```

Here's what it looks like plotted with `Pmf`

, which makes a step function.

```
In [15]:
```thinkplot.Pmf(pmf)
thinkplot.Config(xlabel='Pregnancy length (weeks)', ylabel='Pmf')

```
```

We can use `MakeFrames`

to return DataFrames for all live births, first babies, and others.

```
In [16]:
```live, firsts, others = first.MakeFrames()

Here are the distributions of pregnancy length.

```
In [17]:
```first_pmf = thinkstats2.Pmf(firsts.prglngth, label='firsts')
other_pmf = thinkstats2.Pmf(others.prglngth, label='others')

And here's the code that replicates one of the figures in the chapter.

```
In [18]:
```width=0.45
axis = [27, 46, 0, 0.6]
thinkplot.PrePlot(2, cols=2)
thinkplot.Hist(first_pmf, align='right', width=width)
thinkplot.Hist(other_pmf, align='left', width=width)
thinkplot.Config(xlabel='Pregnancy length(weeks)', ylabel='PMF', axis=axis)
thinkplot.PrePlot(2)
thinkplot.SubPlot(2)
thinkplot.Pmfs([first_pmf, other_pmf])
thinkplot.Config(xlabel='Pregnancy length(weeks)', axis=axis)

```
```

```
In [19]:
```weeks = range(35, 46)
diffs = []
for week in weeks:
p1 = first_pmf.Prob(week)
p2 = other_pmf.Prob(week)
diff = 100 * (p1 - p2)
diffs.append(diff)
thinkplot.Bar(weeks, diffs)
thinkplot.Config(xlabel='Pregnancy length(weeks)', ylabel='Difference (percentage points)')

```
```

```
In [20]:
```d = { 7: 8, 12: 8, 17: 14, 22: 4,
27: 6, 32: 12, 37: 8, 42: 3, 47: 2 }
pmf = thinkstats2.Pmf(d, label='actual')

```
In [21]:
```def BiasPmf(pmf, label):
new_pmf = pmf.Copy(label=label)
for x, p in pmf.Items():
new_pmf.Mult(x, x)
new_pmf.Normalize()
return new_pmf

The following graph shows the difference between the actual and observed distributions.

```
In [22]:
```biased_pmf = BiasPmf(pmf, label='observed')
thinkplot.PrePlot(2)
thinkplot.Pmfs([pmf, biased_pmf])
thinkplot.Config(xlabel='Class size', ylabel='PMF')

```
```

The observed mean is substantially higher than the actual.

```
In [23]:
```print('Actual mean', pmf.Mean())
print('Observed mean', biased_pmf.Mean())

```
```

```
In [24]:
```def UnbiasPmf(pmf, label=None):
new_pmf = pmf.Copy(label=label)
for x, p in pmf.Items():
new_pmf[x] *= 1/x
new_pmf.Normalize()
return new_pmf

We can unbias the biased PMF:

```
In [25]:
```unbiased = UnbiasPmf(biased_pmf, label='unbiased')
print('Unbiased mean', unbiased.Mean())

```
```

And plot the two distributions to confirm they are the same.

```
In [26]:
```thinkplot.PrePlot(2)
thinkplot.Pmfs([pmf, unbiased])
thinkplot.Config(xlabel='Class size', ylabel='PMF')

```
```

```
In [27]:
```import numpy as np
import pandas
array = np.random.randn(4, 2)
df = pandas.DataFrame(array)
df

```
Out[27]:
```

We can specify column names when we create the DataFrame:

```
In [28]:
```columns = ['A', 'B']
df = pandas.DataFrame(array, columns=columns)
df

```
Out[28]:
```

We can also specify an index that contains labels for the rows.

```
In [29]:
```index = ['a', 'b', 'c', 'd']
df = pandas.DataFrame(array, columns=columns, index=index)
df

```
Out[29]:
```

Normal indexing selects columns.

```
In [30]:
```df['A']

```
Out[30]:
```

We can use the `loc`

attribute to select rows.

```
In [31]:
```df.loc['a']

```
Out[31]:
```

`iloc`

attribute:

```
In [32]:
```df.iloc[0]

```
Out[32]:
```

`loc`

can also take a list of labels.

```
In [33]:
```indices = ['a', 'c']
df.loc[indices]

```
Out[33]:
```

If you provide a slice of labels, `DataFrame`

uses it to select rows.

```
In [34]:
```df['a':'c']

```
Out[34]:
```

If you provide a slice of integers, `DataFrame`

selects rows by integer index.

```
In [35]:
```df[0:2]

```
Out[35]:
```

But notice that one method includes the last elements of the slice and one does not.

In general, I recommend giving labels to the rows and names to the columns, and using them consistently.

**Exercise:** Something like the class size paradox appears if you survey children and ask how many children are in their family. Families with many children are more likely to appear in your sample, and families with no children have no chance to be in the sample.

Use the NSFG respondent variable `numkdhh`

to construct the actual distribution for the number of children under 18 in the respondents' households.

Now compute the biased distribution we would see if we surveyed the children and asked them how many children under 18 (including themselves) are in their household.

Plot the actual and biased distributions, and compute their means.

```
In [36]:
```resp = nsfg.ReadFemResp()

```
In [37]:
``````
# Solution goes here
```

```
In [38]:
``````
# Solution goes here
```

```
In [39]:
``````
# Solution goes here
```

```
In [40]:
``````
# Solution goes here
```

```
In [41]:
``````
# Solution goes here
```

```
In [42]:
``````
# Solution goes here
```

**Exercise:** I started this book with the question, "Are first babies more likely to be late?" To address it, I computed the difference in means between groups of babies, but I ignored the possibility that there might be a difference between first babies and others for the same woman.

To address this version of the question, select respondents who have at least two live births and compute pairwise differences. Does this formulation of the question yield a different result?

Hint: use `nsfg.MakePregMap`

:

```
In [43]:
```live, firsts, others = first.MakeFrames()

```
In [44]:
```preg_map = nsfg.MakePregMap(live)

```
In [45]:
``````
# Solution goes here
```

```
In [46]:
``````
# Solution goes here
```

```
In [47]:
``````
# Solution goes here
```

**Exercise:** In most foot races, everyone starts at the same time. If you are a fast runner, you usually pass a lot of people at the beginning of the race, but after a few miles everyone around you is going at the same speed.
When I ran a long-distance (209 miles) relay race for the first time, I noticed an odd phenomenon: when I overtook another runner, I was usually much faster, and when another runner overtook me, he was usually much faster.

At first I thought that the distribution of speeds might be bimodal; that is, there were many slow runners and many fast runners, but few at my speed.

Then I realized that I was the victim of a bias similar to the effect of class size. The race was unusual in two ways: it used a staggered start, so teams started at different times; also, many teams included runners at different levels of ability.

As a result, runners were spread out along the course with little relationship between speed and location. When I joined the race, the runners near me were (pretty much) a random sample of the runners in the race.

So where does the bias come from? During my time on the course, the chance of overtaking a runner, or being overtaken, is proportional to the difference in our speeds. I am more likely to catch a slow runner, and more likely to be caught by a fast runner. But runners at the same speed are unlikely to see each other.

Write a function called `ObservedPmf`

that takes a `Pmf`

representing the actual distribution of runners’ speeds, and the speed of a running observer, and returns a new `Pmf`

representing the distribution of runners’ speeds as seen by the observer.

To test your function, you can use `relay.py`

, which reads the results from the James Joyce Ramble 10K in Dedham MA and converts the pace of each runner to mph.

Compute the distribution of speeds you would observe if you ran a relay race at 7 mph with this group of runners.

```
In [48]:
```import relay
results = relay.ReadResults()
speeds = relay.GetSpeeds(results)
speeds = relay.BinData(speeds, 3, 12, 100)

```
In [49]:
```pmf = thinkstats2.Pmf(speeds, 'actual speeds')
thinkplot.Pmf(pmf)
thinkplot.Config(xlabel='Speed (mph)', ylabel='PMF')

```
```

```
In [50]:
``````
# Solution goes here
```

```
In [51]:
``````
# Solution goes here
```

```
In [ ]:
```