This HTML version of is provided for convenience, but it is not the best format for the book. In particular, some of the symbols are not rendered correctly.

You might prefer to read the PDF version, or you can buy a hardcopy here.

The distributions we have used so far are called **empirical distributions**
because they are based on empirical observations, which are necessarily finite
samples.

The alternative is a **continuous distribution**, which is characterized by a
CDF that is a continuous function (as opposed to a step function). Many real
world phenomena can be approximated by continuous distributions.

I’ll start with the exponential distribution because it is easy to work with.
In the real world, exponential distributions come up when we look at a series
of events and measure the times between events, which are called
**interarrival times**. If the events are equally likely to occur at any time,
the distribution of interarrival times tends to look like an exponential
distribution.

The CDF of the exponential distribution is:

$$ CDF(x) = 1 − e^{−\lambda x} $$The parameter, $\lambda$, determines the shape of the distribution. Figure 4.1 shows what this CDF looks like with $\lambda = 2$.

```
In [1]:
```%matplotlib inline
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
# plot the exponential distribution
def exp_cdf(x,l):
return 1-np.exp(-l*x)
X = np.linspace(0,2.5,100)
pd.DataFrame(exp_cdf(X,2), X).plot()

```
Out[1]:
```

In general, the mean of an exponential distribution is $\frac{1}{\lambda}$, so the mean of this distribution is 0.5. The median is $\frac{log(2)}{\lambda}$, which is roughly 0.35.

To see an example of a distribution that is approximately exponential, we will
look at the interarrival time of babies. On December 18, 1997, 44 babies were
born in a hospital in Brisbane, Australia(This example is based on information and data from Dunn, “A Simple Dataset for Demonstrating Common Distributions,” Journal of Statistics Education v.7, n.3 (1999)). The times of birth for all 44
babies were reported in the local paper; you can download the data from
`http://thinkstats.com/babyboom.dat`

.

Figure 4.2 shows the CDF of the interarrival times in minutes. It seems to have the general shape of an exponential distribution, but how can we tell?

One way is to plot the complementary CDF, 1 − CDF($x$), on a log-$y$ scale. For data from an exponential distribution, the result is a straight line. Let's see why that works.

If you plot the complementary CDF (CCDF) of a dataset that you think is exponential, you expect to see a function like:

$$ y \approx e^{-\lambda x} $$Taking the log of both sides yields:

$$ log(y) \approx -\lambda x$$So on a log-*y* scale the CCDF is a straight line with slope $-\lambda$.

```
In [2]:
```raw_data = open("data/babyboom.txt", 'r').read()
# cleaning and creating a dataframe off this data

```
In [3]:
```babyboom = pd.DataFrame([[int(y) for y in x.strip().split()] for x in raw_data.split('\n')],
columns = ['Time in 24hr', 'Sex', 'Weight (grams)', 'Mins after Midnight'])
# sex: 1 for girl, 2 for boy

```
In [4]:
```# plot of time between births
X = sorted(babyboom['Mins after Midnight']-babyboom['Mins after Midnight'].shift(1))[1:]
plt.step(X, np.linspace(0,1, len(X)), label='post')
plt.show()

```
```

```
In [5]:
```plt.step(X[::-1], np.linspace(0,1, len(X)), label='post')
plt.xscale('log')
plt.show()

```
```

Figure 4.3 shows the CCDF of the interarrivals on a log-*y* scale. It is not
exactly straight, which suggests that the exponential distribution is only an
approximation. Most likely the underlying assumption—that a birth is equally
likely at any time of day—is not exactly true.

**Exercise 1** For small values of $n$, we don’t expect an empirical distribution to fit a continuous distribution exactly. One way to evaluate the quality of fit is to generate a sample from a continuous distribution and see how well it matches the data.

The function `expovariate`

in the `random`

module generates random values
from an exponential distribution with a given value of λ. Use it to generate
44 values from an exponential distribution with mean 32.6. Plot the CCDF on a
log-$y$ scale and compare it to Figure 4.3.

Hint: You can use the function `pyplot.yscale`

to plot the $y$ axis on a log
scale.

Or, if you use `myplot`

, the `Cdf`

function takes a boolean option,
`complement`

, that determines whether to plot the CDF or CCDF, and string
options, `xscale`

and `yscale`

, that transform the axes; to plot a CCDF on a
log-$y$ scale:

` myplot.Cdf(cdf, complement=True, xscale='linear', yscale='log') `

```
In [6]:
```# Exercise 1 Solution
from random import expovariate
X = [expovariate(32.6) for _ in xrange(44)]
X = sorted(X)
plt.step(X[::-1], np.linspace(0,1, len(X)), label='post')

```
Out[6]:
```

**Exercise 2** Collect the birthdays of the students in your class, sort them, and compute the interarrival times in days. Plot the CDF of the interarrival times and the CCDF on a log-$y$ scale. Does it look like an exponential distribution?

The Pareto distribution is named after the economist Vilfredo Pareto, who used
it to describe the distribution of wealth (see
`http://wikipedia.org/wiki/Pareto_distribution`

). Since then, it has been used
to describe phenomena in the natural and social sciences including sizes of
cities and towns, sand particles and meteorites, forest fires and earthquakes.

The CDF of the Pareto distribution is:

$$CDF(x) = 1- \left(\frac{x}{x_m}\right)^{-\alpha} $$The parameters $x_m$ and $\alpha$ determine the location and shape of the distribution. $x_m$ is the minimum possible value. Figure 4.4 shows the CDF of a Pareto distribution with parameters $x_m = 0.5$ and $\alpha = 1$.

```
In [7]:
```def pareto_cdf(x, x_m, a):
return 1-((x/x_m)**(-a))
X = np.linspace(0.5,10, 100)
pd.DataFrame(pareto_cdf(X,0.5,1), X).plot()

```
Out[7]:
```

The median of this distribution is $x_m 2^{1/\alpha}$ , which is 1, but the 95th percentile is 10. By contrast, the exponential distribution with median 1 has 95th percentile of only 1.5.

There is a simple visual test that indicates whether an empirical distribution fits a Pareto distribution: on a log-log scale, the CCDF looks like a straight line. If you plot the CCDF of a sample from a Pareto distribution on a linear scale, you expect to see a function like:

$$ y \approx \left( \frac{x}{x_m} \right)^{-\alpha}$$Taking the log of both sides yields:

$$ log y \approx \alpha (log(x) - log(x_m)) $$So if you plot log $y$ versus log $x$, it should look like a straight line with slope $−\alpha$ and intercept $\alpha log(x_m) $.

**Exercise 3** The `random`

module provides `paretovariate`

, which generates random values from a Pareto distribution. It takes a parameter for $\alpha$, but not $x_m$. The default value for $x_m$ is 1; you can generate a distribution with a different parameter by multiplying by $x_m$.

Write a wrapper function named `paretovariate`

that takes $\alpha$ and $x_m$ as
parameters and uses `random.paretovariate`

to generate values from a two-
parameter Pareto distribution.

Use your function to generate a sample from a Pareto distribution. Compute the CCDF and plot it on a log-log scale. Is it a straight line? What is the slope?

```
In [8]:
```# excercise 3
import random
def paretovariate(a, xm):
"""
Generates a random pareto variable, we shape `a` and scale `xm`.
"""
return random.paretovariate(a)*xm

```
In [9]:
```X = [paretovariate(2,2) for _ in xrange(44)]
X = sorted(X)
plt.step(X[::-1], np.linspace(0,1, len(X)), label='post')

```
Out[9]:
```

**Exercise 4** To get a feel for the Pareto distribution, imagine what the world would be like if the distribution of human height were Pareto. Choosing the parameters $x_m = 100 cm$ and $\alpha = 1.7$, we get a distribution with a reasonable minimum, 100 cm, and median, 150 cm.

Generate 6 million random values from this distribution. What is the mean of this sample? What fraction of the population is shorter than the mean? How tall is the tallest person in Pareto World?

```
In [10]:
```X = [paretovariate(1.7,100) for _ in xrange(6000000)]
mX = np.mean(X)
print "Mean is: %.2f" % (mX/100)
print "Fraction less than mean is: %.2f" % (float(sum([1 for x in X if x < mX]))/6000000.0)
print "Tallest person is %.2f m" % (max(X)/100.0)

```
```

**Exercise 5** Zipf’s law is an observation about how often different words are used. The most common words have very high frequencies, but there are many unusual words, like “hapaxlegomenon,” that appear only a few times. Zipf’s law predicts that in a body of text, called a “corpus,” the distribution of word frequencies is roughly Pareto.

Find a large corpus, in any language, in electronic format. Count how many times each word appears. Find the CCDF of the word counts and plot it on a log-log scale. Does Zipf’s law hold? What is the value of $\alpha$, approximately?

```
In [11]:
```# we shall use Holy Bible KJV as our corpus.
import re
bible = "data/HolyBible-KJV.txt"
dat = open(bible).read().lower()
dat = re.sub(r"[^a-z' ]", ' ', dat)
dat = re.sub(r"'", '', dat) #remove apostrophe 's
dat = re.sub(r'\s+', ' ', dat)
wordfreq = {}
for word in dat.split():
if len(word) < 3: continue
wordfreq[word] = wordfreq.get(word, 0) + 1
wordf = wordfreq.values()

```
In [12]:
```# then use the wordfreq.values() to plot histogram and estimate alpha
# using MLE estimator for xm and alpha we get:
xm = min(wordf)
n = len(wordf)
a = float(n)/sum(np.log(wordf) - np.log(xm))

```
In [13]:
```# plot CCDF
X = sorted(wordf)
plt.step(X[::-1], np.linspace(0,1, len(X)), label='post')
plt.xscale('log')
plt.show()

```
```

**Exercise 6**

The Weibull distribution is a generalization of the exponential distribution that comes up in failure analysis (see http://wikipedia.org/wiki/Weibull_distribution). Its CDF is

$$ CDF(x) = 1 − e^{(x/\lambda)k} $$Can you find a transformation that makes a Weibull distribution look like a straight line? What do the slope and intercept of the line indicate?

Use `random.weibullvariate`

to generate a sample from a Weibull distribution
and use it to test your transformation.

```
In [14]:
```shape = 2
scale = 1 # scale doesn't matter for the transformation
t = [random.weibullvariate(scale, shape) for x in range(100000)]
plt.hist(t, bins=100)
plt.show()

```
```

```
In [15]:
```def transform(x, k):
"""
In order to create a straight line, consider that to generate a weibull variate, the transformation would be:
(this is the quantile or inverse cumulative distribution)
X = scale * (-log(U))^(1/shape)
where U ~ Uniform(0,1)
rearranging and considering that scale only makes the range greater or smaller (i.e. has no effect whether the line is "straight"
or not) you can see that the tranformation is:
U is proportional to exp(x^shape)
"""
return np.exp(-x**k)
tt=[transform(x, shape) for x in t]
plt.hist(tt,bins=100)
plt.show()

```
```

The normal distribution, also called Gaussian, is the most commonly used because it describes so many phenomena, at least approximately. It turns out that there is a good reason for its ubiquity, which we will get to in Section 6.6.

The normal distribution has many properties that make it amenable for
analysis, but the CDF is not one of them. Unlike the other distributions we
have looked at, there is no closed-form expression for the normal CDF; the
most common alternative is to write it in terms of the **error function**,
which is a special function written erf($x$):

The parameters $\mu$ and $\sigma$ determine the mean and standard deviation of the distribution.

If these formulas make your eyes hurt, don’t worry; they are easy to implement
in Python. There are many fast and accurate ways to approximate erf($x$). You
can download one of them from `http://thinkstats.com/erf.py`

, which provides
functions named `erf`

and `NormalCdf`

.

Figure 4.5 shows the CDF of the normal distribution with parameters $\mu = 2.0$ and $\sigma = 0.5 $. The sigmoid shape of this curve is a recognizable characteristic of a normal distribution.

```
In [16]:
```from scipy.stats import norm
X = np.linspace(0.0,4.0, 100000)
t = norm.cdf(X, 2.0, 0.5)
pd.DataFrame(t, X).plot()

```
Out[16]:
```

```
In [17]:
```import gzip
def makeRecord(line, fields):
obs = {}
for (field, start, end, cast) in fields:
try:
s = line[start-1:end]
val = cast(s)
except ValueError:
val = np.nan #make use of numpy's nan
obs[field]=val
return obs
fresp = gzip.open('./data/2002FemResp.dat.gz')
resp_fields = [
('caseid', 1, 12, int),
]
fpreg = gzip.open('./data/2002FemPreg.dat.gz')
preg_fields = [ ('caseid', 1, 12, int),
('nbrnaliv', 22, 22, int),
('babysex', 56, 56, int),
('birthwgt_lb', 57, 58, int),
('birthwgt_oz', 59, 60, int),
('prglength', 275, 276, int),
('outcome', 277, 277, int),
('birthord', 278, 279, int),
('agepreg', 284, 287, int),
('finalwgt', 423, 440, float)]
respondents = pd.DataFrame([makeRecord(line, resp_fields) for line in fresp])
pregnancies = pd.DataFrame([makeRecord(line, preg_fields) for line in fpreg])
#recode
pregnancies['agepreg'] = pregnancies.agepreg/100.0
pregnancies['totalwgt_oz'] = pregnancies.birthwgt_lb * 16 + pregnancies.birthwgt_oz

```
In [18]:
```tot = [x for x in pregnancies['totalwgt_oz'] if not np.isnan(x) and x < 250]
tot = sorted(tot)
X = np.linspace(0.0, 250.0, len(tot))
Y = norm.cdf(X, np.mean(tot), np.std(tot))
plt.step(tot, np.linspace(0,1, len(tot)), label='post')
plt.plot(X, Y)
plt.show()

```
```

The normal distribution is a good model for this dataset. A **model** is a
useful simplification. In this case it is useful because we can summarize the
entire distribution with just two numbers, $\mu$ = 116.5 and $\sigma$ = 19.9, and the
resulting error (difference between the model and the data) is small.

Below the 10th percentile there is a discrepancy between the data and the model; there are more light babies than we would expect in a normal distribution. If we are interested in studying preterm babies, it would be important to get this part of the distribution right, so it might not be appropriate to use the normal model.

**Exercise 7** The Wechsler Adult Intelligence Scale is a test that is intended to measure intelligence. Results are transformed so that the distribution of scores in the general population is normal with µ = 100 and σ = 15.

Use `erf.NormalCdf`

to investigate the frequency of rare events in a normal
distribution. What fraction of the population has an IQ greater than the mean?
What fraction is over 115? 130? 145?

A “six-sigma” event is a value that exceeds the mean by 6 standard deviations, so a six-sigma IQ is 190. In a world of 6 billion people, how many do we expect to have an IQ of 190 or more?

```
In [19]:
```def px_over(iq):
return 1-norm.cdf(iq, 100, 15)
print "IQ greater than 115: %.2f" % px_over(115)
print "IQ greater than 130: %.2f" % px_over(130)
print "IQ greater than 145: %.2f" % px_over(145)

```
```

```
In [20]:
```print "approximately %.0f people have an IQ 190 or greater " % (6000000000*(1-norm.cdf(190, 100, 15)))

```
```

**Exercise 8** Plot the CDF of pregnancy lengths for all live births. Does it look like a normal distribution?

Compute the mean and standard deviation of the sample and plot the normal distribution with the same parameters. Is the normal distribution a good model for this data? If you had to summarize this distribution with two statistics, what statistics would you choose?

```
In [21]:
```tot = [x for x in pregnancies['prglength'] if not np.isnan(x)]
tot = sorted(tot)
X = np.linspace(0.0, 50.0, len(tot))
Y = norm.cdf(X, np.mean(tot), np.std(tot))
plt.step(tot, np.linspace(0,1, len(tot)), label='post')
plt.plot(X, Y)
plt.show()

```
```

```
In [22]:
```plt.hist(tot)
# probably the two most frequent groupings of time since it is bimodal?

```
Out[22]:
```

For the exponential, Pareto and Weibull distributions, there are simple transformations we can use to test whether a continuous distribution is a good model of a dataset.

For the normal distribution there is no such transformation, but there is an
alternative called a **normal probability plot**. It is based on **rankits**:
if you generate $n$ values from a normal distribution and sort them, the $k$th
rankit is the mean of the distribution for the $k$th value.

**Exercise 9** Write a function called `Sample`

that generates 6 samples from a normal distribution with $\mu = 0$ and $\sigma = 1$. Sort and return the values.

Write a function called `Samples`

that calls `Sample`

1000 times and returns
a list of 1000 lists.

If you apply `zip`

to this list of lists, the result is 6 lists with 1000
values in each. Compute the mean of each of these lists and print the results.
I predict that you will get something like this:

If you increase the number of times you call `Sample`

, the results should
converge on these values.

```
In [23]:
```def sample():
sample = np.random.normal(size=6)
sample = sorted(sample)
return sample
def samples():
return [sample() for x in range(1000)]
[np.mean(x) for x in zip(*samples())]

```
Out[23]:
```

Computing rankits exactly is moderately difficult, but there are numerical methods for approximating them. And there is a quick-and-dirty method that is even easier to implement:

- From a normal distribution with $\mu = 0$ and $\sigma = 1$, generate a sample with the same size as your dataset and sort it.
- Sort the values in the dataset.
- Plot the sorted values from your dataset versus the random values.

For large datasets, this method works well. For smaller datasets, you can improve it by generating $m(n+1) − 1$ values from a normal distribution, where $n$ is the size of the dataset and $m$ is a multiplier. Then select every $m$th element, starting with the $m$th.

This method works with other distributions as well, as long as you know how to generate a random sample.

Figure 4.7 is a quick-and-dirty normal probability plot for the birth weight data.

```
In [24]:
```tot = [x for x in pregnancies['totalwgt_oz'] if not np.isnan(x) and x < 250]
tot = sorted(tot)
Y = np.random.normal(size=len(tot))
Y = sorted(Y)
plt.scatter(tot,Y, s=.2)
plt.show()

```
```

The curvature in this plot suggests that there are deviations from a normal distribution; nevertheless, it is a good (enough) model for many purposes.

**Exercise 10** * Write a function called NormalPlot that takes a sequence of values and generates a normal probability plot. You can download a solution from http://thinkstats.com/rankit.py. *

_Use the running speeds from `relay.py`

to generate a normal probability plot.
Is the normal distribution a good model for this data? You can download a
solution from `http://thinkstats.com/relay_normal.py`

._

```
In [25]:
```# generate a sequence of values
gen_seq = np.random.normal(size=1000)
gen_seq = sorted(gen_seq)
#create plot...
Y = np.random.random(size=len(gen_seq))
Y = sorted(Y)
plt.scatter(gen_seq, Y, s=.2)
plt.show()

```
```

If the logarithms of a set of values have a normal distribution, the values
have a **lognormal** distribution. The CDF of the lognormal distribution is
the same as the CDF of the normal distribution, with $log(x)$ substituted for
$x$.

The parameters of the lognormal distribution are usually denoted $\mu$ and $\sigma$. But
remember that these parameters are *not* the mean and standard deviation; the
mean of a lognormal distribution is $e^{\mu + \frac{\sigma^s}{2}}$ and the standard deviation
is ugly.

It turns out that the distribution of weights for adults is approximately lognormal.

The National Center for Chronic Disease Prevention and Health Promotion conducts an annual survey as part of the Behavioral Risk Factor Surveillance System (BRFSS)(Centers for Disease Control and Prevention (CDC). Behavioral Risk Factor Surveillance System Survey Data. Atlanta, Georgia: U.S. Department of Health and Human Services, Centers for Disease Control and Prevention, 2008.). In 2008, they interviewed 414,509 respondents and asked about their demographics, health and health risks.

```
In [35]:
```adult_weight = gzip.open('./data/CDBRFS08.ASC.gz')
aw_fields = [('age', 101, 102, int),
('weight2', 119, 122, int),
('wtyrago', 127, 130, int),
('wtkg2', 1254, 1258, int),
('htm3', 1251, 1253, int),
('sex', 143, 143, int),]
aw_data = pd.DataFrame([makeRecord(line, aw_fields) for line in adult_weight])

```
In [36]:
```# recode and cleanse data
def CleanWeight(weight):
if weight in [7777, 9999]:
return 'NA'
elif weight < 1000:
return weight / 2.2
elif 9000 < weight < 9999:
return weight - 9000
else:
return weight
aw_data['wtkg2'] = aw_data['wtkg2'].apply(lambda x: np.nan if x in ['NA', 99999] else x/100.0)
aw_data['weight2'] = aw_data['weight2'].apply(CleanWeight)
aw_data['wtyrago'] = aw_data['wtyrago'].apply(CleanWeight)
aw_data['htm3'] = aw_data['htm3'].apply(lambda x: np.nan if x == 999 else x)
aw_data['age'] = aw_data['age'].apply(lambda x: np.nan if x in [7,9] else x)

```
Out[36]:
```

```
In [56]:
```weights = np.log(sorted([x for x in aw_data['weight2'] if x != 'NA' and not np.isnan(x)]))
X = np.linspace(min(weights), max(weights), len(weights))
Y = norm.cdf(X, np.mean(weights), np.std(weights))
plt.step(weights, np.linspace(0,1, len(weights)), label='post')
plt.plot(X,Y)
plt.show()

```
```

Among the data they collected are the weights in kilograms of 398,484
respondents. Figure 4.8 shows the distribution of log *w*, where *w* is weight
in kilograms, along with a normal model.

The normal model is a good fit for the data, although the highest weights
exceed what we expect from the normal model even after the log transform.
Since the distribution of log *w* fits a normal distribution, we conclude that
*w* fits a lognormal distribution.

**Exercise 11** Download the BRFSS data from http://thinkstats.com/CDBRFS08.ASC.gz, and my code for reading it from http://thinkstats.com/brfss.py. Run `brfss.py`

and confirm that it prints summary statistics for a few of the variables.

Write a program that reads adult weights from the BRFSS and generates normal probability plots for $w$ and $log(w)$. You can download a solution from http://thinkstats.com/brfss_figs.py.

```
In [57]:
```#create plot... (this is for log weight)
Y = np.random.random(size=len(weights))
Y = sorted(Y)
plt.scatter(weights, Y, s=.2)
plt.show()

```
```

```
In [58]:
```# for non-transformed weight
weights = sorted([x for x in aw_data['weight2'] if x != 'NA' and not np.isnan(x)])
Y = np.random.random(size=len(weights))
Y = sorted(Y)
plt.scatter(weights, Y, s=.2)
plt.show()

```
```

**Exercise 12** The distribution of populations for cities and towns has been proposed as an example of a real-world phenomenon that can be described with a Pareto distribution.

The U.S. Census Bureau publishes data on the population of every incorporated city and town in the United States. I have written a small program that downloads this data and stores it in a file. You can download it from http://thinkstats.com/populations.py.

*Read over the program to make sure you know what it does; then run it to download and process the data.**Write a program that computes and plots the distribution of populations for the 14,593 cities and towns in the dataset.**Plot the CDF on linear and $log(x)$ scales so you can get a sense of the shape of the distribution. Then plot the CCDF on a log-log scale to see if it has the characteristic shape of a Pareto distribution.**Try out the other transformations and plots in this chapter to see if there is a better model for this data.*

What conclusion do you draw about the distribution of sizes for cities and towns? You can download a solution from http://thinkstats.com/populations_cdf.py.

```
In [105]:
```population = open('./data/populations.csv').readlines()
#ignore empty lines
title_ = "City"+population[3].strip()
population = [x.strip() for x in population[4:-6]]

```
In [106]:
```pattern = r',(?=")|(?<=\w),(?=\w)|(?<="),(?=\w)'
title = re.compile(pattern).split(title_)

```
In [112]:
```if re.compile(r'city|town', re.IGNORECASE).search('city'):
print 'yes'

```
```

```
In [114]:
```import re
# remove all commas which are between two numbers
def clean_string(s):
pattern = r'(?<=\d),(?=\d\d\d")'
s = re.sub(pattern, '', s)
s = re.sub(r'"', '', s)
return s.split(',')
#population = pd.DataFrame([clean_string(x) for x in population], columns = title)
pop = [clean_string(x) for x in population]
# remove any row in pop which doesn't have exactly 11 elements.
pop = [x for x in pop if len(x) == 11 and re.compile(r'city|town', re.IGNORECASE).search(x[0])]
popdf = pd.DataFrame(pop, columns = title)

```
In [127]:
```# there are entries which are not numbers, like "(X)"
def to_int(x):
try:
return np.float32(x)
except:
return np.nan
pop_census = []
for x in popdf['Census']:
try:
pop_census.append(float(x))
except:
pass
pop_census = sorted(pop_census)
plt.step(pop_census, np.linspace(0,1, len(pop_census)), label='post')
plt.show()

```
```

```
In [126]:
```# log transform
plt.step(pop_census, np.linspace(0,1,len(pop_census)), label='post')
plt.xscale('log')
plt.show()

```
```

```
In [130]:
```# plot the CCDF (log-log)
plt.step(pop_census[::-1], np.linspace(0,1,len(pop_census)), label='post')
plt.xscale('log')
plt.yscale('log')
plt.show()

```
```

**Exercise 13**

The Internal Revenue Service of the United States (IRS) provides data about income taxes at http://irs.gov/taxstats.

One of their files, containing information about individual incomes for 2008,
is available from http://thinkstats.com/08in11si.csv. I converted it to a
text format called CSV, which stands for “comma-separated values;” you can
read it using the `csv`

module.

Extract the distribution of incomes from this dataset. Are any of the
continuous distributions in this chapter a good model of the data? You can
download a solution from `http://thinkstats.com/irs.py`

.

At the beginning of this chapter I said that many real world phenomena can be modeled with continuous distributions. “So,” you might ask, “what?”

Like all models, continuous distributions are abstractions, which means they leave out details that are considered irrelevant. For example, an observed distribution might have measurement errors or quirks that are specific to the sample; continuous models smooth out these idiosyncrasies.

Continuous models are also a form of data compression. When a model fits a dataset well, a small set of parameters can summarize a large amount of data.

It is sometimes surprising when data from a natural phenomenon fit a
continuous distribution, but these observations can lead to insight into
physical systems. Sometimes we can explain why an observed distribution has a
particular form. For example, Pareto distributions are often the result of
generative processes with positive feedback (so-called preferential attachment
processes: see `http://wikipedia.org/wiki/Preferential_attachment`

.).

Continuous distributions lend themselves to mathematical analysis, as we will see in Chapter 6.

Continuous CDFs are also useful for generating random numbers. If there is an
efficient way to compute the inverse CDF, ICDF(*p*), we can generate random
values with the appropriate distribution by choosing from a uniform
distribution from 0 to 1, then choosing

For example, the CDF of the exponential distribution is

$$ p = 1- e^{-\lambda x} $$Solving for *x* yields:

So in Python we can write

```
def expovariate(lam):
p = random.random()
x = -math.log(1-p) / lam
return x
```

I called the parameter `lam`

because `lambda`

is a Python keyword. Most
implementations of `random.random`

can return 0 but not 1, so $1-p$ can be 1
but not 0, which is good, because $log(0)$ is undefined.

**Exercise 14** Write a function named `weibullvariate`

that takes `lam`

and `k`

and returns a random value from the Weibull distribution with those parameters.

```
In [26]:
```import math
def weibullvariate(lam, k):
"""the CDF for weibull is 1-e^{(-x/lambda)^k},
solving for p and rearrange..."""
p = random.random()
return ((-lam**k)*math.log(1-p))**(1/float(k))

```
In [27]:
```pd.DataFrame(sorted([weibullvariate(2,1) for _ in range(100)])).plot()

```
Out[27]:
```

**empirical distribution:** The distribution of values in a sample.

**continuous distribution:** A distribution described by a continuous function.

**interarrival time:** The elapsed time between two events.

**error function:** A special mathematical function, so-named because it comes up in the study of measurement errors.

**normal probability plot:** A plot of the sorted values in a sample versus the expected value for each if their distribution is normal.

**rankit:** The expected value of an element in a sorted list of values from a normal distribution.

**model:** A useful simplification. Continuous distributions are often good models of more complex empirical distributions.

**corpus:** A body of text used as a sample of a language.

**hapaxlegomenon:** A word that appears only once in a corpus. It appears twice in this book, so far.