Monte Carlo Calculation of π

Christina C. Lee

Category: Numerics

Monte Carlo Physics Series

Monte Carlo- Random Numbers to Improve Calculations

When one hears "Monte Carlo", most people might think of something like this:

Monte Carlo, Monaco: known for extremely large amounts of money, car racing, no income taxes,and copious gambling.

In addition to Monaco, Europe, Las Vegas decided to host a Monte Carlo-themed casino as well. So during the Manhattan project, when the best minds in the United States were camped out in the New Mexican desert, they had plenty of inspiration from Las Vegas, and plenty of difficult problems to work on in the form of quantifying the inner workings of nuclei. Enrico Fermi first played with these ideas, but Stanislaw Ulam invented the modern Monte Carlo Markov Chain later.

At the same time, these scientists now had computers at their disposal. John von Neumann programmed Ulam's algorithm onto ENIAC (Electronic Numerical Integrator and Computer), the very first electronic, general purpose computer, even though it did still run on vacuum tubes.

That still doesn't answer, why do random numbers actually help us solve problems?

Imagine you are visiting a new city for the first time (maybe Monte Carlo). You only have a day or two, and you want to really get to know the city. You have two options for your visit

  • Hit the tourist sites you researched online
  • Wander around. Try and communicate with the locals. Find an out-of-the-way restaurant and sample food not tamed for foreigners. Watch people interact. Get lost.

Both are legitimate ways to see the city. But depending on what you want, you might choose a different option. The same goes for exploring physics problems. Sometimes you want to go in and study just everything you knew about beforehand, but sometimes you need to wander around, not hit everything, but get a better feeling for what everything might be like.

Buffon's Needle: Calculation of π

Even back in the 18th century, Georges-Louis Leclerc, Comte de Buffon posed a problem in geometric probability. Nowadays, we use a slightly different version of that problem to calculate π and illustrate Monte Carlo simulations.

Suppose we have a square dartboard, and someone with really bad, completely random aim, even though he/she always at least hits inside the dartboard. We then inscribe a circle inside that dartboard. After an infinite number of hits, what is the ratio between hits in the circle, and hits in the square?

\begin{equation} f= \frac{N_{circle}}{N_{square}} =\frac{\text{Area of circle}}{\text{Area of square}} =\frac{\pi r^2}{4 r^2}= \frac{\pi}{4} \end{equation}\begin{equation} \pi = 4 f \end{equation}

Onto the Code!


In [1]:
using Statistics
using Plots
gr()


Out[1]:
Plots.GRBackend()

We will generate our random numbers on the unit interval. Thus the radius in our circumstance is $.5$.

Write a function incircle(r2) such that if r2 is in the circle, it returns true, else, it returns false. We will use this with the Julia function filter. Assume r2 is the radius squared, and already centered around the middle of the unit circle


In [2]:
function incircle(r2)
    if r2<.25 
        return true
    else
        return false
    end
end


Out[2]:
incircle (generic function with 1 method)

In [3]:
#The number of darts we will throw at the board.  We will see how accurate different numbers are
N=[10,25,50,75,100,250,500,750,1000,2500,5000,7500,10000];
# We will perform each number multiple times in order to calculate error bars
M=15;

In [4]:
πapprox=zeros(Float64,length(N),M);

for ii in 1:length(N) 
    for jj in 1:M
        
        #popular our array with random numbers on the unit interval
        X=rand(N[ii],2)
        
        #calculate their radius squared
        R2=(X[:,1].-0.5).^2.0.+(X[:,2].-0.5).^2
        
        # 4*number in circle / total number
        πapprox[ii,jj]=4.0*length(filter(incircle,R2))/N[ii];
        
    end
end

In [5]:
# Get our averages and standard deviations
πave=mean(πapprox,dims=2);
πstd=std(πapprox,dims=2);

Analysis

So that was a nice, short little piece of code. Lets plot it now to see means.


In [6]:
plot(N,π*ones(length(N)),xscale=:log10);

for j in 1:M
    scatter!(N,πapprox[:,j]);
end

scatter!(N,πave,yerr=πstd)
plot!(xlabel="Number of Darts",ylabel="pi Estimate",
        title="Monte Carlo Estimate of pi",legend=:false)


Out[6]:
10 1 10 2 10 3 10 4 2.5 3.0 3.5 4.0 Monte Carlo Estimate of pi Number of Darts pi Estimate

When we have fewer numbers of points, our estimates vary much more wildly, and much further from 3.1415926 . But, at least, the guesses from our different runs all seem equally distributed around the correct value, so it seems we have no systematic error.

As we get up to $10^4$, our estimate starts getting much more accurate and consistent.


In [7]:
plot(N,πstd,xscale=:log10)

plot!(xlabel="N points"
    ,ylabel="standard deviation"
    ,title="Dependence of Monte Carlo Error on Number of Points")


Out[7]:
10 1 10 2 10 3 10 4 0.0 0.1 0.2 0.3 0.4 0.5 Dependence of Monte Carlo Error on Number of Points N points standard deviation y1

So what we guessed in the first plot about dispersion in estimate, we quantify here in this plot. When we only have 10 darts, the guesses vary by up to .3, but when we get down to 1,000 trials, we are starting to be consistent to .0002


In [8]:
plot(N,πave,xscale=:log10,label="π Guess")
plot!(N,π*ones(length(N)),label="π")

plot!(xlabel="N steps"
    ,ylabel="Average of 15 runs"
    ,title="Overall Averages")


Out[8]:
10 1 10 2 10 3 10 4 3.050 3.075 3.100 3.125 3.150 Overall Averages N steps Average of 15 runs ? Guess ?

Now lets just make a graphical representation of what we've been doing this whole time. Plot our points on a unit square, and color the ones inside the circle a different color.


In [9]:
X=rand(1000);
Y=rand(1000);
R2=(X.-0.5).^2.0.+(Y.-0.5).^2;
Xc=[];
Yc=[]
for ii in 1:length(X)
    if R2[ii]<.25
        push!(Xc,X[ii]);
        push!(Yc,Y[ii]);
    end
end

In [10]:
scatter(X,Y)
scatter!(Xc,Yc)
plot!(aspect_ratio=1,xlabel="X",ylabel="Y",legend=:false,
        title="Dartboard")


Out[10]:
-0.2 0.0 0.2 0.4 0.6 0.8 1.0 1.2 0.00 0.25 0.50 0.75 1.00 Dartboard X Y

That's all folks! Now here's a picture of some pie to congratulate you on calculating π.

By Scott Bauer, USDA ARS - This image was released by the Agricultural Research Service, the research agency of the United States Department of Agriculture, with the ID K7252-47 (next).This tag does not indicate the copyright status of the attached work. A normal copyright tag is still required. See Commons:Licensing for more information.English | français | македонски | +/−, Public Domain, https://commons.wikimedia.org/w/index.php?curid=264106

@article{Markov, author = {A.  Markov , A}, year = {2006}, month = {12}, pages = {591 - 600}, title = {An Example of Statistical Investigation of the Text Eugene Onegin Concerning the Connection of Samples in Chains}, volume = {19}, journal = {Science in Context}, doi = {10.1017/S0269889706001074} }


In [ ]: