Exploratory analysis of the US Airport Dataset

This dataset contains data for 25 years[1995-2015] of flights between various US airports and metadata about these routes. Taken from Bureau of Transportation Statistics, United States Department of Transportation.

Let's see what can we make out of this!


In [ ]:
%matplotlib inline
import networkx as nx
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
import warnings
warnings.filterwarnings('ignore')

pass_air_data = pd.read_csv('datasets/passengers.csv')

In the pass_air_data dataframe we have the information of number of people that fly every year on a particular route.


In [ ]:
pass_air_data.head()

In [ ]:
# Create a MultiDiGraph from this dataset

passenger_graph = nx.from_pandas_edgelist(pass_air_data, source='ORIGIN', target='DEST', edge_attr=['YEAR', 'PASSENGERS', 'UNIQUE_CARRIER_NAME'], create_using=nx.MultiDiGraph())

Cleveland to Chicago, how many people fly this route?


In [ ]:
passenger_graph['CLE']['ORD']

In [ ]:
temp = [(i['YEAR'], i['PASSENGERS'])for i in dict(passenger_graph['CLE']['ORD']).values()]
x, y = zip(*temp)
plt.plot(x, y)
plt.show()

Exercise

Find the busiest route in 1990 and in 2015 according to number of passengers, and plot the time series of number of passengers on these routes.

You can use the DataFrame instead of working with the network. It will be faster ;) [5 mins]


In [ ]:


In [ ]:


In [ ]:


In [ ]:

So let's have a look at the important nodes in this network, i.e. important airports in this network. We'll use pagerank, betweenness centrality and degree centrality.


In [ ]:
nx.pagerank(passenger_graph)

In [ ]:
def year_network(G, year):
    temp_g = nx.DiGraph()
    for i in G.edges(data=True):
        if i[2]['YEAR'] == year:
            temp_g.add_edge(i[0], i[1], weight=i[2]['PASSENGERS'])
    return temp_g

In [ ]:
pass_2015 = year_network(passenger_graph, 2015)

In [ ]:
len(pass_2015)

In [ ]:
len(pass_2015.edges())

In [ ]:
# Load in the GPS coordinates of all the airports
lat_long = pd.read_csv('datasets/GlobalAirportDatabase.txt', delimiter=':', header=None)

In [ ]:
lat_long[lat_long[1].isin(list(pass_2015.nodes()))]

In [ ]:
pos_dict = {}
for airport in lat_long[lat_long[1].isin(list(pass_2015.nodes()))].iterrows():
    pos_dict[airport[1][1]] = (airport[1][15], airport[1][14])

In [ ]:
pos_dict

Exercise

Using the position dictionary pos_dict create a plot of the airports, only the nodes not the edges.

  • As we don't have coordinates for all the airports we have to create a subgraph first.
  • Use nx.subgraph(Graph, iterable of nodes) to create the subgraph
  • Use nx.draw_networkx_nodes(G, pos) to map the nodes.

or

  • Just use a scatter plot :)

In [ ]:


In [ ]:

What about degree distribution of this network?


In [ ]:
plt.hist(list(nx.degree_centrality(pass_2015).values()))
plt.show()

Let's plot a log log plot to get a better overview of this.


In [ ]:
d = {}
for i, j in dict(nx.degree(pass_2015)).items():
    if j in d:
        d[j] += 1
    else:
        d[j] = 1
x = np.log2(list((d.keys())))
y = np.log2(list(d.values()))
plt.scatter(x, y, alpha=0.4)
plt.show()

Directed Graphs


In [ ]:
G = nx.DiGraph()

G.add_edge(1, 2, weight=1)

# print(G.edges())
# G[1][2]
# G[2][1]
# G.is_directed()
# type(G)

In [ ]:
G.add_edges_from([(1, 2), (3, 2), (4, 2), (5, 2), (6, 2), (7, 2)])
nx.draw_circular(G, with_labels=True)

In [ ]:
G.in_degree()

In [ ]:
nx.pagerank(G)

In [ ]:
G.add_edge(5, 6)
nx.draw_circular(G, with_labels=True)

In [ ]:
nx.pagerank(G)

In [ ]:
G.add_edge(2, 8)
nx.draw_circular(G, with_labels=True)

In [ ]:
nx.pagerank(G)

Moving back to Airports


In [ ]:
sorted(nx.pagerank(pass_2015, weight=None).items(), key=lambda x:x[1], reverse=True)[:10]

In [ ]:
sorted(nx.betweenness_centrality(pass_2015).items(), key=lambda x:x[1], reverse=True)[0:10]

In [ ]:
sorted(nx.degree_centrality(pass_2015).items(), key=lambda x:x[1], reverse=True)[0:10]

'ANC' is the airport code of Anchorage airport, a place in Alaska, and according to pagerank and betweenness centrality it is the most important airport in this network Isn't that weird? Thoughts?

related blog post: https://toreopsahl.com/2011/08/12/why-anchorage-is-not-that-important-binary-ties-and-sample-selection/

Let's look at weighted version, i.e taking into account the number of people flying to these places.


In [ ]:
sorted(nx.betweenness_centrality(pass_2015, weight='weight').items(), key=lambda x:x[1], reverse=True)[0:10]

In [ ]:
sorted(nx.pagerank(pass_2015, weight='weight').items(), key=lambda x:x[1], reverse=True)[0:10]

How reachable is this network?

We calculate the average shortest path length of this network, it gives us an idea about the number of jumps we need to make around the network to go from one airport to any other airport in this network.


In [ ]:
nx.average_shortest_path_length(pass_2015)

Wait, What??? This network is not connected. That seems like a really stupid thing to do.


In [ ]:
list(nx.weakly_connected_components(pass_2015))

SPB, SSB, AIK anyone?


In [ ]:
pass_air_data[(pass_air_data['YEAR'] == 2015) & (pass_air_data['ORIGIN'] == 'AIK')]

In [ ]:
pass_2015.remove_nodes_from(['SPB', 'SSB', 'AIK'])

In [ ]:
nx.is_weakly_connected(pass_2015)

In [ ]:
nx.is_strongly_connected(pass_2015)

Strongly vs weakly connected graphs.


In [ ]:
G = nx.DiGraph()
G.add_edge(1, 2)
G.add_edge(2, 3)
G.add_edge(3, 1)
nx.draw(G)

In [ ]:
G.add_edge(3, 4)
nx.draw(G)

In [ ]:
nx.is_strongly_connected(G)

In [ ]:
list(nx.strongly_connected_components(pass_2015))

In [ ]:
pass_air_data[(pass_air_data['YEAR'] == 2015) & (pass_air_data['DEST'] == 'TSP')]

In [ ]:
pass_2015_strong = max(nx.strongly_connected_component_subgraphs(pass_2015), key=len)

In [ ]:
len(pass_2015_strong)

In [ ]:
nx.average_shortest_path_length(pass_2015_strong)

Exercise! (Actually this is a game :D)

How can we decrease the avg shortest path length of this network?

Think of an effective way to add new edges to decrease the avg shortest path length. Let's see if we can come up with a nice way to do this, and the one who gets the highest decrease wins!!!

The rules are simple:

  • You can't add more than 2% of the current edges( ~500 edges)

[10 mins]


In [ ]:


In [ ]:

What about airlines? Can we find airline specific reachability?


In [ ]:
passenger_graph['CLE']['SFO'][25]

In [ ]:
def str_to_list(a):
    return a[1:-1].split(', ')

In [ ]:
for i in str_to_list(passenger_graph['JFK']['SFO'][25]['UNIQUE_CARRIER_NAME']):
    print(i)

In [ ]:
%%time
for origin, dest in passenger_graph.edges():
    for key in passenger_graph[origin][dest]:
        passenger_graph[origin][dest][key]['airlines'] = str_to_list(passenger_graph[origin][dest][key]['UNIQUE_CARRIER_NAME'])

Exercise

Play around with United Airlines network.

  • Extract a network for United Airlines flights from the metagraph passenger_graph for the year 2015
  • Make sure it's a weighted network, where weight is the number of passengers.
  • Find the number of airports and connections in this network
  • Find the most important airport, according to PageRank and degree centrality.

In [ ]:
united_network = nx._________
for _______, _______ in passenger_graph.edges():
    if 25 in passenger_graph[______][_______]: # 25 key is for the year 2015
        if "'United Air Lines Inc.'" in ____________________:
            united_network.add_edge(_____, ______, weight= __________)

In [ ]:
# number of nodes

In [ ]:
# number of edges

In [ ]:
# top 10 according to pagerank

In [ ]:
# top 10 according to degree centrality

Exercise

We are in Cleveland so what should we do?

Obviously we will make a time series of number of passengers flying out of Cleveland with United Airlines over the years.

There are 2 ways of doing it.

  • Create a new multidigraph specifically for this exercise.

OR

  • exploit the pass_air_data dataframe.

In [ ]: