In [ ]:
%load_ext autoreload
%autoreload 2
%matplotlib inline
%config InlineBackend.figure_format = 'retina'
In this chapter, we will look at the relationship between graphs and linear algebra.
The deep connection between these two topics is super interesting, and I'd like to show it to you through an exploration of three topics:
Before we go deep into the linear algebra piece though, we have to first make sure some ideas are clear.
The most important thing that we need when treating graphs in linear algebra form is the adjacency matrix. For example, for four nodes joined in a chain:
In [ ]:
import networkx as nx
nodes = list(range(4))
G1 = nx.Graph()
G1.add_nodes_from(nodes)
G1.add_edges_from(zip(nodes, nodes[1:]))
we can visualize the graph:
In [ ]:
nx.draw(G1, with_labels=True)
and we can visualize its adjacency matrix:
In [ ]:
import nxviz as nv
m = nv.MatrixPlot(G1)
m.draw()
and we can obtain the adjacency matrix as a NumPy array:
In [ ]:
A1 = nx.to_numpy_array(G1, nodelist=sorted(G1.nodes()))
A1
In the Paths chapter, we can use the breadth-first search algorithm to find a shortest path between any two nodes.
As it turns out, using adjacency matrices, we can answer a related question, which is how many paths exist of length K between two nodes.
To see how, we need to see the relationship between matrix powers and graph path lengths.
Let's take the adjacency matrix above, raise it to the second power, and see what it tells us.
In [ ]:
import numpy as np
np.linalg.matrix_power(A1, 2)
In [ ]:
from nams.solutions.linalg import adjacency_matrix_power
from nams.functions import render_html
render_html(adjacency_matrix_power())
In [ ]:
np.linalg.matrix_power(A1, 3)
You should be able to convince yourself that:
In [ ]:
G2 = nx.DiGraph()
G2.add_nodes_from(nodes)
G2.add_edges_from(zip(nodes, nodes[1:]))
nx.draw(G2, with_labels=True)
Convince yourself that the resulting adjacency matrix power contains the same semantic meaning as that for an undirected graph, that is, the number of ways to go from "row" node to "column" node in K steps. (I have provided three different matrix powers for you.)
In [ ]:
A2 = nx.to_numpy_array(G2)
np.linalg.matrix_power(A2, 2)
In [ ]:
np.linalg.matrix_power(A2, 3)
In [ ]:
np.linalg.matrix_power(A2, 4)
Let's now dive into the second topic here, that of message passing.
To show how message passing works on a graph, let's start with the directed linear chain, as this will make things easier to understand.
Our graph adjacency matrix contains nodes ordered in a particular fashion along the rows and columns. We can also create a "message" matrix $M$, using the same ordering of nodes along the rows, with columns instead representing a "message" that is intended to be "passed" from one node to another:
In [ ]:
M = np.array([1, 0, 0, 0])
M
Notice where the position of the value 1
is - at the first node.
If we take M and matrix multiply it against A2, let's see what we get:
In [ ]:
M @ A2
The message has been passed onto the next node! And if we pass the message one more time:
In [ ]:
M @ A2 @ A2
Now, the message lies on the 3rd node!
We can make an animation to visualize this more clearly. There are comments in the code to explain what's going on!
In [ ]:
def propagate(G, msg, n_frames):
"""
Computes the node values based on propagation.
Intended to be used before or when being passed into the
anim() function (defined below).
:param G: A NetworkX Graph.
:param msg: The initial state of the message.
:returns: A list of 1/0 representing message status at
each node.
"""
# Initialize a list to store message states at each timestep.
msg_states = []
# Set a variable `new_msg` to be the initial message state.
new_msg = msg
# Get the adjacency matrix of the graph G.
A = nx.to_numpy_array(G)
# Perform message passing at each time step
for i in range(n_frames):
msg_states.append(new_msg)
new_msg = new_msg @ A
# Return the message states.
return msg_states
In [ ]:
from IPython.display import HTML
import matplotlib.pyplot as plt
from matplotlib import animation
def update_func(step, nodes, colors):
"""
The update function for each animation time step.
:param step: Passed in from matplotlib's FuncAnimation. Must
be present in the function signature.
:param nodes: Returned from nx.draw_networkx_edges(). Is an
array of colors.
:param colors: A list of pre-computed colors.
"""
nodes.set_array(colors[step].ravel())
return nodes
def anim(G, initial_state, n_frames=4):
"""
Animation function!
"""
# First, pre-compute the message passing states over all frames.
colors = propagate(G, initial_state, n_frames)
# Instantiate a figure
fig = plt.figure()
# Precompute node positions so that they stay fixed over the entire animation
pos = nx.kamada_kawai_layout(G)
# Draw nodes to screen
nodes = nx.draw_networkx_nodes(G, pos=pos, node_color=colors[0].ravel(), node_size=20)
# Draw edges to screen
ax = nx.draw_networkx_edges(G, pos)
# Finally, return the animation through matplotlib.
return animation.FuncAnimation(fig, update_func, frames=range(n_frames), fargs=(nodes, colors))
# Initialize the message
msg = np.zeros(len(G2))
msg[0] = 1
# Animate the graph with message propagation.
HTML(anim(G2, msg, n_frames=4).to_html5_video())
The section on message passing above assumed unipartite graphs, or at least graphs for which messages can be meaningfully passed between nodes.
In this section, we will look at bipartite graphs.
Recall from before the definition of a bipartite graph:
Bipartite graphs have a natural matrix representation, known as the biadjacency matrix. Nodes on one partition are the rows, and nodes on the other partition are the columns.
NetworkX's bipartite
module provides a function for computing the biadjacency matrix of a bipartite graph.
Let's start by looking at a toy bipartite graph, a "customer-product" purchase record graph, with 4 products and 3 customers. The matrix representation might be as follows:
In [ ]:
# Rows = customers, columns = products, 1 = customer purchased product, 0 = customer did not purchase product.
cp_mat = np.array([[0, 1, 0, 0],
[1, 0, 1, 0],
[1, 1, 1, 1]])
From this "bi-adjacency" matrix, one can compute the projection onto the customers, matrix multiplying the matrix with its transpose.
In [ ]:
c_mat = cp_mat @ cp_mat.T # c_mat means "customer matrix"
c_mat
What we get is the connectivity matrix of the customers, based on shared purchases. The diagonals are the degree of the customers in the original graph, i.e. the number of purchases they originally made, and the off-diagonals are the connectivity matrix, based on shared products.
To get the products matrix, we make the transposed matrix the left side of the matrix multiplication.
In [ ]:
p_mat = cp_mat.T @ cp_mat # p_mat means "product matrix"
p_mat
You may now try to convince yourself that the diagonals are the number of times a customer purchased that product, and the off-diagonals are the connectivity matrix of the products, weighted by how similar two customers are.
In the following exercises, you will now play with a customer-product graph from Amazon. This dataset was downloaded from UCSD's Julian McAuley's website, and corresponds to the digital music dataset.
This is a bipartite graph. The two partitions are:
customers
: The customers that were doing the reviews.products
: The music that was being reviewed.In the original dataset (see the original JSON in the datasets/
directory), they are referred to as:
customers
: reviewerID
products
: asin
In [ ]:
from nams import load_data as cf
G_amzn = cf.load_amazon_reviews()
Remember that with bipartite graphs, it is useful to obtain nodes from one of the partitions.
In [ ]:
from nams.solutions.bipartite import extract_partition_nodes
In [ ]:
customer_nodes = extract_partition_nodes(G_amzn, "customer")
mat = nx.bipartite.biadjacency_matrix(G_amzn, row_order=customer_nodes)
You'll notice that this matrix is extremely large! There are 5541 customers and 3568 products, for a total matrix size of $5541 \times 3568 = 19770288$, but it is stored in a sparse format because only 64706 elements are filled in.
In [ ]:
mat
In [ ]:
customer_mat = mat @ mat.T
Next, get the diagonals of the customer-customer matrix. Recall here that in customer_mat
, the diagonals correspond to the degree of the customer nodes in the bipartite matrix.
SciPy sparse matrices provide a .diagonal()
method that returns the diagonal elements.
In [ ]:
# Get the diagonal.
degrees = customer_mat.diagonal()
Finally, find the index of the customer that has the highest degree.
In [ ]:
cust_idx = np.argmax(degrees)
cust_idx
We can verify this independently by sorting the customer nodes by degree.
In [ ]:
import pandas as pd
import janitor
# There's some pandas-fu we need to use to get this correct.
deg = (
pd.Series(dict(nx.degree(G_amzn, customer_nodes)))
.to_frame()
.reset_index()
.rename_column("index", "customer")
.rename_column(0, "num_reviews")
.sort_values('num_reviews', ascending=False)
)
deg.head()
Indeed, customer 294 was the one who had the most number of reviews!
Let's now also compute which two customers are similar, based on shared reviews. To do so involves the following steps:
scipy.sparse.diags(elements)
will construct a sparse diagonal matrix based on the elements inside elements
.
In [ ]:
import scipy.sparse as sp
In [ ]:
# Construct diagonal elements.
customer_diags = sp.diags(degrees)
# Subtract off-diagonals.
off_diagonals = customer_mat - customer_diags
# Compute index of most similar individuals.
np.unravel_index(np.argmax(off_diagonals), customer_mat.shape)
In [ ]:
from time import time
start = time()
# Compute the projection
G_cust = nx.bipartite.weighted_projected_graph(G_amzn, customer_nodes)
# Identify the most similar customers
most_similar_customers = sorted(G_cust.edges(data=True), key=lambda x: x[2]['weight'], reverse=True)[0]
end = time()
print(f'{end - start:.3f} seconds')
print(f'Most similar customers: {most_similar_customers}')
In [ ]:
start = time()
# Compute the projection using matrices
mat = nx.bipartite.matrix.biadjacency_matrix(G_amzn, customer_nodes)
cust_mat = mat @ mat.T
# Identify the most similar customers
degrees = customer_mat.diagonal()
customer_diags = sp.diags(degrees)
off_diagonals = customer_mat - customer_diags
c1, c2 = np.unravel_index(np.argmax(off_diagonals), customer_mat.shape)
end = time()
print(f'{end - start:.3f} seconds')
print(f'Most similar customers: {customer_nodes[c1]}, {customer_nodes[c2]}, {cust_mat[c1, c2]}')
On a modern PC, the matrix computation should be about 10-50X faster using the matrix form compared to the object-oriented form. (The web server that is used to build the book might not necessarily have the software stack to do this though, so the time you see reported might not reflect the expected speedups.) I'd encourage you to fire up a Binder session or clone the book locally to test out the code yourself.
You may notice that it's much easier to read the "objects" code, but the matrix code way outperforms the object code. This tradeoff is common in computing, and shouldn't surprise you. That said, the speed gain alone is a great reason to use matrices!