Getting started with TensorFlow

In this notebook, you play around with the TensorFlow Python API.


In [ ]:
# Ensure the right version of Tensorflow is installed.
!pip freeze | grep tensorflow==2.1 || pip install tensorflow==2.1

While re-running the above cell you will see the output tensorflow==2.1.0 that is the installed version of tensorflow.


In [1]:
import tensorflow as tf
import numpy as np

print(tf.__version__)


2.1.0

Adding two tensors

First, let's try doing this using numpy, the Python numeric package. numpy code is immediately evaluated.


In [2]:
a = np.array([5, 3, 8])
b = np.array([3, -1, 2])
c = np.add(a, b)
print(c)


[ 8  2 10]

The equivalent code in TensorFlow consists of two steps:

Step 1: Build the graph


In [3]:
tf.compat.v1.disable_eager_execution() # Need to disable eager in TF2.x
a = tf.compat.v1.constant([5, 3, 8])
b = tf.compat.v1.constant([3, -1, 2])
c = tf.add(a,b,name='c')
print(c)


Tensor("Add:0", shape=(3,), dtype=int32)

c is an Op ("Add") that returns a tensor of shape (3,) and holds int32. The shape is inferred from the computation graph.

Try the following in the cell above:

  1. Change the 5 to 5.0, and similarly the other five numbers. What happens when you run this cell?
  2. Add an extra number to a, but leave b at the original (3,) shape. What happens when you run this cell?
  3. Change the code back to a version that works

Step 2: Launch the graph in a session.


In [4]:
sess = tf.compat.v1.Session()
# Evaluate the tensor c.
print(sess.run(c))


[ 8  2 10]

Using a feed_dict

Same graph, but without hardcoding inputs at build stage


In [5]:
a = tf.compat.v1.placeholder(tf.int32, name='a')
b = tf.compat.v1.placeholder(tf.int32, name='b')
c = tf.add(a,b,name='c')
sess = tf.compat.v1.Session()
print(sess.run(c, feed_dict={
      a: [3, 4, 5],
      b: [-1, 2, 3]
   }))


[2 6 8]

Heron's Formula in TensorFlow

The area of triangle whose three sides are $(a, b, c)$ is $\sqrt{s(s-a)(s-b)(s-c)}$ where $s=\frac{a+b+c}{2}$

Look up the available operations at https://www.tensorflow.org/api_docs/python/tf


In [6]:
def compute_area(sides):
  # slice the input to get the sides
  a = sides[:,0]  # 5.0, 2.3
  b = sides[:,1]  # 3.0, 4.1
  c = sides[:,2]  # 7.1, 4.8
  
  # Heron's formula
  s = (a + b + c) * 0.5   # (a + b) is a short-cut to tf.add(a, b)
  areasq = s * (s - a) * (s - b) * (s - c) # (a * b) is a short-cut to tf.multiply(a, b), not tf.matmul(a, b)
  return tf.sqrt(areasq)

sess = tf.compat.v1.Session()
  # pass in two triangles
area = compute_area(tf.constant([
      [5.0, 3.0, 7.1],
      [2.3, 4.1, 4.8]
    ]))

print(sess.run(area))


[6.278497 4.709139]

Placeholder and feed_dict

More common is to define the input to a program as a placeholder and then to feed in the inputs. The difference between the code below and the code above is whether the "area" graph is coded up with the input values or whether the "area" graph is coded up with a placeholder through which inputs will be passed in at run-time.


In [7]:
sess = tf.compat.v1.Session()
sides = tf.compat.v1.placeholder(tf.float32, shape=(None, 3))  # batchsize number of triangles, 3 sides
area = compute_area(sides)
print(sess.run(area, feed_dict = {
      sides: [
        [5.0, 3.0, 7.1],
        [2.3, 4.1, 4.8]
     ]
     }))


[6.278497 4.709139]

tf.eager

tf.eager allows you to avoid the build-then-run stages. However, most production code will follow the lazy evaluation paradigm because the lazy evaluation paradigm is what allows for multi-device support and distribution.

One thing you could do is to develop using tf.eager and then comment out the eager execution and add in the session management code.

You may need to restart the session to try this out.


In [1]:
import tensorflow as tf

tf.compat.v1.enable_eager_execution()

def compute_area(sides):
  # slice the input to get the sides
  a = sides[:,0]  # 5.0, 2.3
  b = sides[:,1]  # 3.0, 4.1
  c = sides[:,2]  # 7.1, 4.8
  
  # Heron's formula
  s = (a + b + c) * 0.5   # (a + b) is a short-cut to tf.add(a, b)
  areasq = s * (s - a) * (s - b) * (s - c) # (a * b) is a short-cut to tf.multiply(a, b), not tf.matmul(a, b)
  return tf.sqrt(areasq)

area = compute_area(tf.constant([
      [5.0, 3.0, 7.1],
      [2.3, 4.1, 4.8]
    ]))


print(area)


tf.Tensor([6.278497 4.709139], shape=(2,), dtype=float32)

Challenge Exercise

Use TensorFlow to find the roots of a fourth-degree polynomial using Halley's Method. The five coefficients (i.e. $a_0$ to $a_4$) of

$f(x) = a_0 + a_1 x + a_2 x^2 + a_3 x^3 + a_4 x^4$

will be fed into the program, as will the initial guess $x_0$. Your program will start from that initial guess and then iterate one step using the formula:

If you got the above easily, try iterating indefinitely until the change between $x_n$ and $x_{n+1}$ is less than some specified tolerance. Hint: Use tf.while_loop

Copyright 2020 Google Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License