Learning Objectives
The Keras sequential API allows you to create Tensorflow models layer-by-layer. This is useful for building most kinds of machine learning models but it does not allow you to create models that share layers, re-use layers or have multiple inputs or outputs.
In this lab, we'll see how to build a simple deep neural network model using the keras sequential api and feature columns. Once we have trained our model, we will deploy it using AI Platform and see how to call our model for online prediciton.
In [ ]:
!sudo chown -R jupyter:jupyter /home/jupyter/training-data-analyst
In [ ]:
# Ensure the right version of Tensorflow is installed.
!pip freeze | grep tensorflow==2.1 || pip install tensorflow==2.1
Start by importing the necessary libraries for this lab.
In [ ]:
import datetime
import os
import shutil
import numpy as np
import pandas as pd
import tensorflow as tf
from matplotlib import pyplot as plt
from tensorflow import keras
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, DenseFeatures
from tensorflow.keras.callbacks import TensorBoard
print(tf.__version__)
%matplotlib inline
In [ ]:
!ls -l ../data/*.csv
In [ ]:
!head ../data/taxi*.csv
We wrote these functions for reading data from the csv files above in the previous notebook.
In [ ]:
CSV_COLUMNS = [
'fare_amount',
'pickup_datetime',
'pickup_longitude',
'pickup_latitude',
'dropoff_longitude',
'dropoff_latitude',
'passenger_count',
'key'
]
LABEL_COLUMN = 'fare_amount'
DEFAULTS = [[0.0], ['na'], [0.0], [0.0], [0.0], [0.0], [0.0], ['na']]
UNWANTED_COLS = ['pickup_datetime', 'key']
def features_and_labels(row_data):
label = row_data.pop(LABEL_COLUMN)
features = row_data
for unwanted_col in UNWANTED_COLS:
features.pop(unwanted_col)
return features, label
def create_dataset(pattern, batch_size=1, mode='eval'):
dataset = tf.data.experimental.make_csv_dataset(
pattern, batch_size, CSV_COLUMNS, DEFAULTS)
dataset = dataset.map(features_and_labels)
if mode == 'train':
dataset = dataset.shuffle(buffer_size=1000).repeat()
# take advantage of multi-threading; 1=AUTOTUNE
dataset = dataset.prefetch(1)
return dataset
We will use feature columns to connect our raw data to our keras DNN model. Feature columns make it easy to perform common types of feature engineering on your raw data. For example, you can one-hot encode categorical data, create feature crosses, embeddings and more. We'll cover these in more detail later in the course, but if you want to a sneak peak browse the official TensorFlow feature columns guide.
In our case we won't do any feature engineering. However, we still need to create a list of feature columns to specify the numeric values which will be passed on to our model. To do this, we use tf.feature_column.numeric_column()
We use a python dictionary comprehension to create the feature columns for our model, which is just an elegant alternative to a for loop.
Lab Task #1: Create a feature column dictionary that we will use when building our deep neural network below. The keys should be the element of the INPUT_COLS
list, while the values should be numeric feature columns.
In [ ]:
INPUT_COLS = [
'pickup_longitude',
'pickup_latitude',
'dropoff_longitude',
'dropoff_latitude',
'passenger_count',
]
# Create input layer of feature columns
# TODO 1
feature_columns = # TODO -- Your code here.
Next, we create the DNN model. The Sequential model is a linear stack of layers and when building a model using the Sequential API, you configure each layer of the model in turn. Once all the layers have been added, you compile the model.
Lab Task #2a: Create a deep neural network using Keras's Sequential API. In the cell below, use the tf.keras.layers
library to create all the layers for your deep neural network.
In [ ]:
# Build a keras DNN model using Sequential API
# TODO 2a
model = # TODO -- Your code here.
Next, to prepare the model for training, you must configure the learning process. This is done using the compile method. The compile method takes three arguments:
rmsprop
or adagrad
), or an instance of the Optimizer class.We will add an additional custom metric called rmse
to our list of metrics which will return the root mean square error.
Lab Task #2b: Compile the model you created above. Create a custom loss function called rmse
which computes the root mean squared error between y_true
and y_pred
. Pass this function to the model as an evaluation metric.
In [ ]:
# TODO 2b
# Create a custom evalution metric
def rmse(y_true, y_pred):
return # TODO -- Your code here.
# Compile the keras model
# TODO -- Your code here.
To train your model, Keras provides three functions that can be used:
.fit()
for training a model for a fixed number of epochs (iterations on a dataset)..fit_generator()
for training a model on data yielded batch-by-batch by a generator.train_on_batch()
runs a single gradient update on a single batch of data. The .fit()
function works well for small datasets which can fit entirely in memory. However, for large datasets (or if you need to manipulate the training data on the fly via data augmentation, etc) you will need to use .fit_generator()
instead. The .train_on_batch()
method is for more fine-grained control over training and accepts only a single batch of data.
The taxifare dataset we sampled is small enough to fit in memory, so can we could use .fit
to train our model. Our create_dataset
function above generates batches of training examples, so we could also use .fit_generator
. In fact, when calling .fit
the method inspects the data, and if it's a generator (as our dataset is) it will invoke automatically .fit_generator
for training.
We start by setting up some parameters for our training job and create the data generators for the training and validation data.
We refer you the the blog post ML Design Pattern #3: Virtual Epochs for further details on why express the training in terms of NUM_TRAIN_EXAMPLES
and NUM_EVALS
and why, in this training code, the number of epochs is really equal to the number of evaluations we perform.
In [ ]:
TRAIN_BATCH_SIZE = 1000
NUM_TRAIN_EXAMPLES = 10000 * 5 # training dataset will repeat, wrap around
NUM_EVALS = 50 # how many times to evaluate
NUM_EVAL_EXAMPLES = 10000 # enough to get a reasonable sample
trainds = create_dataset(
pattern='../data/taxi-train*',
batch_size=TRAIN_BATCH_SIZE,
mode='train')
evalds = create_dataset(
pattern='../data/taxi-valid*',
batch_size=1000,
mode='eval').take(NUM_EVAL_EXAMPLES//1000)
There are various arguments you can set when calling the .fit method. Here x
specifies the input data which in our case is a tf.data
dataset returning a tuple of (inputs, targets). The steps_per_epoch
parameter is used to mark the end of training for a single epoch. Here we are training for NUM_EVALS epochs. Lastly, for the callback
argument we specify a Tensorboard callback so we can inspect Tensorboard after training.
Lab Task #3: In the cell below, you will train your model. First, define the steps_per_epoch
then train your model using .fit()
, saving the model training output to a variable called history
.
In [ ]:
# TODO 3
%time
steps_per_epoch = # TODO -- Your code here.
LOGDIR = "./taxi_trained"
history = # TODO -- Your code here.
In [ ]:
model.summary()
Running .fit
(or .fit_generator
) returns a History object which collects all the events recorded during training. Similar to Tensorboard, we can plot the training and validation curves for the model loss and rmse by accessing these elements of the History object.
In [ ]:
RMSE_COLS = ['rmse', 'val_rmse']
pd.DataFrame(history.history)[RMSE_COLS].plot()
In [ ]:
LOSS_COLS = ['loss', 'val_loss']
pd.DataFrame(history.history)[LOSS_COLS].plot()
To make predictions with our trained model, we can call the predict method, passing to it a dictionary of values. The steps
parameter determines the total number of steps before declaring the prediction round finished. Here since we have just one example, we set steps=1
(setting steps=None
would also work). Note, however, that if x is a tf.data
dataset or a dataset iterator, and steps is set to None, predict will run until the input dataset is exhausted.
In [ ]:
model.predict(x={"pickup_longitude": tf.convert_to_tensor([-73.982683]),
"pickup_latitude": tf.convert_to_tensor([40.742104]),
"dropoff_longitude": tf.convert_to_tensor([-73.983766]),
"dropoff_latitude": tf.convert_to_tensor([40.755174]),
"passenger_count": tf.convert_to_tensor([3.0])},
steps=1)
Of course, making individual predictions is not realistic, because we can't expect client code to have a model object in memory. For others to use our trained model, we'll have to export our model to a file, and expect client code to instantiate the model from that exported file.
We'll export the model to a TensorFlow SavedModel format. Once we have a model in this format, we have lots of ways to "serve" the model, from a web application, from JavaScript, from mobile applications, etc.
Lab Task #4: Use tf.saved_model.save
to export the trained model to a Tensorflow SavedModel format. Reference the documentation for tf.saved_model.save
as you fill in the code for the cell below.
Next, print the signature of your saved model using the SavedModel Command Line Interface command saved_model_cli
. You can read more about the command line interface and the show
and run
commands it supports in the documentation here.
In [ ]:
# TODO 4a
OUTPUT_DIR = "./export/savedmodel"
shutil.rmtree(OUTPUT_DIR, ignore_errors=True)
EXPORT_PATH = os.path.join(OUTPUT_DIR,
datetime.datetime.now().strftime("%Y%m%d%H%M%S"))
tf.saved_model.save( # TODO -- Your code here.
In [ ]:
# TODO 4b
!saved_model_cli show \
--tag_set # TODO -- Your code here.
--signature_def # TODO -- Your code here.
--dir # TODO -- Your code here.
!find {EXPORT_PATH}
os.environ['EXPORT_PATH'] = EXPORT_PATH
Lab Task #5a: Complete the code in the cell below to deploy your trained model to AI Platform using the gcloud ai-platform versions create
command. Have a look at the documentation for how to create model version with gcloud.
In [ ]:
%%bash
# TODO 5a
PROJECT= #TODO: Change this to your PROJECT
BUCKET=${PROJECT}
REGION=us-east1
MODEL_NAME=taxifare
VERSION_NAME=dnn
## Create GCS bucket if it doesn't exist already...
exists=$(gsutil ls -d | grep -w gs://${BUCKET}/)
if [ -n "$exists" ]; then
echo -e "Bucket exists, let's not recreate it."
else
echo "Creating a new GCS bucket."
gsutil mb -l ${REGION} gs://${BUCKET}
echo "\nHere are your current buckets:"
gsutil ls
fi
if [[ $(gcloud ai-platform models list --format='value(name)' | grep $MODEL_NAME) ]]; then
echo "$MODEL_NAME already exists"
else
echo "Creating $MODEL_NAME"
gcloud ai-platform models create --regions=$REGION $MODEL_NAME
fi
if [[ $(gcloud ai-platform versions list --model $MODEL_NAME --format='value(name)' | grep $VERSION_NAME) ]]; then
echo "Deleting already existing $MODEL_NAME:$VERSION_NAME ... "
echo yes | gcloud ai-platform versions delete --model=$MODEL_NAME $VERSION_NAME
echo "Please run this cell again if you don't see a Creating message ... "
sleep 2
fi
echo "Creating $MODEL_NAME:$VERSION_NAME"
gcloud ai-platform versions create \
--model= # TODO -- Your code here.
--framework= # TODO -- Your code here.
--python-version= # TODO -- Your code here.
--runtime-version= # TODO -- Your code here.
--origin= # TODO -- Your code here.
--staging-bucket= # TODO -- Your code here.
In [ ]:
%%writefile input.json
{"pickup_longitude": -73.982683, "pickup_latitude": 40.742104,"dropoff_longitude": -73.983766,"dropoff_latitude": 40.755174,"passenger_count": 3.0}
Lab Task #5b: Complete the code in the cell below to call prediction on your deployed model for the example you just created in the input.json
file above.
In [ ]:
# TODO 5b
!gcloud ai-platform predict \
--model # TODO -- Your code here.
--json-instances # TODO -- Your code here.
--version # TODO -- Your code here.
Copyright 2019 Google Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License