In [0]:
#@title Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
In [0]:
import tensorflow as tf
import numpy as np
Tensors are multi-dimensional arrays with a uniform type (called a dtype
). You can see all supported dtypes
at tf.dtypes.DType
.
If you're familiar with NumPy, tensors are (kind of) like np.arrays
.
All tensors are immutable like python numbers and strings: you can never update the contents of a tensor, only create a new one.
Here is a "scalar" or "rank-0" tensor . A scalar contains a single value, and no "axes".
In [0]:
# This will be an int32 tensor by default; see "dtypes" below.
rank_0_tensor = tf.constant(4)
print(rank_0_tensor)
A "vector" or "rank-1" tensor is like a list of values. A vector has 1-axis:
In [0]:
# Let's make this a float tensor.
rank_1_tensor = tf.constant([2.0, 3.0, 4.0])
print(rank_1_tensor)
A "matrix" or "rank-2" tensor has 2-axes:
In [0]:
# If we want to be specific, we can set the dtype (see below) at creation time
rank_2_tensor = tf.constant([[1, 2],
[3, 4],
[5, 6]], dtype=tf.float16)
print(rank_2_tensor)
A scalar, shape: [] |
A vector, shape: [3] |
A matrix, shape: [3, 2] |
---|---|---|
|
|
|
Tensors may have more axes, here is a tensor with 3-axes:
In [0]:
# There can be an arbitrary number of
# axes (sometimes called "dimensions")
rank_3_tensor = tf.constant([
[[0, 1, 2, 3, 4],
[5, 6, 7, 8, 9]],
[[10, 11, 12, 13, 14],
[15, 16, 17, 18, 19]],
[[20, 21, 22, 23, 24],
[25, 26, 27, 28, 29]],])
print(rank_3_tensor)
There are many ways you might visualize a tensor with more than 2-axes.
A 3-axis tensor, shape: [3, 2, 5] |
||
---|---|---|
|
|
|
You can convert a tensor to a NumPy array either using np.array
or the tensor.numpy
method:
In [0]:
np.array(rank_2_tensor)
In [0]:
rank_2_tensor.numpy()
Tensors often contain floats and ints, but have many other types, including:
The base tf.Tensor
class requires tensors to be "rectangular"---that is, along each axis, every element is the same size. However, there are specialized types of Tensors that can handle different shapes:
We can do basic math on tensors, including addition, element-wise multiplication, and matrix multiplication.
In [0]:
a = tf.constant([[1, 2],
[3, 4]])
b = tf.constant([[1, 1],
[1, 1]]) # Could have also said `tf.ones([2,2])`
print(tf.add(a, b), "\n")
print(tf.multiply(a, b), "\n")
print(tf.matmul(a, b), "\n")
In [0]:
print(a + b, "\n") # element-wise addition
print(a * b, "\n") # element-wise multiplication
print(a @ b, "\n") # matrix multiplication
Tensors are used in all kinds of operations (ops).
In [0]:
c = tf.constant([[4.0, 5.0], [10.0, 1.0]])
# Find the largest value
print(tf.reduce_max(c))
# Find the index of the largest value
print(tf.argmax(c))
# Compute the softmax
print(tf.nn.softmax(c))
Tensors have shapes. Some vocabulary:
Note: Although you may see reference to a "tensor of two dimensions", a rank-2 tensor does not usually describe a 2D space.
Tensors and tf.TensorShape
objects have convenient properties for accessing these:
In [0]:
rank_4_tensor = tf.zeros([3, 2, 4, 5])
A rank-4 tensor, shape: [3, 2, 4, 5] |
|
---|---|
|
|
In [0]:
print("Type of every element:", rank_4_tensor.dtype)
print("Number of dimensions:", rank_4_tensor.ndim)
print("Shape of tensor:", rank_4_tensor.shape)
print("Elements along axis 0 of tensor:", rank_4_tensor.shape[0])
print("Elements along the last axis of tensor:", rank_4_tensor.shape[-1])
print("Total number of elements (3*2*4*5): ", tf.size(rank_4_tensor).numpy())
While axes are often referred to by their indices, you should always keep track of the meaning of each. Often axes are ordered from global to local: The batch axis first, followed by spatial dimensions, and features for each location last. This way feature vectors are contiguous regions of memory.
Typical axis order |
---|
|
TensorFlow follow standard python indexing rules, similar to indexing a list or a string in python, and the bacic rules for numpy indexing.
0
:
, are used for slices start:stop:step
In [0]:
rank_1_tensor = tf.constant([0, 1, 1, 2, 3, 5, 8, 13, 21, 34])
print(rank_1_tensor.numpy())
Indexing with a scalar removes the dimension:
In [0]:
print("First:", rank_1_tensor[0].numpy())
print("Second:", rank_1_tensor[1].numpy())
print("Last:", rank_1_tensor[-1].numpy())
Indexing with a :
slice keeps the dimension:
In [0]:
print("Everything:", rank_1_tensor[:].numpy())
print("Before 4:", rank_1_tensor[:4].numpy())
print("From 4 to the end:", rank_1_tensor[4:].numpy())
print("From 2, before 7:", rank_1_tensor[2:7].numpy())
print("Every other item:", rank_1_tensor[::2].numpy())
print("Reversed:", rank_1_tensor[::-1].numpy())
Higher rank tensors are indexed by passing multiple indices.
The single-axis exact same rules as in the single-axis case apply to each axis independently.
In [0]:
print(rank_2_tensor.numpy())
Passing an integer for each index the result is a scalar.
In [0]:
# Pull out a single value from a 2-rank tensor
print(rank_2_tensor[1, 1].numpy())
You can index using any combination integers and slices:
In [0]:
# Get row and column tensors
print("Second row:", rank_2_tensor[1, :].numpy())
print("Second column:", rank_2_tensor[:, 1].numpy())
print("Last row:", rank_2_tensor[-1, :].numpy())
print("First item in last column:", rank_2_tensor[0, -1].numpy())
print("Skip the first row:")
print(rank_2_tensor[1:, :].numpy(), "\n")
Here is an example with a 3-axis tensor:
In [0]:
print(rank_3_tensor[:, :, 4])
Selecting the last feature across all locations in each example in the batch | |
---|---|
|
|
In [0]:
# Shape returns a `TensorShape` object that shows the size on each dimension
var_x = tf.Variable(tf.constant([[1], [2], [3]]))
print(var_x.shape)
In [0]:
# You can convert this object into a Python list, too
print(var_x.shape.as_list())
You can reshape a tensor into a new shape. Reshaping is fast and cheap as the underlying data does not need to be duplicated.
In [0]:
# We can reshape a tensor to a new shape.
# Note that we're passing in a list
reshaped = tf.reshape(var_x, [1, 3])
In [0]:
print(var_x.shape)
print(reshaped.shape)
The data maintains it's layout in memory and a new tensor is created, with the requested shape, pointing to the same data. TensorFlow uses C-style "row-major" memory ordering, where incrementing the right-most index corresponds to a single step in memory.
In [0]:
print(rank_3_tensor)
If you flatten a tensor you can see what order it is laid out in memory.
In [0]:
# A `-1` passed in the `shape` argument says "Whatever fits".
print(tf.reshape(rank_3_tensor, [-1]))
Typically the only reasonable uses of tf.reshape
are to combine or split adjacent axes (or add/remove 1
s).
For this 3x2x5 tensor, reshaping to (3x2)x5 or 3x(2x5) are both reasonable things to do, as the slices do not mix:
In [0]:
print(tf.reshape(rank_3_tensor, [3*2, 5]), "\n")
print(tf.reshape(rank_3_tensor, [3, -1]))
Some good reshapes. | ||
---|---|---|
|
|
|
Reshaping will "work" for any new shape with the same total number of elements, but it will not do anything useful if you do not respect the order of the axes.
Swapping axes in tf.reshape
does not work, you need tf.transpose
for that.
In [0]:
# Bad examples: don't do this
# You can't reorder axes with reshape.
print(tf.reshape(rank_3_tensor, [2, 3, 5]), "\n")
# This is a mess
print(tf.reshape(rank_3_tensor, [5, 6]), "\n")
# This doesn't work at all
try:
tf.reshape(rank_3_tensor, [7, -1])
except Exception as e:
print(f"{type(e).__name__}: {e}")
Some bad reshapes. | ||
---|---|---|
|
|
|
You may run across not-fully-specified shapes. Either the shape contains a None
(a dimension's length is unknown) or the shape is None
(the rank of the tensor is unknown).
Except for tf.RaggedTensor, this will only occur in the context of TensorFlow's, symbolic, graph-building APIs:
DTypes
To inspect a tf.Tensor
's data type use the Tensor.dtype
property.
When creating a tf.Tensor
from a Python object you may optionally specify the datatype.
If you don't, TensorFlow chooses a datatype that can represent your data. TensorFlow converts Python integers to tf.int32
and python floating point numbers to tf.float32
. Otherwise TensorFlow uses the same rules NumPy uses when converting to arrays.
You can cast from type to type.
In [0]:
the_f64_tensor = tf.constant([2.2, 3.3, 4.4], dtype=tf.float64)
the_f16_tensor = tf.cast(the_f64_tensor, dtype=tf.float16)
# Now, let's cast to an uint8 and lose the decimal precision
the_u8_tensor = tf.cast(the_f16_tensor, dtype=tf.uint8)
print(the_u8_tensor)
Broadcasting is a concept borrowed from the equivalent feature in NumPy. In short, under certain conditions, smaller tensors are "stretched" automatically to fit larger tensors when running combined operations on them.
The simplest and most common case is when you attempt to multiply or add a tensor to a scalar. In that case, the scalar is broadcast to be the same shape as the other argument.
In [0]:
x = tf.constant([1, 2, 3])
y = tf.constant(2)
z = tf.constant([2, 2, 2])
# All of these are the same computation
print(tf.multiply(x, 2))
print(x * y)
print(x * z)
Likewise, 1-sized dimensions can be stretched out to match the other arguments. Both arguments can be stretched in the same computation.
In this case a 3x1 matrix is element-wise multiplied by a 1x4 matrix to produce a 3x4 matrix. Note how the leading 1 is optional: The shape of y is [4]
.
In [0]:
# These are the same computations
x = tf.reshape(x,[3,1])
y = tf.range(1, 5)
print(x, "\n")
print(y, "\n")
print(tf.multiply(x, y))
A broadcasted add: a [3, 1] times a [1, 4] gives a [3,4] |
---|
|
Here is the same operation without broadcasting:
In [0]:
x_stretch = tf.constant([[1, 1, 1, 1],
[2, 2, 2, 2],
[3, 3, 3, 3]])
y_stretch = tf.constant([[1, 2, 3, 4],
[1, 2, 3, 4],
[1, 2, 3, 4]])
print(x_stretch * y_stretch) # Again, operator overloading
Most of the time, broadcasting is both time and space efficient, as the broadcast operation never materializes the expanded tensors in memory.
You see what broadcasting looks like using tf.broadcast_to
.
In [0]:
print(tf.broadcast_to(tf.constant([1, 2, 3]), [3, 3]))
Unlike a mathematical op, for example, broadcast_to
does nothing special to save memory. Here, you are materializing the tensor.
It can get even more complicated. This section of Jake VanderPlas's book Python Data Science Handbook shows more broadcasting tricks (again in NumPy).
Most ops, like tf.matmul
and tf.reshape
take arguments of class tf.Tensor
. However, you'll notice in the above case, we frequently pass Python objects shaped like tensors.
Most, but not all, ops call convert_to_tensor
on non-tensor arguments. There is a registry of conversions, and most object classes like NumPy's ndarray
, TensorShape
, Python lists, and tf.Variable
will all convert automatically.
See tf.register_tensor_conversion_function
for more details, and if you have your own type you'd like to automatically convert to a tensor.
A `tf.RaggedTensor`, shape: [4, None] |
---|
|
In [0]:
ragged_list = [
[0, 1, 2, 3],
[4, 5],
[6, 7, 8],
[9]]
In [0]:
try:
tensor = tf.constant(ragged_list)
except Exception as e:
print(f"{type(e).__name__}: {e}")
Instead create a tf.RaggedTensor
using tf.ragged.constant
:
In [0]:
ragged_tensor = tf.ragged.constant(ragged_list)
print(ragged_tensor)
The shape of a tf.RaggedTensor
contains unknown dimensions:
In [0]:
print(ragged_tensor.shape)
tf.string
is a dtype
, which is to say we can represent data as strings (variable-length byte arrays) in tensors.
The strings are atomic and cannot be indexed the way Python strings are. The length of the string is not one of the dimensions of the tensor. See tf.strings
for functions to manipulate them.
Here is a scalar string tensor:
In [0]:
# Tensors can be strings, too here is a scalar string.
scalar_string_tensor = tf.constant("Gray wolf")
print(scalar_string_tensor)
And a vector of strings:
A vector of strings, shape: [3,] |
---|
|
In [0]:
# If we have three string tensors of different lengths, this is OK.
tensor_of_strings = tf.constant(["Gray wolf",
"Quick brown fox",
"Lazy dog"])
# Note that the shape is (3,). The string length is not included.
print(tensor_of_strings)
In the above printout the b
prefix indicates that tf.string
dtype is not a unicode string, but a byte-string. See the Unicode Tutorial for more about working with unicode text in TensorFlow.
If you pass unicode characters they are utf-8 encoded.
In [0]:
tf.constant("🥳👍")
Some basic functions with strings can be found in tf.strings
, including tf.strings.split
.
In [0]:
# We can use split to split a string into a set of tensors
print(tf.strings.split(scalar_string_tensor, sep=" "))
In [0]:
# ...but it turns into a `RaggedTensor` if we split up a tensor of strings,
# as each string might be split into a different number of parts.
print(tf.strings.split(tensor_of_strings))
Three strings split, shape: [3, None] |
---|
|
And tf.string.to_number
:
In [0]:
text = tf.constant("1 10 100")
print(tf.strings.to_number(tf.strings.split(text, " ")))
Although you can't use tf.cast
to turn a string tensor into numbers, you can convert it into bytes, and then into numbers.
In [0]:
byte_strings = tf.strings.bytes_split(tf.constant("Duck"))
byte_ints = tf.io.decode_raw(tf.constant("Duck"), tf.uint8)
print("Byte strings:", byte_strings)
print("Bytes:", byte_ints)
In [0]:
# Or split it up as unicode and then decode it
unicode_bytes = tf.constant("アヒル 🦆")
unicode_char_bytes = tf.strings.unicode_split(unicode_bytes, "UTF-8")
unicode_values = tf.strings.unicode_decode(unicode_bytes, "UTF-8")
print("\nUnicode bytes:", unicode_bytes)
print("\nUnicode chars:", unicode_char_bytes)
print("\nUnicode values:", unicode_values)
The tf.string
dtype is used for all raw bytes data in TensorFlow. The tf.io
module contains functions for converting data to and from bytes, including decoding images and parsing csv.
A `tf.SparseTensor`, shape: [3, 4] |
---|
|
In [0]:
# Sparse tensors store values by index in a memory-efficient manner
sparse_tensor = tf.sparse.SparseTensor(indices=[[0, 0], [1, 2]],
values=[1, 2],
dense_shape=[3, 4])
print(sparse_tensor, "\n")
# We can convert sparse tensors to dense
print(tf.sparse.to_dense(sparse_tensor))