chapter
stringclasses 1
value | page
stringlengths 12
12
| content
stringlengths 946
3.15k
| file_type
stringclasses 1
value |
---|---|---|---|
ch13 | page_001.txt | Chapter 13
Neural Nets and Deep
Learning
In Sections 12.2 and 12.3 we discussed the design of single “neurons” (percep-
trons). These take a collection of inputs and, based on weights associated with
those inputs, compute a number that, compared with a threshold, determines
whether to output “yes” or “no.” These methods allow us to separate inputs
into two classes, as long as the classes are linearly separable. However, most
problems of interest and importance are not linearly separable. In this chapter,
we shall consider the design of neural nets, which are collections of perceptrons,
or nodes, where the outputs of one rank (or layer of nodes becomes the inputs
to nodes at the next layer. The last layer of nodes produces the outputs of
the entire neural net. The training of neural nets with many layers requires
enormous numbers of training examples, but has proven to be an extremely
powerful technique, referred to as deep learning, when it can be used.
We also consider several specialized forms of neural nets that have proved
useful for special kinds of data. These forms are characterized by requiring that
certain sets of nodes in the network share the same weights. Since learning all
the weights on all the inputs to all the nodes of the network is in general a
hard and time-consuming task, these special forms of network greatly simplify
the process of training the network to recognize the desired class or classes of
inputs. We shall study convolutional neural networks (CNN’s), which are spe-
cially designed to recognize classes of images. We shall also study recurrent neu-
ral networks (RNN’s) and long short-term memory networks (LSTM’s), which
are designed to recognize classes of sequences, such as sentences (sequences of
words).
509
| text |
ch13 | page_002.txt | 510
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
13.1
Introduction to Neural Nets
We begin the discussion of neural nets with an extended example. After that,
we introduce the general plan of a neural net and some important terminology.
Example 13.1 : The problem we discuss is to learn the concept that “good”
bit-vectors are those that have two consecutive 1’s. Since we want to deal with
only tiny example instances, we shall assume bit-vectors have length four. Our
training examples will thus have the form ([x1, x2, x3, x4], y), where each of the
xi’s are bits, 0 or 1. There are 16 possible training examples, and we shall
assume we are given some subset of these as our training set. Notice that eight
of the possible bit vectors are good – they do have consecutive 1’s, and there
are also eight “bad” examples. For instance, 0111 and 1100 are good; 1001 and
0100 are bad.1
To start, let us look at a neural net that solves this simple problem exactly.
How we might design this net from training examples is the true subject for
discussion, but this net will serve as an example of what we would like to
achieve. The net is shown in Fig. 13.1.
x 1
x 2
x 3
x 4
x 1
x 2
x 3
x 4
0.5
y
1
1
2.5
1.5
1
1
2
0
0
0
1
1
Figure 13.1: A neural net that tells whether a bit-vector has consecutive 1’s
The net has two layers, the first consisting of two nodes, and the second
with a single node that produces the output y. Each node is a perceptron,
exactly as was described in Section 12.2. In the first layer, the first node is
characterized by weight vector [w1, w2, w3, w4] = [1, 2, 1, 0] and threshold 2.5.
Since each input xi is either 0 or 1, we note that the only way to reach a sum
P4
i=1 xiwi as high as 2.5 is if x2 = 1 and at least one of x1 and x3 is also 1.
The output of this node is 1 if and only if the input is one of 1100, 1101, 1110,
1111, 0110, or 0111. That is, it recognizes those bit-vectors that either begin
with two 1’s or have two 1’s in the middle. The only good inputs it does not
1We shall show bit vectors as bit strings in what follows, so we can avoid the commas
between components, each of which is 0 or 1.
| text |
ch13 | page_003.txt | 13.1. INTRODUCTION TO NEURAL NETS
511
recognize are those that end with 11 but do not have 11 elsewhere. these are
0011 and 1011.
Fortunately, the second node in the first layer, with weights [0, 0, 1, 1] and
threshold 1.5 gives output 1 whenever x3 = x4 = 1, and not otherwise. This
node thus recognizes the inputs 0011 and 1011, as well as some other good
inputs that are also recognized by the first node.
Now, let us turn to the second layer, with a single node; that node has
weights [1, 1] and threshold 0.5. It thus behaves as an “OR-gate.” It gives
output y = 1 whenever either or both of the nodes in the first layer have output
1, but gives output y = 0 if both of the first-layer nodes give output 0. Thus,
the neural net of Fig. 13.1 gives output 1 for all the good inputs but none of
the bad inputs.
✷
x 1
x 2
x 3
x 4
x 1
x 2
x 3
x 4
y
1
1
1
2
0
0
0
1
1
1
1
−0.5
1
−2.5
1
−1.5
Figure 13.2: Making the threshold 0 for all nodes
It is useful in many situations to assume that nodes have a threshold of
0. Recall from Section 12.2.4 that we can always convert a perceptron with a
nonzero threshold t to one with a 0 threshold if we add an additional input.
That input always has value 1 and a weight equal to −t. For example, we can
convert the net of Fig. 13.1 to that in Fig. 13.2.
13.1.1
Neural Nets, in General
Example 13.1 and its net of Fig. 13.1 is much simpler than anything that would
be a useful application of neural nets. The general case is suggested by Fig. 13.3.
The first, or input layer, is the input, which is presumed to be a vector of some
length n. Each component of the vector [x1, x2, . . . , xn] is an input to the net.
There are one or more hidden layers and finally, at the end, an output layer,
which gives the result of the net. Each of the layers can have a different number
of nodes, and in fact, choosing the right number of nodes at each layer is an
| text |
ch13 | page_004.txt | 512
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
important part of the design process for neural nets. Especially, note that the
output layer can have many nodes. For instance, the neural net could classify
inputs into many different classes, with one output node for each class.
.
.
.
.
.
.
.
.
.
. . .
x
Hidden layers
Output
layer
x
1
2
.
.
.
x n
y
y
y 1
2
k
Input
layer
Figure 13.3: The general case of a neural network
Each layer, except for the input layer, consists of one or more nodes, which
we arrange in the column that represents that layer. We can think of each node
as a perceptron. The inputs to a node are outputs of some or all of the nodes in
the previous layer. So that we can assume the threshold for each node is zero,
we can also allow a node to have an input that is a constant, typically 1, as
we suggested in Fig. 13.2. Associated with each input to each node is a weight.
The output of the node depends on P xiwi, where the sum is over all the inputs
xi, and wi is the weight of that input. Sometimes, the output is either 0 or 1;
the output is 1 if that sum is positive and 0 otherwise. However, as we shall
see in Section 13.2, it is often convenient, when trying to learn the weights for
a neural net that solves some problem, to have outputs that are almost always
close to 0 or 1, but may be slightly different. The reason, intuitively, is that
it is then possible for the output of a node to be a continuous function of its
inputs. We can then use gradient descent to converge to the ideal values of all
the weights in the net.
| text |
ch13 | page_005.txt | 13.1. INTRODUCTION TO NEURAL NETS
513
13.1.2
Interconnections Among Nodes
Neural nets can differ in how the nodes at one layer are connected to nodes
at the layer to its right. The most general case is when each node receives as
inputs the outputs of every node of the previous layer. A layer that receives
all outputs from the previous layer is said to be fully connected. Some other
options for choosing interconnections are:
1. Random. For some m, we pick for each node m nodes from the previous
layer and make those, and only those, be inputs to this node.
2. Pooled. Partition the nodes of one layer into some number of clusters. In
the next layer, which is called a pooled layer, there is one node for each
cluster, and this node has all and only the member of its cluster as inputs.
3. Convolutional.
This approach to interconnection, which we discuss in
more detail in the next section and Section 13.4, views the nodes of each
layer as arranged in a grid, typically two-dimensional. In a convolutional
layer, a node corresponding to coordinates (i, j) receive as inputs the
nodes of the previous layer that have coordinates in some small region
around (i, j). For example, the node (i, j) at one convolutional layer may
have as inputs those nodes from the previous layer that correspond to
coordinates (p, q), where i ≤p ≤i + 2 and j ≤q ≤j + 2 (i.e., the square
of side 3 whose lower-left corner is the point (i, j).
13.1.3
Convolutional Neural Networks
A convolutional neural network, or CNN, contains one or more convolutional
layers. There can also be nonconvolutional layers, such as fully connected layers
and pooled layers. However, there is an important additional constraint: the
weights on the inputs must be the same for all nodes of a single convolutional
layer. More precisely, suppose that each node (i, j) in a convolutional layer
recieves (i + u, j + v) as one of its inputs, where u and v are small constants.
Then there is a weight w associated with u and v (but not with i and j). For
any i and j, the weight on the input to the node for (i, j) coming from the
output of the node i + u, j + v) from the previous layer must be w.
This restriction makes training a CNN much more efficient than training a
general neural net. The reason is that there are many fewer parameters at each
layer, and therefore, many fewer training examples can be used, than if each
node or each layer has its own weights for the training process to discover.
CNN’s have been found extremely useful for tasks such as image recognition.
In fact, the CNN draws inspiration from the way the human eye processes
images. The neurons of the eye are arranged in layers, similarly to the layers
of a neural net. The first layer takes inputs that are essentially pixels of the
image, each pixel the result of a sensor in the retina. The nodes of the first
layer recognize very simple features, such as edges between light and dark.
Notice that a small square of pixels, say 3-by-3, might exhibit an edge at a
| text |
ch13 | page_006.txt | 514
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
particular angle, e.g., if the upper left corner is light and the other eight pixels
dark. Moreover, the algorithm for recognizing an edge of a certain type is the
same, regardless of where in the field of vision this little square appears. That
observation justifies the CNN constraint that all the nodes of a layer use the
same weights. In the eye, additional layers combine results from the previous
layers to recognize more and more complex structures: long boundaries, regions
of similar color, and finally faces and all the familiar objects that we see daily.
We shall have more to say about convolutional neural networks in Sec-
tion 13.4. Moreover, CNN’s are only one example of a kind of neural network
where certain families of nodes are constrained to have the same weights. For
example, in Section 13.5, we shall consider recurrent neural networks and long
short-term memory networks, which are specially adapted to recognizing prop-
erties of sequences, such as sentences (sequences of words).
13.1.4
Design Issues for Neural Nets
Building a neural net to solve a given problem is partially art and partially
science. Before we can begin to train a net by finding the weights on the inputs
that serve our goals best, we need to make a number of design decisions. These
include answering the following questions:
1. How many hidden layers should we use?
2. How many nodes will there be in each of the hidden layers?
3. In what manner will we interconnect the outputs of one layer to the inputs
of the next layer?
Further, in later sections we shall see that there are other decisions that need
to be made when we train the neural net. These include:
4. What cost function should we minimize to express what weights are best?
5. How should we compute the outputs of each gate as a function of the
inputs? We have suggested that the normal way to compute output is to
take a weighted sum of inputs and compare it to 0. But there are other
computations that serve better in common circumstances.
6. What algorithm do we use to exploit the training examples in order to
optimize the weights?
13.1.5
Exercises for Section 13.1
!! Exercise 13.1.1: Prove that the problem of Example 13.1 cannot be solved
by a perceptron; i.e., the good and bad points are not linearly separable.
| text |
ch13 | page_007.txt | 13.2. DENSE FEEDFORWARD NETWORKS
515
! Exercise 13.1.2: Consider the general problem of identifying bit-vectors of
length n having two consecutive 1’s. Assume a single hidden layer with some
number of gates. What is the smallest number of gates you can have in the
hidden layer if (a) n = 5 (b) n = 6?
! Exercise 13.1.3: Design a neural net that functions as an exclusive-or gate,
that is, it takes two inputs and gives output 1 if exactly one of the inputs is
1 gives output 0 otherwise. Hint: remember that both weights and thresholds
can be negative.
! Exercise 13.1.4: Prove that there is no single preceptron that behaves like
an exclusive-or gate.
! Exercise 13.1.5: Design a neural net to compute the exclusive-or of three
inputs; that is, output 1 if an odd number of the three inputs is 1 and output
0 if an even number of the inputs are 1.
13.2
Dense Feedforward Networks
In the previous section, we simply exhibited a neural net that worked for the
“consecutive 1’s” problem. However, the true value of neural nets comes from
our ability to design them given training data. To design a net, there are many
choices that must be made, such as the number of layers and the number of
nodes for each layer, as was discussed in Section 13.1.4.
These choices can
be more art than science. The computational part of training, which is more
science than art, is primarily the choice of weights for the inputs to each node.
The techniques for selecting weights usually involve convergence using gra-
dient descent. But gradient descent requires a cost function, which must be a
continuous function of the weights. The nets discussed in Section 13.1.4, on the
other hand, use perceptrons whose output is either 0 or 1, so the outputs are
not normally continuous functions of the inputs. In this section, we shall dis-
cuss the various ways one can modify the behavior of the nodes in a net so the
outputs become continuous functions of the inputs, and therefore a reasonable
cost function applied to the outputs will also be continuous.
13.2.1
Linear Algebra Notation
We can succinctly describe the neural network we used for the consecutive
1’s problem using linear algebra notation.
The input nodes form a vector2
x = [x1, x2, x3, x4], while the hidden nodes form a vector h = [h1, h2]. The
4 edges connecting the input to hidden node 1 form a weight vector w1 =
[w11, w12, w13, w14], and similarly we have weight vector w2 for hidden node 2.
2We assume all vectors are column vectors by default. However, it is often more convenient
to write row vectors, and we shall do so in the text. But in formulas, we shall use the transpose
operator when we actually want to use the vector as a row rather than a column.
| text |
ch13 | page_008.txt | 516
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Why Use Linear Algebra?
Notational brevity is one reason to use linear algebra notation for neu-
ral networks. Another is performance. It turns out that graphics pro-
cessing units (GPU’s) have circuitry that allows for highly parallelized
linear-algebra operations. Mutiplying a matrix and a vector using a sin-
gle linear algebra operator is much faster than coding the same operator
using nested loops. Modern deep-learning frameworks (e.g., TensorFlow,
PyTorch, Caffe) harness the power of GPU’s to dramatically speed up
neural network computations.
The threshold inputs to the hidden layer nodes (i.e., the negatives of the thresh-
olds) form a 2-vector b = [b1, b2], often called the bias vector. The perceptron
applies the nonlinear step function to produce its output, defined as:
step(z) =
(
1
when z > 0
0
otherwise
Each hidden node hi can now be described using the expression:
hi = step(wT
i x + bi) for i = 1, 2
We could organize the weight vectors w1 and w2 into a 2 × 4 weight matrix
W, where the ith row of W is wT
i . The hidden nodes can thus be described
using the expression:
h = step(Wx + b)
In the case of a vector input, the step function just operates element-wise on
the vector to produce an output vector of the same length. We can use a similar
arrangement to describe the transformation that produces the final output from
the hidden layer. In this case, the final output is a scalar y, so instead of a weight
matrix W we need only a weight vector u = [u1, u2] and a single bias c. We
thus have:
y = step(uTh + c)
Linear-algebra notation works just as well when we have larger inputs and
many more nodes in one hidden layer. We just need to scale the weight matrix
and bias vector appropriately. That is, the matrix W has one row for each
node in the layer and one column for each output from the previous layer (or
for each input, if this is the first layer); the bias vector has one component
for each node. It is also easy to handle the case where there is more than one
node in the output layer. For example, in a multiclass classification problem, we
might have an output node yi corresponding to target class i. For a given input,
the outputs specify the probability that the input belongs to the corresponding
| text |
ch13 | page_009.txt | 13.2. DENSE FEEDFORWARD NETWORKS
517
class. This arrangement results in an output vector y = [y1, y2, . . . , yn], where
n is number of classes. The simple network from the prior section had a boolean
output, corresponding to two output classes (true and false), so we could have
modeled the output equally well as a 2-vector. In the case of a vector output,
we connect the hidden and output layers by a weight matrix of appropriate
dimensions in place of the weight vector we used for the example.
The perceptrons in our example used a nonlinear step function. More gen-
erally, we can use any other nonlinear function, called the activation function,
following the linear transformation.
We describe commonly used activation
functions starting in Section 13.2.2.
Our simple example used a single hidden layer of nodes between the input
and output layers. In general, we can have many hidden layers, as was sugested
in Fig. 13.3. Each hidden layer introduces an additional matrix of weights and
vector of biases, as well as its own activation function. This kind of network is
called a feedforward network, since all edges are oriented “forward,” from input
to output, without cycles.
Suppose there are ℓhidden layers and an additional output layer, numbered
ℓ+ 1. Let the weight matrix for the ith layer be Wi and let the bias vector
for that layer be bi. The weights W1, W2, . . . , Wl+1 and biases b1, b2, . . . , bl+1
constitute the parameters of the model. Our objective is to learn the best values
for these parameters to achieve the task at hand. We will soon describe how to
go about learning the model parameters.
13.2.2
Activation Functions
A node (perceptron) in a neural net is designed to give a 0 or 1 (yes or no)
output. Often, we want to modify that output is one of several ways, so we apply
an activation function to the output of a node. In some cases, the activation
function takes all the outputs of a layer and modifies them as a group. the
reason we need an activation function is as follows. The approach we shall use
to learn good parameter values for the network is gradient descent. Thus, we
need activation functions that “play well” with gradient descent. In particular,
we look for activation functions with the following properties:
1. The function is continuous and differentiable everywhere (or almost ev-
erywhere).
2. The derivative of the function does not saturate (i.e., become very small,
tending towards zero) over its expected input range. Very small deriva-
tives tend to stall out the learning process.
3. The derivative does not explode (i.e., become very large, tending towards
infinity), since this would lead to issues of numerical instability.
The step function does not satisfy conditions (2) and (3). Its derivative explodes
at 0 and is 0 everywhere else. Thus the step function does not play well with
gradient descent and is not a good choice for a deep neural network.
| text |
ch13 | page_010.txt | 518
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
13.2.3
The Sigmoid
Given that we cannot use the step function, we look for alternatives in the
class of sigmoid functions – so called because of the S-shaped curve that these
functions exhibit. The most commonly used sigmoid function is the logistic
sigmoid:
σ(x) =
1
1 + e−x =
ex
1 + ex
Notice that the sigmoid has the value 1/2 at x = 0. For large x, the sigmoid
approaches 1, and for large, negtive x, the sigmoid approaches 0.
The logistic sigmoid, like all the functions we shall discuss, are applied to
vectors elementwise, so if x = [x1, x2, . . . , xn] then
σ(x) = [σ(x1), σ(x2), . . . , σ(xn)]
The logistic sigmoid has several advantages over the step function as a way
to define the output of a perceptron. The logistic sigmoid is continuous and dif-
ferentiable, so it enables us to use gradient descent to discover the best weights.
Since its value is in the range [0, 1], it is possible to interpret the outputs of the
network as a probability. However, the logistic sigmoid saturates very quickly
as we move away from the “critical region” around 0. So the derivative goes
towards zero and gradient-based learning can stall out. That is, weights almost
stop changing, once they get away from 0.
In Section 13.3.3, when we describe the backpropagation algorithm, we shall
see that we need the derivatives of activation functions and loss functions. As
an exercise, you can verify that if y = σ(x), then
dy
dx = y(1 −y)
13.2.4
The Hyperbolic Tangent
Closely related to sigmoid is the hyperbolic tangent function, defined by:
tanh(x) = ex −e−x
ex + e−x
Simple algebraic manipulation yields:
tanh(x) = 2σ(2x) −1
So the hyperbolic tangent is just a scaled and shifted version of the sigmoid.
It has two desirable properties that make it attractive in some situations: its
output is in the range [−1, 1] and is symmetric around 0. It also shares the
good properties and the saturation problem of the sigmoid. You may show that
if y = tanh(x) then
dy
dx = 1 −y2
| text |
ch13 | page_011.txt | 13.2. DENSE FEEDFORWARD NETWORKS
519
6
4
2
0
2
4
6
0.5
1.0
3
2
1
1
2
3
1.0
0.5
0.5
1.0
(a)
(b)
Figure 13.4: The logistic sigmoid (a) and hyperbolic tangent (b) functions
Figure 13.4 shows the logistic sigmoid and hyperbolic tangent functions.
Note the difference in scale along the x-axis between the two charts. It is easy
to see that the functions are identical after shifting and scaling.
13.2.5
Softmax
The softmax function differs from sigmoid functions in that it does not operate
element-wise on a vector. Rather, the softmax function applies to an entire
vector. If x = [x1, x2, . . . , xn], then its softmax µ(x) = [µ(x1), µ(x2), . . . , µ(xn)]
where
µ(xi) =
exi
P
j exj
Softmax pushes the largest component of the vector towards 1 while pushing all
the other components towards zero. Also, all the outputs sum to 1, regardless
of the sum of the components of the input vector. Thus, the output of the
softmax function can be intepreted as a probability distribution.
A common application is to use softmax in the output layer for a classi-
fication problem. The output vector has a component corresponding to each
target class, and the softmax output is interpreted as the probability of the
input belonging to the corresponding class.
Softmax has the same saturation problem as the sigmoid function, since one
component gets larger than all the others. There is a simple workaround to this
problem, however, when softmax is used at the output layer. In this case, it is
usual to pick cross entropy as the loss function, which undoes the exponentiation
in the definition of softmax and avoids saturation. Cross entropy is explained in
| text |
ch13 | page_012.txt | 520
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Accuracy of Softmax Calculation
The denominator of the softmax function involves computing a sum of the
form P
j exj. When the xj’s take a wide range of values, their exponents
exj take on an even wider range of values – some tiny and some very large.
Adding very large and very small floating point numbers leads to numerical
inaccuracy issues in fixed-width floating point representations (such as 32-
bit or 64-bit). Fortunately, there is a trick to avoid this problem. We
observe that
µ(xi) =
exi
P
j exj =
exi−c
P
j exj−c
for any constant c. We pick c = maxj xj, so that xj −c ≤0 for all j.
This ensures that exj−c is always between 0 and 1, and leads to a more
accurate calculation.
Most deep learning frameworks will take care to
compute softmax in this manner.
Section 13.2.9. We address the problem of differentiating the softmax function
in Section 13.3.3.
13.2.6
Recified Linear Unit
The rectified linear unit, or ReLU, is defined as:
f(x) = max(0, x) =
(
x,
for x ≥0
0,
for x < 0
The name of this function derives from the analogy to half-wave rectification in
electrical engineering. The function is not differentiable at 0 but is differentiable
everywhere else, including at points arbitrarily close to 0. In practice, we “set”
the derivative at 0 to be either 0 (the left derivative) or 1 (the right derivative).
In modern neural nets, a version of ReLU has replaced sigmoid as the de-
fault choice of activation function. The popularity of ReLU derives from two
properties:
1. The gradient of ReLU remains constant and never saturates for positive
x, speeding up training.
It has been found in practice that networks
that use ReLU offer a significat speedup in training compared to sigmoid
activation.
2. Both the function and its derivative can be computed using elementary
and efficient mathematical operations (no exponentiation).
ReLU does suffer from a problem related to the saturation of its derivative
when x < 0. Once a node’s input values become negative, it is possible that the
| text |
ch13 | page_013.txt | 13.2. DENSE FEEDFORWARD NETWORKS
521
3
2
1
1
2
3
1
1
2
3
3
2
1
1
2
3
1
1
2
3
(a)
(b)
Figure 13.5: The ReLU (a) and ELU (b), with α = 1 functions
node’s output get “stuck” at 0 through the rest of the training. This is called
the dying ReLU problem.
The Leaky ReLU attempts to fix this problem by defining the activation
function as follows:
f(x) =
(
x,
for x ≥0
αx,
for x < 0
where α is typically a small positive value such as 0.01. The Parametric ReLU
(PReLU) makes α a parameter to be optimized as part of the learning process.
An improvement on both the original and leaky ReLU functions is Expo-
nential Linear Unit, or ELU. This function is defined as:
f(x) =
(
x,
for x ≥0
α(ex −1),
for x < 0
where α ≥0 is a hyperparameter. That is, α is held fixed during the learning
process, but we can repeat the learning process with different values of α to
find the best value for our problem.
The node’s value saturates to −α for
large negative values of x, and a typical choice is α = 1.
ELU’s drive the
mean activation of nodes towards zero, which appears to speed up the learning
process compared to other ReLU variants.
13.2.7
Loss Functions
A loss function quantifies the difference between a model’s predictions and the
output values observed in the real world (i.e., in the training set). Suppose
| text |
ch13 | page_014.txt | 522
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
the observed output corresponding to input x is ˆy and the predicted output is
y. Then a loss fucntion L(y, ˆy) quantifies the prediction error for this single
input. Typically, we consider the loss over a large set of observations, such
as the entire training set. In that case, we usually average the losses over all
training examples.
We shall consider separately two cases. In the first case, there is a single
output node, and it produces a real value. In this case we study “regression
loss.” In the second case, there are several output nodes, each of which indicates
that the input is a member of a particular class; we study this matter under
“classification loss” in Section 13.2.9.
13.2.8
Regression Loss
Suppose the model has a single continuous-valued output, and (x, ˆy) is a training
example. For the same input x, suppose the predicted output of the neural net
is y. Then the squared error loss L(y, ˆy) of this prediction is:
L(y, ˆy) = (y −ˆy)2
In general, we compute the loss for a set of predictions. Suppose the observed
(i.e., training set) input-output pairs are T = {(x1, ˆy1), (x2, ˆy2), . . . , (xn, ˆyn)},
while the corresponding input-output pairs predicted by the model are P =
{(x1, y1), (x2, y2), . . . , (xn, yn)}. The mean squared error (MSE) for the set is:
L(P, T ) = 1
n
n
X
i=1
(yi −ˆyi)2
Note that the mean squared error is just square of the RMSE. It is convenient
to omit the square root to simplify the derivative of the function, which we shall
use during training. In any case, when we minimize MSE we also automatically
minimize RMSE.
One problem with MSE is that it is very sensitive to outliers due the squared
term. A few outliers can contribute very highly to the loss and swamp out the
effect of other points, making the training process susceptible to wild swings.
One way to deal with this issue is to use the Huber Loss. Suppose z = y −ˆy,
and δ is a constant. The Huber Loss Lδ is given by:
Lδ(z) =
(
z2
if |z| ≤δ
2δ(|z| −1
2δ)
otherwise
Figure 13.6 contrasts the squared error and Huber loss functions.
In the case where we have a vector y of outputs rather than a single output,
we use ∥y −ˆy∥in place of |y −ˆy| in the definitions of mean squared error and
Huber loss.
| text |
ch13 | page_015.txt | 13.2. DENSE FEEDFORWARD NETWORKS
523
5
4
3
2
1
1
2
3
4
5
5
10
15
20
25
Figure 13.6: Huber Loss (solid line, δ = 1) and Squared Error (dotted line) as
functions of z = y −ˆy
13.2.9
Classification Loss
Consider a multiclass classification problem with target classes C1, C2, . . . , Cn.
Suppose each point in the training set is of the form (x, p) where x is the input
and p = [p1, p2, . . . , pn] is the output. Here pi gives the probability that the
input x belongs to class Ci, with P
i pi = 1. In many cases, we are certain that
an input belongs to a particular class Ci; in this case pi = 1 and pj = 0 for
i ̸= j. In general, we may interpret pi as our level of certainty that input x
belongs to class Ci, and p as a probability distribution over the target classes.
We design our neural network to produce as output a vector
q = [q1, q2, . . . , qn]
of probabilities, with P
i qi = 1. As before, we interpret q as a probability
distribution over the target classes, with qi denoting the model’s probability
that input x belongs to target class Ci. In Section 13.2.5 we described a sim-
ple method to produce such a probability vector as output: use the softmax
activation function in the output layer of the network.
Since both the labeled output and the model’s output are probability dis-
tributions, it is natural to look for a loss function that quantifies the distance
between two probability distributions. Recall the definition of entropy from
Section ??. That is, H(p), the entropy of a discrete probability distribution p
is:
H(p) = −
n
X
i=1
pi log pi
Imagine an alphabet of n symbols, and messages using these symbols. Sup-
pose at each location in the message, the probability that symbol i appears is
| text |
ch13 | page_016.txt | 524
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
pi. Then a key result from information theory is that if we encode messages
using an optimal binary code, the average number of bits per symbol needed to
encode messages is H(p).
Suppose we did not know the symbol probability distribution p when we
design the coding scheme. Instead, we believe that symbols appear following
a different probability distribution q. We might ask what the average number
of bits per symbol will be if we use this suboptimal encoding scheme. A well-
known result from information theory states that the average number of bits in
this case is the cross entropy H(p, q), defined as:
H(p, q) = −
n
X
i=1
pi log qi
Note that H(p, p) = H(p), and in general H(p, q) ≥H(p). The difference
between the cross entropy and the entropy is the average number of additional
bits needed per symbol. It is a reasonable measure of the distance between the
distributions p and q, called the Kullblack-Liebler divergence (KL-divergence)
and denoted D(p∥q):
D(p∥q) = H(p, q) −H(p) =
n
X
i=1
pi log pi
qi
Even though KL-divergence is often regarded as a distance, it is not truly
a distance measure because it is not commutative.
However, it is perfectly
adequate as a loss function for our purposes, since there is in fact an inherent
assymmetry in the situation: p is the ground truth while q is the predicted
output. Notice that minimizing the KL-divergence loss of a model is equivalent
to minimizing the cross-entropy loss, since the term H(p) depends only on the
input and is independent of the model that is learned.
In practice, cross entropy is the most commonly used loss function for clas-
sification problems. Networks designed for classification often use a softmax
activation function in the output layer.
These choices are so common that
many implementations, such as TensorFlow, offer a single function that com-
bines softmax with cross entropy. In addition to convenience, one reason to do
so is that the combined function is more stable numerically, and its derivative
also takes a simple form, as we show in Section 13.3.3.
13.2.10
Exercises for Section 13.2
Exercise 13.2.1: Show that for the logistic sigmoid σ, if y = σ(x), then
dy
dx = y(1 −y)
Exercise 13.2.2: Show that if y = tanh(x) then
dy
dx = 1 −y2
| text |
ch13 | page_017.txt | 13.3. BACKPROPAGATION AND GRADIENT DESCENT
525
Exercise 13.2.3: Show that tanh(x) = 2σ(2x) −1.
Exercise 13.2.4: Show that σ(x) = 1 −σ(−x).
Exercise 13.2.5: Show that for any vector [v1, v2, . . . , vk], Pk
i=1 µ(vi) = 1.
1
y
x 1
x 2
x 3
x 4
y
3
y
2
1
=1
=0
=1
=0
1
2
3
−4
1
−1
2
3
−2
−2
1 −2
1
1
1
4
1
1
2
2
3
Figure 13.7: Neural net for Exercise 13.2.6
Exercise 13.2.6: In Fig. 13.7 is a neural net with paricular values shown
for all the weights and inputs. Suppose that we use the sigmoid function to
compute outputs of nodes at the first layer, and we use softmax to compute the
outputs of the nodes in the output layer.
(a) Compute the values of the outputs for each of the five nodes.
! (b) Assuming each of the weights and each of the xi’s is a variable, express
the output of the first (top) output node in terms of the weights and the
xi’s,
! (c) Find the derivative of your function from part (b) with respect to the
weight on the first (top) input to the first (top) node in the first layer.
13.3
Backpropagation and Gradient Descent
We now turn to the problem of training a deep network. Training a network
means finding good values for the parameters (weights and thresholds) of the
network. Usually, we shall have access to a training set of labeled input/output
pairs. The training process tries to find parameter values that minimize the
average loss on the training set. The hope is that the training set is represen-
tative of the data the model will encounter in the future, and therefore, the
average loss on the training set is a good measure of the average error on all
possible inputs. We must be careful, however; since deep networks have many
| text |
ch13 | page_018.txt | 526
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
parameters, it is possible to find parameters that yield low training loss but
nevertheless perform poorly in the real world. This phenomenon is called over-
fitting, a problem we have mentioned several times, starting in Section 9.4.4.
For the moment, we assume that our goal is to find parameters that minimize
the expected loss on the training set. This goal is achieved by gradient descent.
There is an elegant algorithm called backpropagation that allows us to compute
these gradients efficiently. Before we describe backpropagation, we need a few
preliminaries.
13.3.1
Compute Graphs
A compute graph captures the data flow of a deep network. Formally, a compute
graph is a directed, acyclic graph (DAG). Each node in the compute graph has
an operand and, optionally, an operator. The operand can be a scalar, a vector,
or a matrix. The operator is a linear-algebra operator (such as + or ×), an
activation function (such as σ) or a loss function (such as MSE). When a node
has both an operand and an operator, the operator is written above the operand.
When a node has only an operand, its output is the value associated with
the operand. The output of a node with an operator is the result of applying
its operator to its immediate predecessors in the graph and then assigning the
result to the operand. In general, the operator can be an arbitrary expression
that uses its inputs to produce an output.3
Example 13.2 : Figure 13.8 shows the compute graph for a single-layer dense
network described by y = σ(Wx + b) where x is the input and y is the output.
We then compute an MSE loss against the training-set output ˆy. That is, we
have a single layer of n nodes. The vector y is of length n and represents the
outputs of each of these nodes. There are k inputs, and (x, ˆy) represents one
training example. Matrix W represents the weights on the inputs of the nodes;
that is, Wij is the weight for input j at the ith node. Finally, b represents the
n biases, so its ith element is the negative of the threshold of the ith node.
u
v
y
y
x
W
b
L
MSE
σ
Figure 13.8: Compute graph for a single-layer dense network
For the graph in Fig. 13.8, we have:
3Sometimes the order of operands to the operator matters. We shall ignore that detail
here and assume it is understood from the context.
| text |
ch13 | page_019.txt | 13.3. BACKPROPAGATION AND GRADIENT DESCENT
527
u = Wx
v = u + b
y = σ(v)
L = MSE(y, ˆy)
Each of these steps corresponds to one of the four nodes in the middle row,
in order from the left. The first step corresponds to the node with operand
u and operator ×. Here is an example where it must be understood that the
node labeled W is the first argument. If necessary, we could label each incoming
edge with a number to indicate its order, but in this case the order should be
obvious, since a column vector x could not multiply a matrix W unless the
matrix happened to have only one row. The second step corresponds to the
node with operator + and operand v. Here, the order of arguments does not
matter, since + on vectors is commutative.
✷
13.3.2
Gradients, Jacobians, and the Chain Rule
The goal of the backpropagation algorithm is to compute the gradient of the
loss function with respect to the parameters of the network. Then, we can
adjust the parameters slightly in the directions that will reduce the loss, and
repeat the process until we reach a selection of parameter values for which little
improvement in the loss is possible. Recall the definition of the gradient: given a
function f : RN →R from a real-valued vector to a scalar, if x = [x1, x2, . . . , xn]
and y = f(x) then the gradient of y with respect to x, denoted by ∇xy is given
by
∇xy =
h ∂y
∂x1
, ∂y
∂x2
, . . . , ∂y
∂xn
i
Example 13.3 : We could let function f be the loss function, e.g., the squared-
error loss, which we denote L. This loss is a scalar-valued function of the output
y:
L(y) = ∥y −ˆy∥2 =
n
X
i=1
(yi −ˆyi)2
So we can easily write down the gradient of L with respect to y:
∇yL = [2(y1 −ˆy1), (y2 −ˆy2), . . . , 2(yn −ˆyn)] = 2(y −ˆy)
✷
The generalization of the gradient to vector-valued functions is called the
Jacobian.
Suppose we have a function f : Rm →Rn and y = f(x).
The
Jacobian Jx(y) is given by:4
4The Jacobian is sometimes defined as the transpose of our definition. The formulations
are equivalent. Recall that we are assuming all vectors are column vectors unless transposed,
but we show them as row vectors so they can be written in-line.
| text |
ch13 | page_020.txt | 528
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Jx(y) =
∂y1
∂x1
. . .
∂yn
∂x1
...
...
...
∂y1
∂xm
. . .
∂yn
∂xm
We shall make use of the chain rule for derivatives from calculus. If y = g(x)
and z = f(y) = f(g(x)), then the chain rule says:
dz
dx = dz
dy
dy
dx
Also, if z = f(u, v) where u = g(x) and v = h(x), then
dz
dx = ∂z
∂u
du
dx + ∂z
∂v
dv
dx
For functions of vectors, we can restate the chain rule in terms of gradients and
Jacobians. Suppose y = g(x) and z = f(y) = f(g(x)) then:
∇xz = Jx(y)∇yz
If z = f(u, v) where u = g(x) and v = h(x), then
∇xz = Jx(u)∇uz + Jx(v)∇vz
13.3.3
The Backpropagation Algorithm
The goal of the backpropagation algorithm is to compute the gradient of the loss
function with respect to the parameters of the network. Consider the compute
graph from Fig. 13.8. Here the loss function L is the MSE function. We shall
use the notation g(z) to stand for ∇z(L), that is, the gradient of the the loss
function L with respect to some vector z. We already know the gradient of L
with respect to the output y:
g(y) = ∇y(L) = 2(y −ˆy)
We work backwards through the compute graph, applying the chain rule at
each stage. At each point we pick a node all of whose successors have already
been processed. Suppose a is such a node, and suppose it has just one immediate
successor b in the graph (note that in the simple compute graph of Fig. 13.8,
each node has just one immediate successor). Since we have already processed
node b, we have already computed g(b). We can now compute g(a) using the
chain rule:
g(a) = Ja(b)g(b)
In the case where node a has more than one successor node, we use the more
general sum version of the chain rule. That is, g(a) would be the sum of the
above terms for each successor b of a.
| text |
ch13 | page_021.txt | 13.3. BACKPROPAGATION AND GRADIENT DESCENT
529
Since we shall need to compute these gradients several times, once for each
iteration of gradient descent, we can avoid repeated computation by adding
additional nodes to the compute graph for backpropagation: one node for each
gradient computation.
In general, the Jacobian Ja(b) is a function of both
a and b, and so the node for g(a) will have arcs to it from the nodes for a,
b and g(b).
Popular frameworks for deep learning (e.g., TensorFlow) know
how to compute the functional expression for the Jacobians and gradients of
commonly used operators such as those that appear in Fig. 13.8. In that case,
the developer needs only to provide the compute graph and the framework
will add the new nodes for backpropagation. Figure 13.9 shows the resulting
compute graph with added gradient nodes.
g( W )
g( v )
g( y)
u
v
y
σ
L
MSE
y
( b)
g
g( u )
W
x
b
Figure 13.9: Compute graph with gradient nodes
Example 13.4 : We shall work out the functional expressions for the gradients
of all the nodes in Fig. 13.8. We already know g(y). So the next node we choose
to process is v.
g(v) = ∇v(L) = Jv(y)∇y(L) = Jv(y)g(y)
We know that y = σ(v). Since σ is an element-wise operator, the Jacobian
Jv(y) takes a particularly simple form. Using the derivative for the logistic
sigmoid function from Section 13.2.2, we see that
∂yi
∂vj
=
(
yi(1 −yi)
if i = j
0
otherwise
The Jacobian is therefore a diagonal matrix:
Jv(y) =
y1(1 −y1)
0
. . .
0
0
y2(1 −y2)
. . .
0
...
...
...
...
0
0
. . .
yn(1 −yn)
| text |
ch13 | page_022.txt | 530
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Suppose s = [s1, s2, . . . , sn] is a vector defined by si = yi(1 −yi) (i.e., the
diagonal of the Jacobian matrix). We can express g(v) simply as
g(v) = s ◦g(y)
where a ◦b is the vector resulting from the element-wise product of a and b.5
Now that we have g(v), we can compute g(b) and g(u). We have g(b) =
Jb(v)g(v) and g(u) = Ju(v)g(v). Since
v = u + b
it is straightforward to verify that
Jb(v) = Ju(v) = In
where In is the n × n identity matrix. So we have
g(b) = g(u) = g(v)
We finally come to the matrix W. Recall that u = Wx. There is a potential
problem here, because all the machinery we have set up works for vectors, while
W is a matrix. But recall from Section 13.2.1 that we assembled the matrix W
from a set of vectors w1, w2, . . . , wn, where wT
i is the ith row of W. The trick
is to consider each of these vectors separately and compute its gradient using
the usual formula.
g(wi) = Jwi(u)g(u)
We know that ui = wT
i x and none of the other uj’s have any dependency on
wi for i ̸= j. Therefore, the Jacobian Jwi(u) is zero everywhere except the ith
column, which is equal to x. Thus we have
g(wi) = g(ui)x
✷
Example 13.5 : As we mentioned, neural networks for classification often use
a softmax activation in the final layer followed by a cross-entropy loss. We will
now compute the gradient of the combined operator.
Suppose the input to the combined operator is y; let q = µ(y), and let l =
H(p, q), where p represents the true probability vector for the corresponding
training example. We have:
log(qi)
=
log(
eyi
P
j eyj )
=
yi −log(
X
j
eyj)
5This operation sometimes called the Hadamard product, so as not to confuse it with the
more usual dot product, which is the sum of the components of the Hadamard product.
| text |
ch13 | page_023.txt | 13.3. BACKPROPAGATION AND GRADIENT DESCENT
531
Therefore, noting that P
i pi = 1, we have:
l
=
H(p, q)
=
−
X
i
pi log qi
=
−
X
i
pi(yi −log(
X
j
eyj))
=
−
X
i
piyi −log(
X
j
eyj)
X
i
pi
=
−
X
i
piyi −log(
X
j
eyj)
Differentiating, we get:
∂l
∂yk
=
−pk +
eyk
P
j eyj
=
−pk + µ(yk)
=
qk −pk
Therefore, we end up with the rather neat result:
∇yl = q −p
This combined gradient does not saturate or explode, and leads to good learning
behavior. That observation explains why softmax and cross entropy loss work
so well together in practice.
✷
13.3.4
Iterating Gradient Descent
Given a set of training examples, we run the compute graph in both directions
for each example: forward (to compute the loss) and backwards (to compute
the gradients). We average the loss and gradients across the training set to
compute the average loss and the average gradient for each parameter vector.
At each iteration we update each parameter vector in the direction opposite
to its gradient, so the loss will tend to decrease. Suppose z is a parameter
vector. We set:
z ←z −ηg(z)
Here η is a hyperparameter, the learning rate. We stop gradient descent either
when the loss between successive iterations changes by only a tiny amount (i.e.,
we have reached a local minimum) or after a fixed number of iterations.
It is important to pick the learning rate carefully. Too small a value means
gradient descent might take a very large number of iterations to converge. Too
large a value might cause large oscillations in the parameter values and never
| text |
ch13 | page_024.txt | 532
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
lead to convergence. Usually picking the right learning rate is a matter of trial
and error. It is also possible and common to vary the learning rate. Start with
an initial learning rate η0. Then, at each iteration, multiply the learning rate
by a factor β (0 < β < 1) until the learning rate reaches a sufficiently low value.
When we have a large training set, we may not want to use the entire
training set for each iteration, as it might be too time-consuming. So for each
iteration we randomly sample a “minibatch” of training examples. This variant
is called stochastic gradient descent,” as was discussed in Section 12.3.5, since
we estimate the gradients using a different sample of the training set at each
iteration.
We have left open the question of how the parameter values are initialized
before we start gradient descent.
The usual approach is to choose them at
random. Popular approachaes include sampling each entry uniformly at ran-
dom in [−1, 1], or choosing randomly using a normal distribution. Notice that
initializing all the weights to the same value would cause all nodes in a layer
to behave the same way, and thus we would never reap the benefit of having
different nodes in a layer recognize different features of the input.
13.3.5
Tensors
Previously, we have imagined that the inputs to a neural net are one-dimensional
vectors. But there is no reason why we cannot view the input as having a higher
dimension.
Example 13.6 : A grey-scale photo might be represented by a two-dimensional
array of real numbers, corresponding to the intensity of each pixel. Each pixel
of a color image typically requires 3 dimensions. That is, each pixel itself is
a vector with three components, say for the intensity of the pixel in the red,
green, and blue colors. One useful way to view a color image as input to a
neural net is to think of each training example as a two-dimensional array of
pixels, where the value of each pixel is not a real number, as we have heretofore
imagined, but a vector with three dimensions, one for each of the three colors.
✷
Similarly, we have viewed each layer of a neural net as a column of nodes.
But there is no reason we cannot imagine the nodes in a layer to be organized
as a two-dimensional array or even an array of dimension greater than two.
Finally, we have viewed the input values as real numbers and similarly viewed
the values produced by each node as a real number. But we could also think of
the values attached to each input or output of a node as a vector or a higher-
dimensional structure. The natural generalization of vectors and matrices is
the tensor, which is an n-dimensional array of scalars.
Unfortunately, the backpropagation algorithm we described works for vec-
tors, not higher-dimensional tensors. In such cases, we resort to the same trick
we used in Section 13.3.3, where we unrolled the matrix W into collection of
| text |
ch13 | page_025.txt | 13.3. BACKPROPAGATION AND GRADIENT DESCENT
533
vectors. Just as we regard an m × n matrix as a set of m n-vectors, we can re-
gard a 3-dimensional tensor of dimensionality l×m×n as a set of lm n-vectors,
and similarly for tensors of higher dimension.
Example 13.7 : This example is based on the MNIST dataset.6 This dataset
consists of 28 × 28 monochrome images, each represented by a two-dimensional
square bit array whose sides are of length 28. Our goal is to build a neural net
that determines whether an image corresponds to a handwritten digit (0-9) and
if so which one. Consider a single image X, which is a 28 × 28 matrix. Suppose
the first layer of our network is a dense layer7 consisting of 49 hidden nodes,
which we shall imagine is arranged in a 7 × 7 array. We model the hidden layer
as a 7 × 7 matrix H, where the output of the node in row i and column j is hij.
We can model the weights for each of the 28 × 28 inputs and each of the
7×7 nodes as a weight tensor W with dimensions 7×7×28×28. That is, Wijkl
represents the weight for the input pixel in row i and column j of the image to
the node whose position in the array of nodes is row k and column l. Then:
hij =
28
X
k=1
28
X
l=1
wijklxkl for 1 ≤i, j ≤7
where we omit the bias term for simplicity (i.e., we assume all thresholds of all
nodes are 0.
An equivalent way to think about this structure to flatten the input X into
a vector x of length 784 (since 28 × 28 = 784) and the hidden layer H into a
vector h of length 49. We flatten the weight tensor W as follows: the last two
dimensions are flattened into a single dimension to match x, and its first two
dimensions are flattened into a single dimension to match the hidden vector h,
resulting in a 49 × 784 weight matrix. Suppose as in Section 13.2.1, we have
wT
i denote the ith row of this new weight matrix. We can now write:
hi = wT
i x for 1 ≤i ≤49
It’s straightforward to see that there is a 1-to-1 mapping between the hidden
nodes in the old and new arrangements. Moreover, the output of each hidden
node is determined by a dot product, just as in Section 13.2.1. Thus the tensor
notation is just a convenient way to group vectors.
✷
Thus the tensors used in neural networks have little in common with the
tensors used in Physics and other mathematical sciences. A tensor in our con-
text is just a nested collection of vectors. The only tensor operation we shall
need is the flattening of a tensor by merging dimensions as in Example 13.7. We
can use the backpropagation algorithm described in Section 13.3.3 for tensors
once we have flattened them approriately.
6See yann.lecun.com/exdb/mnist/.
7In reality, the first network layer for this problem is likely to be a convolutional layer.
See Section 13.4
| text |
ch13 | page_026.txt | 534
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
13.3.6
Exercises for Section 13.3
Exercise 13.3.1: This exercise uses the neural net from Fig. 13.7. However,
assume that the weights on all inputs are variables rather than the constants
shown. Note, however, that some inputs do not feed one of the nodes in the
first layer, so these weights are fixed at 0. Assume that the input vector is x,
the output vector is y, and the output of the two nodes in the hidden layer is
the vector z. Also, let the matrix and bias vector connecting x to z be W1 and
b1, while the matrix and bias vector connecting z to y are W2 and b2. Assume
that the activation function at the hidden layer is the hyperbolic tangent, and
the activation function at the output is the identity function (i.e., no change to
the outputs is made). Finally, assume the loss function is mean-squared error,
where ˆy is the true output vector for a given input vector x. Draw the compute
graph that shows how y is computed from x.
Exercise 13.3.2: For the network described in Exercise 13.3.1:
(a) What is Jy(z)?
(b) What is Jz(x)?
(c) Express g(x) in terms of g(tanh (z)).
(d) Express g(x) in terms of the loss function.
(e) Draw the compute graph with gradient computation for the entire net-
work.
13.4
Convolutional Neural Networks
Consider a fully-connected network layer for processing 224 × 224 images, with
each pixel encoded using 3 color values (often called channels – the values of
intensity for red, green, and blue)).8 The number of weight parameters in the
connection layer for each output node is 224 × 224 × 3 = 150, 528. Suppose
we had 224 nodes in the output layer. We would end up with a total of over
33 million parameters! Given that our training set of images is usually of the
order of tens or hundreds of thousands of images, the huge number of parameters
would quickly lead to overfitting even using just a single layer.
A Convolutional Neural Network (CNN) greatly reduces the number of pa-
rameters by taking advantage of the properties of image data. CNN’s introduce
two new varieties of network layers: convolutional layers and pooling layers.
8This is the size of most of the images on ImageNet, for example.
| text |
ch13 | page_027.txt | 13.4. CONVOLUTIONAL NEURAL NETWORKS
535
13.4.1
Convolutional Layers
Convolutional layers make use of the fact that image features often are described
by small contiguous areas in the image. For example, at the first convolutional
layer, we might recognize small sections of edges in the image. At later layers,
more complex structures, such as features that look like flower petals or eyes
might be recognized. The idea that simplifies the calculation is the fact that
the recognition of features such as edges does not depend on where in the image
the edge is. Thus, we can train a single node to recognize a small section of
edge, say an edge through a 5-by-5 region of pixels. This idea benefits us in
two ways.
1. The node in question needs only inputs from 25 pixels corresponding to
any 5-by-5 square, not inputs from all 224-by-224 pixels. That saves us a
lot in the representation of the trained CNN.
2. The number of weights that we need to learn in the training process is
greatly reduced. For each node in the layer, we require only one weight
for each input to that node – say 75 weights if a pixel is represented by
RGB values, not the 150,528 weights that we suggested above would be
needed for an ordinary, fully connected layer.
We shall think of the nodes in a convolutional layer as filters for learning
features. A filter examines a small spatially contiguous area of the image –
traditionally, a small square area such as a 5 × 5 square of pixels. Moreover,
since many features of interest may occur anywhere in the input image (and
possibly in more than one location), we apply the same filter at many locations
on the input.
To keep things simple, suppose the input consists of monochromatic images,
that is, grey-scale images whose pixels are each a single value. Each image is
thus encoded by a 2-dimensional pixel array of size 224 × 224. A 5 × 5 filter F
is encoded by a 5 × 5 weight matrix W and single bias parameter b. When the
filter is applied on a similarly-sized square region of input image X with the
top left corner of the filter aligned with the image pixel xij, the response of the
filter at this position, denoted rij, is given by:
rij =
4
X
k=0
4
X
l=0
xi+k,j+lwkl + b
(13.1)
We now slide the filter along the length and width of the input image, applying
the filter at each position, so that we capture every possible 5 × 5 square region
of pixels in the image. Notice that we can apply the filter at input locations
1 ≤i ≤220 and 1 ≤j ≤220, although it does not “fit” at positions with a
higher i or j. The resulting set of responses rij are then passed through an
activation function to form the activation map R of the filter. In most cases,
the activation function is ReLU or one of its variants. When trained, i.e., the
| text |
ch13 | page_028.txt | 536
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
weights wij of the filter are determined, the filter will recognize some feature of
the image, and the activation map tells whether (or to what degree) this feaure
is present at each position of the image.
Example 13.8 : In Fig. 13.10(b), we see a 2 × 2 filter, which is to be applied
to the 4 × 4 image in Fig. 13.10(b). To do so, we lay the filter over all nine of
the 2 × 2 squares of the image. In the figure, we suggest the filter being placed
over the 2 × 2 square in the upper-right corner. After overlaying the filter, we
multiply each of the filter elements by the corresponding image element and
then take the sum of the products. In principle, we then need to add in a bias
term, but in this example, we shall assume the bias is 0.
1 0 1 0
0 1 0 1
1 1 0 0
0 0 0 1
1 0
0 −1
0 0 0
(b) 2−by−2 filter
1 1 −1
(a) 4−by−4 image
(c) 3−by−3 response
−1 1 0
Figure 13.10: Applying a filter to an image
Another way to look at this process is that we turn the filter into a vector
by concatenating its rows, in order, and we do the same to the square of the
image. Then, we take the dot product of the vectors. For instance, the filter
can be thought of as the vector [1, 0, 0, −1], and the square in the upper-left
corner of the image can be thought of as the vector [1, 0, 0, 1]. The dot product
of these vectors is 1 × 1 + 0 × 0 + 0 × 0 + (−1) × 1 = 0. Thus, the result, shown
in Fig. 13.10(c), has a 0 for its upper-left entry.
For another example, if we slide the filter down one row, the dot product of
the filter as a vector and the vector formed from the first two elements of the
second and third rows of the image is 1 × 0 + 0 × 1 + 0 × 1 + (−1) × 1 = −1.
Thus, the first element of the second row of the convolution is −1.
✷
When we deal with color images, the input has three channels. That is, each
pixel is represented by three values, one for each color. Suppose we have a color
image of size 224 × 224 × 3. The filter’s output will also have three channels,
and so the filter is now encoded by a 5 × 5 × 3 weight matrix W and single
bias parameter b. The activation map R still remains 5 × 5, with each response
given by:
rij =
4
X
k=0
4
X
l=0
3
X
d=1
xi+k,j+l,dwkld + b
(13.2)
In our example, we imagined a filter of size 5. In general, the size of the filter is
a hyperparameter of the convolutional layer. Filters of size 3, 4, or 5 are most
| text |
ch13 | page_029.txt | 13.4. CONVOLUTIONAL NEURAL NETWORKS
537
commonly used. Note that the filter size specifies only the width and height of
the filter; the number of channels of the filter always matches the number of
channels of the input.
The activation map in our example is slightly smaller than the input. In
many cases, it is convenient to have the activation map be of the same size as the
input. We can expand the repsonse by using zero padding: adding additional
rows and columns of zeros to pad out the input. A zero padding of p corresponds
to adding p rows of zeros each to the top and bottom, and p columns to the left
and right, increasing the dimensionality of the input by 2p along both width
and height. A zero padding of 2 in our example would augment the input size
to 228 × 228 and result in an activation map of size 224 × 224, the same size as
the original input image.
The third hyperparameter of interest is stride. In our example, we assumed
that we apply the filter at every possible point in the input image. We could
think instead of sliding the filter one pixel at a time along the width and height
of the input, corresponding to a stride s = 1.
Instead, we could slide the
filter along the width and the height of image two or three pixels at a time,
corresponding to a stride s of 2 or 3. The larger the stride, the smaller the
activation map compared to the input.
Suppose the input is an m× m square of pixels, the output an n× n square,
filter size is f, stride is s, and zero padding is p. It is easily seen that:
n = (m −f + 2p)/s + 1
(13.3)
In particular, we must be careful to pick hyperparameters such that s evenly
divides m −f + 2p; else we would have an invalid arrangement for the convo-
lutional layer, and most software implementations would throw an exception.
We can intuitively think of a filter as looking for an image feature, such as
a splotch of color or an edge. Classifying an image usually requires identifying
many features. We therefore use many filters, ideally one for each useful feature.
During the training of the CNN, we hope that each filter will learn to identify
one of these features.
Suppose we use k filters; to keep things simple, we
constrain all filters to use same size, stride, and zero padding. Then the output
contains k activation maps. The dimensionality of the output layer is therefore
n × n × k, where n is given by Equation 13.3.
The set of k filters together constitute a convolutional layer. Given input
with d channels, a filter of size f requires df 2+1 parameters (df 2 weight param-
eters and 1 bias parameter). Therefore, a convolutional layer of k such filters
uses k(df 2 + 1) parameters.
Example 13.9 : Continuing the ImageNet example, suppose the input consists
of 224×224×3 images, and we use a convolutional layer of 64 filters, each of size
5, stride 1, and zero padding 2. The size of the output layer is 224 × 224 × 64.
Each filter needs 3×5×5+1 = 76 parameters (including one for the threshold)
and the convolutional layer contains 64 × 76 = 4864 parameters – orders of
magnitude smaller than the number of parameters for a fully connected layer
with the same input and output sizes.
✷
| text |
ch13 | page_030.txt | 538
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
13.4.2
Convolution and Cross-Correlation
This subsection is a short detour to explain why Convolutional Neural Networks
are so named. It is not a pre-requisite for any of the other material in this
chapter.
The convolutional layer is named because of the resemblance it bears to the
convolution operation from functional analysis, which is often used in signal
processing and probability theory. Given functions f and g, usually defined
over the time domain, their convolution (f ∗g)(t) is defined as:
(f ∗g)(t) =
Z ∞
−∞
f(τ)g(t −τ)dτ =
Z ∞
−∞
f(t −τ)g(τ)dτ
Here we are interested in the discrete version of convolution, where f and g are
defined over the integers:
(f ∗g)(i) =
∞
X
k=−∞
f(k)g(i −k) =
∞
X
k=−∞
f(i −k)g(k)
Often convolution is viewed as using function g to transform function f. In
this context, the function g is sometimes called the kernel. When the kernel is
finite, so g(k) is only defined for k = 0, 1, . . ., m−1, the definition simplifies to:
(f ∗g)(i) =
m−1
X
k=0
f(i −k)g(k)
We can extend the definition to 2-dimensional functions:
(f ∗g)(i, j) =
m−1
X
k=0
m−1
X
l=0
f(i −k, j −l)g(k, l)
Let us define h to be the kernel obtained by flipping g, i.e., h(i, j) = g(−i, −j)
for i, j ∈{0, . . . , m −1}. It can verified that the convolution f ∗h of f with the
flipped kernel h is given by:
(f ∗h)(i, j) =
m−1
X
k=0
m−1
X
l=0
f(i + k, j + l)g(k, l)
(13.4)
Note the similarity of Equation 13.4 to Equation 13.1, ignoring the bias term b.
The operation of the convolutional layer can be thought of as the convolution of
the input with a flipped kernel. This similarity is the reason why convolutional
layers are so named, and filters are sometimes called kernels.
The cross-correlation f ⋆g is defined by (f ⋆g)(x, y) = (f ∗h)(x, y) where
h is the flipped version of g. Thus the operation of the convolutional layer can
also be seen as the cross-correlation of the input with the filter.
| text |
ch13 | page_031.txt | 13.4. CONVOLUTIONAL NEURAL NETWORKS
539
13.4.3
Pooling Layers
A pooling layer takes as input the output of a convolutional layer and produces
an output with smaller spatial extent. The size reduction is accomplished by
using a pooling function to compute aggregates over small contiguous regions
of the input. For example we might use max pooling over nonoverlapping 2 × 2
regions of the input; in this case there is an output node corresponding to
every nonoverlapping 2 × 2 region of the input, with the output value being the
maximum value among the 4 inputs in the region. The aggregation operates
independently on each channel of the input. The resulting output layer is 25%
the size of the input layer. There are three elements in defining a pooling layer:
1. The pooling function, which is most commonly the max function but could
in theory be any aggregate function, such as average.
2. The size f of each pool, which specifies that each pool uses an f ×f square
of inputs.
3. The stride s between pools, analogous to the stride used in the convolu-
tional layer.
The most common use cases in practice are f = 2 and s = 2, which specifies
nonoverlapping 2 × 2 regions, and f = 3, s = 2, which specifies 3 × 3 regions
with some overlap. Higher values of f lead to too much loss of information in
practice. Note that the pooling operation shrinks the height and width of the
input layer, but preserves the number of channels. It operates independently
on each channel of its input. Note that unlike the convolution layer, it is not
common practice to use zero padding for the max pooling layer.
Pooling is appropriate if we we believe that features are approximately in-
variant to small translations. For example, we might care about the relative
locations of features such as legs or wings and not their exact locations. In
such cases pooling can greatly reduce the size of the hidden layer that forms
the input to the subsequent layers of the network.
Example 13.10 : Suppose we apply max pooling with size = 2 and stride =
2 to the 224 × 224 × 64 output of the convolutional layer from Example 13.9.
The resulting output is of size 112 × 112 × 64.
✷
13.4.4
CNN Architecture
Now that we have seen the building blocks of Convolutional Neural Networks,
we can put them together to build deep networks. A typical CNN alternates
convolutional and pooling layers, ending with one or more fully-connected layers
that produce the final output.
Example 13.11 : For instance, Fig. 13.11 is a simple network architecture for
classifying ImageNet images into one of 1000 image classes, loosely inspired by
| text |
ch13 | page_032.txt | 540
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
VGGnet.9 This simple network strictly alternates convolutional and pooling
layers. In practice, high-performing networks are finely tuned to their task,
and may stack convolutional layers directly on top of one another with the
occasional pooling layer in between. Moreover, there is often more than one
fully-connected layer before the final output. The input to the first layer is the
224-by-224, 3-channel image. The input to each subsequent layer is the output
of the previous layer.
Layer Type
Size
Stride
Pad
Filter Count
Output Size
Convolution
3
1
1
64
224 × 224 × 64
Max Pool
2
2
0
64
112 × 112 × 64
Convolution
3
1
1
128
112 × 112 × 128
Max Pool
2
2
0
128
56 × 56 × 128
Convolution
3
1
1
256
56 × 56 × 256
Max Pool
2
2
0
256
28 × 28 × 256
Convolution
3
1
1
512
14 × 14 × 512
Max Pool
2
2
0
512
14 × 14 × 512
Convolution
3
1
1
1024
14 × 14 × 1024
Max Pool
2
2
0
1024
7 × 7 × 1024
Fully Connected
1 × 1 × 1000
Figure 13.11: Layers of a Convolutional Neural Network
The first layer is a convolutional layer, consisting of 64 filters, each with
three channels, as would be the case for any color-image processor. The filters
are 3-by-3, and the stride is 1, so every 3-by-3 square of the image is an input
to the filter. There is one unit of zero-padding, so the number of outputs of
this layer equals the number of inputs. Further, notice that we can view the
output as another 224-by-224 array. Each element of this array consists of 64
filters, each of which is a 3-channel pixel.
The output of the first layer is fed to a max-pool layer, in which we divide
the 224-by-224 array into 2-by-2 squares (because both the size and stride are
2). Thus, the 224-by-224 array has become a 112-by-112 array, and there are
still the same 64 filters.
At the third layer, which is again convolutional, we take the 112-by-112
array of pixels from the second layer as input.
This layer has more filters
than the first layer – 128 to be exact. The intuitive reason for the increase is
that the first layer recognizes very simple structures, like edges, and there are
not too many different simple structures. However, the third layer should be
recognizing somewhat more complex features, and these features can involve a
6-by-6 square of the input image, because of the pooling done at the second
layer.
Similarly, each subsequent convolutional layer takes inputs from the
9K. Simonyan and A. Zussman,“Very Deep Convolutional Networks for Large-Scale Image
Recognition,” arXiv:1409–1556v6.
| text |
ch13 | page_033.txt | 13.4. CONVOLUTIONAL NEURAL NETWORKS
541
How Many Nodes in a Convolutional Layer?
We have referred to a node of a convolutional layer as a “filter.” That
filter may be a single node, or as in Example 13.11, a set of several nodes,
one for each channel. When we train the CNN, we determine weights for
each filter, so there are relatively few nodes. For instance, in the first layer
in Example 13.11, there are 192 nodes, three for each of the 64 filters.
However, when we apply the trained CNN to an input, we apply each
filter to every pixel in the input. Thus, in Example 13.11, each of the 64
filters of the first layer is applied to 224 × 224 = 50, 176 pixels. The key
point to remember is that although a CNN has a succinct representation,
the application of the represented neural net to data requires a significant
amount of computation. The same, by the way, can be said for the other
specialized forms of neural net that we discuss in Section 13.5.
previous pooling layer and its filters can represent structures of progressively
larger sizes and complexities. Thus, the number of filters has been chosen to
double at each convolution layer.
Finally, the eleventh, and last, layer is a fully connected layer. It has 1000
nodes, corresponding to the 1000 images classes we are trying to learn how to
recognize. Being fully connected, each of these 1000 nodes takes all 7 × 7 × 3 =
147 outputs from the 10th layer; the factor 3 is from the fact that all filters of
the previous layers have three channels.
✷
Designing CNN’s and other deep network architectures is still more art than
science. Over the past few years, however, some rules of thumb have emerged
that are worth keeping in mind:
1. Deep networks that use many convolutional layers, each using many small
filters, are better than shallow networks that use large filters.
2. A simple pattern is to use convolutional layers that preserve the spatial
extent of their input by zero padding, and have size reduction done ex-
clusively by pooling layers.
3. Smaller strides work better in practice than larger strides.
4. It’s very useful to have the input size evenly divisible by 2 many times.
13.4.5
Implementation and Training
An examination of Equation 13.1 (generalized so the filter is f × f rather than
5-by-5) suggests that we can express each entry in the output of the convolution
as the dot product of vectors, followed by a scalar sum. To do so, we must flatten
| text |
ch13 | page_034.txt | 542
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
the filter F and the corresponding region in the input into vectors. Consider
the convolution of an m×m×1 tensor X (i.e., X is actually an m×m matrix)
with an f × f filter F and bias b, to produce as output the n × n matrix Z.
We now explain how to implement the convolution operation using a single
vector-matrix multiplication.
We first flatten the filter F into a f 2 × 1 vector g. We then create matrix
Y from X as follows: each square f × f region of X is flattened into a f 2 × 1
vector, and all these vectors are lined up as columns to form a single f 2 × n2
matrix Y . Construct the n2 × 1 vector b so that all its entries are equal to the
bias b. Then
z = Y Tg + b
yields a n2 × 1 vector z. Moreover, each element of z is a single element in the
convolution. Therefore, we can rearrange the entries in z into an n × n matrix
Z that gives the output of the convolution.
Notice that the matrix Y is larger than the input X (approximately by a
factor of f 2), because each entry of X is repeated many times in Y . Thus,
this implementation uses a lot of memory. However, multiplying a matrix and
a vector is extremely fast on modern hardware such as Graphics Processing
Units (GPU’s), and so it is the method used in practice.
This approach to computing convolutions can easily be extended to the case
of inputs with more than one channel. Moreover, we can also handle the case
where we have k filters rather than just 1. We then need to replace the vector
g with a df 2 × k matrix G and use a larger matrix Y (df 2 × n2). We also need
to use an n2 × k bias matrix B, where each column repeats the bias term of
the corresponding filter. Finally, the output of the convolution is expressed by
an n2 × k matrix C, with a column for the output of each filter, where:
C = Y TF + B
We have explained how to perform the forward pass through the convolu-
tional layer. During training, we shall need to backpropagate through the layer.
Since each entry in the output of convolution is a dot product of vectors followed
by a sum, we can use the techniques from Section 13.3.3 to compute deriva-
tives. It turns out that the derivative of a convolution can also be expressed as
a convolution, but we shall not go into the details here.
13.4.6
Exercises for Section 13.4
Exercise 13.4.1: Suppose images are 512-by-512, and we use a filter that is
3-by-3.
(a) How many responses will be computed for this layer of a CNN?
(b) How much zero padding is necessary to produce an output of size equal
to the input?
| text |
ch13 | page_035.txt | 13.5. RECURRENT NEURAL NETWORKS
543
(c) Suppose we do not do any zero padding. If the output of one layer is
input to the next layer, after how many layers will there be no output at
all?
Exercise 13.4.2: Repeat Exercise 13.4.1(a) and (c) for the case when there is
a stride of three.
Exercise 13.4.3: Suppose we have the output of an m-by-m convolutional
layer with k filters, each having d channels. These outputs are fed to a pooling
layer with size f and stride s. How many output values does the pooling layer
produce?
Exercise 13.4.4: For this exercise, assume that inputs are single bits 0 (white)
and 1 (black). Consider a 3-by-3 filter, whose weights are wij, for 0 ≤i ≤2 and
0 ≤j ≤2, and whose bias is b. Suggest wieghts and bias so that the output of
this filter would detect the following simple features:
(a) A vertical boundary, where the left column is 0, and the other two columns
are 1.
(b) A diagonal boundary, where only the triangle of three pixels in the upper
right corner are 1.
(c) a corner, in which the 2-by-2 square in the lower right is 0 and the other
pixels are 1.
13.5
Recurrent Neural Networks
Just as CNN’s are a specialized family of neural networks for processing 2-
dimensional image data, recurrent neural networks (RNN’s) are networks spe-
cially designed for processing sequential data. Sequential data naturally arises
in many settings: a sentence is a sequence of words; a video is a sequence of
images; a stock market ticker is a sequence of prices.
Consider a simple example from the field of language processing. Each input
is a sentence, modeled as a sequence of words. After processing each prefix of
the sequence, we would like to predict the next word in the sentence; the output
at each step is a probability vector across words. Our example suggests two
key design considerations:
1. The output at each point depends on the entire prefix of the sentence
until that point, and not just the last word. The network needs to retain
some “memory” of the past.
2. The underlying language model does not change across positions in the
sequence, so we should use the same parameters (weights for each of the
nodes) at each position.
| text |
ch13 | page_036.txt | 544
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
s
x
y
V
U
s
x
y
V
U
s
x
y
s
s
x
y
V
U
W
n
n
n
W
V
U
W
W
W
(b) The unrolled RNN of length
(a) The basic unit of an RNN.
n.
o o o
0
1
1
1
2
2
2
Figure 13.12: RNN architecture
These considerations lead naturally to a recurrent network model, where we
perform the same operation at each step, with the input to each step being
dependent upon the output from prior steps. Figure 13.12 shows the structure
of a typical recurrent neural network. The input is a sequence x1, x2, . . . , xn,
and the output is also a sequence y1, y2, . . . , yn. In our example, each input
xi represents a word, and each output yi is a probability vector for the next
word in the seentence. The input xi is typically encoded as a 1-hot vector –
a vector of length equal to the number of possible words,10 with a 1 in the
position corresponding to the input word and a 0 in all other positions. There
are two important difference between a general neural network and the RNN:
1. The RNN has inputs at all (or almost all) layers, and not just at the first
layer.
2. The weights at each of the first n layers are constrained to be the same;
these weights are the matrices U and W in Equation 13.5 below. Thus,
each of the first n layers has the same set of nodes, and correspond-
ing nodes from each of the layers share weights (and are thus really the
same node), just as nodes of a CNN representing different locations share
weights and are thus really the same node.
At each step t, we have a hidden state vector st that functions as the memory
in which the network encodes information about the prefix of the sequence it
has seen. The hidden state at time t is a function of the input at time t and
the hidden state at time t −1:
st = f(Uxt + Wst−1 + b)
(13.5)
10Since there could in principle be an infinite number of words, we might in practice devote
components of the vector only to the most common words or the words that are most impor-
tant in the application at hand. Other words would all be represented by a single additional
component of the vector.
| text |
ch13 | page_037.txt | 13.5. RECURRENT NEURAL NETWORKS
545
Here f is a nonlinear activation function such as tanh or sigmoid. U and W are
matrices of weights, and b is a vector of biases. We define s0 to be a vector of
all zeros. The output at time t is a function of the hidden state at time t, after
being transformed by a parameter matrix V and an activation function g:
yt = g(V st + c)
In our example, g might be the softmax function to ensure that the output is a
valid probability vector.
The RNN in Figure 13.12 has an output at each time step. In some appli-
cations, such as machine translation, we need just a single output at the end of
each sentence. In such cases, the RNN’s single output is further processed by
one or more fully-connected layers to generate the final output.
It is simplest to assume that our RNN’s inputs are fixed-length sequences
of length n. In this case, we simply unroll the RNN to contain n time steps.
In practice, many applications have to deal with variable-length sequences e.g.,
sentences of varying lengths. There are two approaches to deal with this situa-
tion:
1. Zero-padding. Fix n to be the longest sequence we process, and pad out
shorter sequences to be length n.
2. Bucketing. Group sequences according to length, and build a separate
RNN for each length.
A combination of these two approaches is used. We can create a bucket for a
small number of different lengths, and assign a sequence to the bucket of the
shortest length that is at least as long as the sequence. Then, within a bucket
we use padding for sequences that are shorter than the maximum length for
that bucket.
13.5.1
Training RNN’s
We use backpropagation to train an RNN, just as we would any neural network.
Let us work through an example. Suppose our input consists of sequences of
length n. Our network uses the activation function tanh for the state update,
softmax for the output, and the loss function is cross-entropy. Since the network
has several outputs, one at each time-step, we seek to minimize the total error
e, defined as the sum of the losses ei at each time step i:
e =
n
X
i=1
ei
To simplify notation, we use the following conventions. Suppose x and y are
vectors, and z is a scalar. We define:
dz
dx
=
∇xz
dy
dx
=
Jx(y)
| text |
ch13 | page_038.txt | 546
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Moreover, suppose W is a matrix, and w is the vector obtained by concatenating
the rows of W. Then:
dz
dW
=
dz
dw
dy
dW
=
dy
dw
These conventions also extend naturally to partial derivatives.
We use backpropagation to compute the gradients of the error with respect
to the network parameters. We focus on
de
dW ; the gradients for U and V are
similar, and left as exercises for the reader. It is clear that:
de
dW =
n
X
t=1
det
dW
Focusing on step t, we have:
det
dW = dst
dW
det
dst
We leave it as an exercise to verify that:
det
dst
= V T(yt −ˆyt)
(13.6)
Setting Rt = dst
dW , we note that st = tanh(zt), where zt = Wst−1 + Uxt + b,
we have:
Rt = dst
dzt
dzt
dW
It is straightforward to verify that dst
dzt is the diagonal matrix A defined by:
aij =
(
1 −s2
ti
when i = j
0
otherwise
We should to be careful to note that zt is a function of W both directly and
indirectly, since st−1 also depends on W. So we must express dzt
dW as a sum of
two terms:
dzt
dW = ∂zt
∂W +
∂zt
∂st−1
dst−1
dW
It is easily verified that:
∂zt
∂st−1
= W T
The form of ∂zt
∂W is a little trickier. It is a matrix B with mostly zero entries,
the nonzero elements being entries from st−1. We leave the computation of B
| text |
ch13 | page_039.txt | 13.5. RECURRENT NEURAL NETWORKS
547
as an exercise for the reader. Now, noting that dst−1
dW
is just Rt−1, we have the
recurrence:
Rt = A(B + W TRt−1)
Setting Pt = AB and Qt = AW T, we end up with:
Rt = Pt + QtRt−1
(13.7)
We can use this recurrence to set up an iterative evaluation of Rt, and thence
de
dW . We initialize the iteration by setting R0 to the matrix with all zeros. This
iterative method for computing the gradients of an RNN is called Backpropa-
gation Through Time (BPTT), since it reveals the effects of earlier time-steps
on later time-steps.
13.5.2
Vanishing and Exploding Gradients
RNN’s are a simple and appealing model for learning from sequences, and
the BPTT algorithm is straightforward to implement. Unfortunately, RNN’s
have a fatal flaw that limits their use in many practical applications. They
are effective only at learning short-term connections between nearby elements
in the sequence and ineffective at learning long-distance connections. Long-
distance connections are crucial in many applications; for example, there can
be an arbitrary number of words or clauses separating a verb or a pronoun from
the subject with which it is associated.
In order to understand the cause of this limitation, let us unroll Equa-
tion 13.7:
Rt
=
Pt + QtRt−1
=
Pt + Qt(Pt−1 + Qt−1Rt−2)
=
Pt + QtPt−1 + QtQt−1Rt−2
. . .
Ultimately yielding:
Rt = Pt +
t−1
X
j=0
Pj
tY
k=j+1
Qk
(13.8)
From Equation 13.8, it is clear that the contribution of step i to Rt is given by:
Ri
t = Pi
tY
k=i+1
Qk
(13.9)
Equation 13.9 includes the product of several matrices that look like diagonal
matrix A. Each entry in A is strictly less than 1. Just as the product of many
numbers, each strictly less than 1, approaches zero as we add more multipli-
cands, the term Qt
k=i+1 Qk approaches zero for i ≪t. In other words, the
gradient at step t is determined entirely by the preceding few time steps, with
| text |
ch13 | page_040.txt | 548
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
very little contribution from much earlier time steps. This phenomenon is called
the problem of vanishing gradients.
Equation 13.9 results in vanishing gradients because we used the tanh ac-
tivation function for state update. If instead we use other activation functions
such as ReLU, we end up with the product of many matrices with large entries,
resulting in the problem of exploding gradients. Exploding gradients are easier
to handle than vanishing gradients, because we can clip the gradient at each
step to lie within a fixed range. However, the resulting RNN’s still have trouble
learning long-distance associations.
13.5.3
Long Short-Term Memory (LSTM)
The LSTM model is a refinement of the basic RNN model to address the prob-
lem of learning long-distance associations. In the past few years, LSTM has
become popular as the de-facto sequence-learning model, and has been used
with success in many applications. Let us understand the intuition behind the
LSTM model before we describe it formally. The main elements of the LSTM
model are:
1. The ability to forget information by purging it from memory. For example,
when analyzing a text we might want to discard information about a
sentence when it ends. Or when analyzing a the sequence of frames in a
movie, we might want to forget about the location of a scene when the
next scene begins.
2. The ability to save selected information into memory. For example, when
we process product reviews, we might want to save only words expressing
opinions (e.g., excellent, terrible) and ignore other words.
3. The ability to focus only on the aspects of memory that are immediately
relevant. For example, focus only on information about the characters of
the current movie scene, or only on the subject of the sentence currently
being analyzed.
We can implement this focus by using a 2-tier archi-
tecture: a long-term memory that retains information about the entire
processed prefix of the sequence, and a working memory that is restricted
to the items of immediate relevance.
The RNN model has a single hidden state vector st at time t.
The LSTM
model adds an additional state vector ct, called the cell state, for each time t.
Intuitively, the hidden state corresponds to working memory and the cell state
corresponds to long-term memory. Both state vectors are of the same length,
and both have entries in the range [−1, 1]. We may imagine the working memory
having most of its entries near zero with only the relevant entries turned “on.”
The architectural ingredient that enables the ability to forget, save, and
focus is the gate. A gate g is just a vector of the same length as a state vector
| text |
ch13 | page_041.txt | 13.5. RECURRENT NEURAL NETWORKS
549
s; each of the gate’s entries is between 0 and 1. The Hadamard product11 s ◦g
allows us to selectively pass through certain parts of the state while filtering out
others. Usually, a gate vector is created by a linear combination of the hidden
state and the current input. We then apply a sigmoid function to “squash” its
entries to lie between 0 and 1. In general, an LSTM may use several different
kinds of gate vectors, each for a different purpose. At time t, we can create a
gate g as follows:
g = σ(Wst−1 + Uxt + b)
Here W and U are weight matrices and b is a bias vector.
At time t, we first compute a candidate state update vector ht based on the
previous hidden state and the current input:
ht = tanh(Whst−1 + Uhxt + bh)
(13.10)
Note that W, U, and b with subscript h are two weight matrices and a bias
vector that we learn and use for just the purpose of computing ht for each t.
We also compute two gates, the forget gate ft and the input gate it. The
forget gate determines which aspects of the long-term memory we retain. The
input gate determines which parts of the candidate state update to save into the
long-term memory. These gates are computed using different weight matrices
and bias vectors, which also must be learned. We indicate these matrices and
vector with subscripts f and i, respectively.
ft
=
σ(Wfst−1 + Ufxt + bf)
(13.11)
it
=
σ(Wist−1 + Uixt + bi)
(13.12)
We update the long-term memory using the gates and the candidate update
vector as follows:12
ct = ct−1 ◦ft + ht ◦it
(13.13)
Now that we have updated the long-term memory, we need to update the work-
ing memory. We do this in two steps. The first step is to create an output gate
ot. The second step is to apply this gate to the long-term memory, followed by
a tanh activation:13
ot
=
σ(Wost−1 + Uoxt + bo)
(13.14)
st
=
tanh(ct ◦ot)
(13.15)
11The Hadamard product of vectors [x1, x2, . . . , xn] and [yq, y2, . . . , yn] is the vector whose
components are the products of the corresponding components of the two argument vectors,
that is, [x1y1, x2y2, . . . , xnyn]. The same operation may be applied to any matrices that have
the same dimensions.
12Technically, an entry of 1 in the forget gate results in retaining the corresponding memory
entry; so the forget gate should really be called the remember gate. Similarly, the input gate
might be better named the save gate. Here we follow the naming convention commonly used
in the literature.
13Once again, the output gate might be better named the focus gate since it focuses the
working memory on certain aspects of the long-term memory.
| text |
ch13 | page_042.txt | 550
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
Here, we use subscript o to indicate another pairs of weight matrices and a bias
vector that must be learned.
Finally, the output at time t is computed in exactly the same manner as the
RNN output:
yt = g(V st + d)
(13.16)
where g is an activation function, V is a weight matrix and d is a bias vector.
Equations 13.10 through 13.16 describe the state update operations at a
single time step t of an LSTM. We can think of plain RNN as a special case
of LSTM. When we set the forget gate to all 0’s (so we throw away all prior
long-term memory) and the input gate to all 1’s (save the entire candidate state
update), and the output gate to all 1’s (working memory is same as long-term
memory), we get something that looks very close to an RNN, the only difference
being an extra tanh factor.
The ability to selectively forget allows LSTM’s to avoid the vanishing-
gradient problem at the expense of introducing many more parameters than
vanilla RNN’s. While we will not provide a rigorous proof here, we note that the
key to avoiding vanishing gradients is the long-term memory update in Equa-
tion 13.13. Several variations of the basic LSTM model have been proposed;
the most common variant is the gated recurrent Unit (GRU) model, which uses
a single state vector instead of two state vectors (long-term and short-term).
A GRU has fewer parameters than an LSTM and might be suitable in some
situations with smaller data sets.
13.5.4
Exercises for Section 13.5
Exercise 13.5.1: In this exercise, you are asked to design the input weights
for one or more nodes of the hidden state of an RNN. The input is a sequence
of bits, 0 or 1 only.14 Note that you can use other nodes to help with the node
requested. Also note that you can apply a transformation to the output of the
node so a “yes” answer has one value and a “no” answer has another.
(a) A node to signal when the input is 1 and the previous input is 0.
! (b) A node to signal when the last three inputs have all been 1.
!! (c) A node to signal when the input is the same as the previous input.
! Exercise 13.5.2: Verify Equation 13.6.
! Exercise 13.5.3: Give the formulas for the gradients de
dU and
de
dV for the gen-
eral RNN of Fig.13.12.
14We should understand that RNN’s, like any neural network, is to be learned from data,
not designed as we are suggesting you do here.
| text |
ch13 | page_043.txt | 13.6. REGULARIZATION
551
13.6
Regularization
Thus far, we have presented our goal as one of minimizing loss (i.e., prediction
error) on the training set. Gradient descent and stochastic gradient descent
help us achieve this objective.
In practice, the real objective of training is
to minimize the loss on new and hitherto unseen inputs.
Our hope is that
our training set is representative of unknown future inputs, so a low loss on the
training set translates into good performance on new inputs. Unfortunately, the
trained model sometimes learns idiosyncrasies of the training data that allow
it have low training loss, but not generalize well to new inputs – the familiar
problem of overfitting.
How can we tell if a model has overfit? In general, we split the available
data into a training set and a test set.
We train the model using only the
training-set data, withholding the test set. We then evaluate the performance
of the model on the test set. If the model performs much worse on the test set
than on the training set, we know the model has overfit. Assuming data points
are independent of one another, we can pick a fraction of the available data
points at random to form the test set. A common ratio for the training:test
split is 80:20. i.e, 80% of the data for training and 20% for test. We have to
be careful, however: in sequence-learning problems (e.g., modeling time series),
the state of the sequence at any point in time encodes information about the
past. In such cases the final piece of the sequence is a better test set.
Overfitting is a general problem that affects all machine-learning models.
However, deep neural networks are particularly susceptible to overfitting, be-
cause they use many more parameters (weights and biases) than other kinds of
models. Several techniques have been developed to reduce overfitting in deep
networks, usually by trading higher training error for better generalization. The
process is referred to as model regularization. In this section we describe some
of the most important regularization methods for deep learning.
13.6.1
Norm Penalties
Gradient descent is not guaranteed to learn parameters (weights and biases)
that reduce the training loss to an absolute minimum. In practice, the pro-
cedure learns parameters that correspond to a local minimum in the training
loss. There are usually many local minima, and some might lead to better gen-
eralization than others. In practice, it has been observed that solutions where
the learned weights have low absolute values tend to generalize better than
solutions with large weights.
We can force gradient descent to favor solutions with low weight values by
adding a term to the loss function. Suppose w is the vector of all the weight
values in the model, and L0 is loss function used by our model. We define a
new loss function L as follows:
L = L0 + α ∥w∥2
(13.17)
| text |
ch13 | page_044.txt | 552
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
The loss function L penalizes large weight values. Here α is a hyperparameter
that trades offbetween minimizing the original loss function L0 and the penalty
associated with the L2-norm of the weights. Instead of the L2-norm, we could
penalize the L1-norm of the weights:
L = L0 + α
X
i
|wi|
(13.18)
In practice, it is observed that the L2-norm penalty works best for most
applications. The L1-norm penalty is useful in some situations calling for model
compression, because it tends to produce models where many of the weights are
zero.
13.6.2
Dropout
Dropout is a technique that reduces overfitting by making random changes to
the underlying deep neural network. Recall that when we train using stochastic
gradient descent, at each step we sample at random a minibatch of inputs to
process. When using dropout, we also select at random a certain fraction (say
half) of all the hidden nodes from the network and delete them, along with
any edges connected to them. We then perform forward propagation and back-
propagation for the minibatch using this modified network, and update the
weights and biases. After processing the minibatch, we restore all the deleted
nodes and edges. When we sample the next minibatch, we delete a different
random subset of nodes and repeat the training process.
The fraction of hidden nodes deleted each time is a hyperparameter called
the dropout rate.
When training is complete, and we actually use the full
network, we need to take into account that the full network contains a larger
number of hidden nodes than the networks used for training. We therefore need
to scale the weight on each outgoing edge from a hidden node by the dropout
rate.
Why does dropout reduce overfitting? Several hypotheses have been put
forward, but perhaps the most convincing argument is that dropout allows a
single neural network to behave effectively as a collection of neural networks.
Imagine that we have a collection of neural networks, each with a different
network topology. Suppose we trained each network independently using the
training data, and used some kind of voting or averaging scheme to create a
higher-level model. Such a scheme would perform better than any of the indi-
vidual networks. The dropout technique simulates this setup without explicitly
creating a collection of neural networks.
13.6.3
Early Stopping
In Section 13.6.1 we suggested iterating through training examples (or mini-
batches) until we reach a local minimum in the loss function. In practice, this
approach leads to overfitting. It has been observed that while the loss on the
| text |
ch13 | page_045.txt | 13.7. SUMMARY OF CHAPTER 13
553
training set (the training loss) decreases through the training process, the loss
on the test set (the test loss) often behaves differently. The test loss falls dur-
ing the initial part of the training, and then many hit a minimum and actually
increase after a large number of training iterations, even as the training loss
keeps falling.
Intuitively, the point at when the test loss starts to increase is the point at
which the training process has started learning idiosyncrasies of the training
data rather than a generalizable model. A simple approach to avoiding this
problem is to stop the training when the test loss stops falling. There is, how-
ever, a subtle problem with this approach: we might inadvertently overfit to the
test data (rather than to the training data) by stopping training at the point
of minimum test loss. Therefore, the test error no longer is a reliable measure
of the true performance of the model on hitherto unseen inputs.
The usual solution is to use a third subset of inputs, the validation set, to
determine the point at which we stop training. We split the data not just into
training and test sets, but into three groups: training, validation, and test.
Both the validation and test sets are withheld from the training process. When
the loss on the validation set stops decreasing, we stop the training process.
Since the test set has played no role at all in the training process, the test error
remains a reliable indicator of the true performance of the model.
13.6.4
Dataset Augmentation
The accuracy of most machine-learning models increases when we provide ad-
ditional training data. Usually, larger training sets also lead to less overfitting.
When the actual training data available is limited, we can often create addi-
tional synthetic training examples by applying transformations or adding noise.
For example, consider the digit-classification problem we encountered in
Example 13.7. It is clear that if we rotate an image corresponding to a digit
by a few degrees, it still remains an image of the same digit. We can augment
the training data by systematically applying transformations of this kind. One
way to think of this process is as a way to encode additional domain knowledge
(e.g., a slightly distorted image of a cat is still an image of a cat).
13.7
Summary of Chapter 13
✦Neural Nets: A neural net is a collection of perceptrons (nodes), usually
organized in layers, where the outputs from one layer provide inputs to
the next layers. The first (inout) layer takes external inputs, and the
last (output) layer indicates the class of the input. Other layers in the
middle are called hidden layers and generally are trained to recognize
intermediate concepts needed to determine the output.
✦Types of Layers: Many layers are fully connected, meaning that each node
in the layer has all the nodes of the previous layer as inputs. Other layers
| text |
ch13 | page_046.txt | 554
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
are pooled, meaning that the nodes of the previous layer are partitioned,
and each node of this layer takes as input only the nodes of one block
of the partition. Convolutional layers are also used, especially in image
processing applications.
✦Convolutional Layers: Convolutional layers can be viewed as if their nodes
were organized into a two-dimensional array of pixels, with each pixel
represented by the same collection of nodes. The weights on corresponding
nodes from different pixels must be the same, so they are in effect the same
node, and we need learn only one set of weights for each family of nodes,
one from each pixel.
✦Activation Functions: The output of a node in a neural net is determined
by first taking the weighted sum of its inputs, using the weights that are
learned during the process of training the net. An activation function
is then applied to this sum. Common activation functions include the
sigmoid function, the hyperbolic tangent, softmax, and various forms of
linear recified unit functions.
✦Loss Functions: These measure the difference between the output of the
net and the correct output according to the training set. Commonly used
loss functions include squared-error loss, Huber loss, classification loss,
and cross-entropy loss.
✦Training a Neural Net: We train a neural net by repeatedly computing
the output of the net on training examples and computing the average
loss on the training examples. Weights on the nodes are then adjusted by
propagating the loss backward through the net, using the backpropagation
algorithm.
✦Backpropagation: By choosing our activation functions and loss functions
to be differentiable, we can compute a deriative of the loss function with
respect to every weight in the network.
Thus, we can determine the
direction in which to adjust each weight to reduce the loss. Using the chain
rule, these directions can be computed layer-by-layer, from the output to
the input.
✦Convolutional Neural Networks: These typically consist of a large num-
ber of convolutional layers, along with pooled layers and fully connected
layers. They are well suited to processing images, where the first convolu-
tional layers recognize simple features, such as boundries, and later layers
recognize progressively more complex features.
✦Recurrent Neural Networks: These are designed to recognize sequences,
such as sentences (sequences of words). There is one layer for each position
in the sequence, and the nodes are divided into families, which each have
one node at each layer. The nodes of a family are constrained to have
| text |
ch13 | page_047.txt | 13.8. REFERENCES FOR CHAPTER 13
555
the same weights, so the training process therefore needs to deal with a
relatively small number of weights.
✦Long Short-Term Memory Networks: These improve on RNN’s by adding
a second state vector – the cell state – to enable some information about
the sequence to be retained, while most information is forgotten after a
while. In addition, we learn gate vectors that control what information is
retained from the input, the state, and the output.
✦Avoiding Overfitting: There are a number of specialized techniques de-
signed to avoid overfitting a deep network. These include penalizing large
weights, randomly dropping some nodes each time we apply a step of
gradient descent, and use of a validation set to enable us to stop training
when the loss on the validation set bottoms out.
13.8
References for Chapter 13
For information on TensorFlow, see [12]. You can learn about Pytorch at [10]
and about Caffe at [1]. The MNIST database is described in [9].
Backpropagation as the way to train deep neural nets is from [11].
The idea of convolutional neural networks begins with [2], which defined
convolutional layers and pooling layers. However, it was [13] that introduced
the idea of requiring nodes in one convolutional layer to share weights. The
application of these networks to character recognition and other important tasks
is from [8] and [7].
Also see [6] for the application to ImageNet of CNN’s.
ImageNet is available from [5].
Recurrent neural networks first appeared in [4]. Long short-term memory
is from [3].
1. http://caffe2.ai
2. Fukushima, K., “Neocognitron, a self-organizing neural network model
for a mechanism of pattern recognition unaffected by shift of position,”
Biological Cybernetics 36:1 (1980), pp. 193–202.
3. Hochreiter, S. and J. Schmidhuber, “Long Short-Term Memory,” Neural
Computation 9:8 (1997), pp. 1735-1780.
4. J. J. Hopfield, “Neural networks and physical systems with emergent col-
lective computational abilities,” Proceedings of the National Academy of
Sciences 79:8 (1982), pp. 2554-2558.
5. http://www.image-net.org.
6. Krizhevsky, A, I. Sutskever, and G.E. Hinton, “Image classification with
deep convolutional neural networks,” Advances in Neural Information
Processing Systems, pp. 1097–1105, 2012.
| text |
ch13 | page_048.txt | 556
CHAPTER 13. NEURAL NETS AND DEEP LEARNING
7. LeCun, Y. and Y. Bengio, “Convolutional networks for images, speech,
and time series,” The Handbook of Brain Theory and Neural Networks
(M. Arbib, ed.) 3361:10 (1995).
8. LeCun, Y., B. Boser, J.S. Denker, D. Henderson, R.E. Howard, and
W. Hubbard, “Backpropagation applied to handwritten zip code recogni-
tion,” Neural Computation 1:4 (1989) pp. 541–551.
9. LeCun, Y., C. Cortes, and C.J.C. Burges, “The MNIST database of hand-
written digits,” http://http://yann.lecun.com/exdb/mnist/
10. http://pytorch.org
11. Rumelhart, D.E., G.E. Hinton, and R.J. Williams, “Learning representa-
tions by back-propagating errors, Nature 323 (1986), pp. 533–536.
12. http://www.tensorflow.org
13. Waibel, A., T. Hanazawa, G.E. Hinton, K. Shikano, and K.J. Lang,
“Phoneme recognition using time-delay neural networks,” IEEE Trans-
actions on Acoustics, Speech, and Signal Processing 37:3 (1989), pp. 328–
339.
| text |
README.md exists but content is empty.
- Downloads last month
- 678