arxiv_db / papers /1311 /1311.2901.tex
taesiri's picture
Initial commit
8cf8144
raw
history blame
42.6 kB
\documentclass{article}
\usepackage{graphicx} \usepackage{subfigure}
\usepackage{natbib,amsmath}
\usepackage{algorithm}
\usepackage{algorithmic}
\usepackage{hyperref}
\newcommand{\theHalgorithm}{\arabic{algorithm}}
\usepackage[accepted]{icml2013}
\def\x{{\mathbf x}}
\def\h{{\mathbf h}}
\def\b{{\mathbf b}}
\def\t{{\mathbf t}}
\def\L{{\cal L}}
\def\BE{\vspace{-0.0mm}\begin{equation}}
\def\EE{\vspace{-0.0mm}\end{equation}}
\def\BEA{\vspace{-0.0mm}\begin{eqnarray}}
\def\EEA{\vspace{-0.0mm}\end{eqnarray}}
\newcommand{\alg}[1]{Algorithm \ref{alg:#1}}
\newcommand{\eqn}[1]{Eqn.~\ref{eqn:#1}}
\newcommand{\fig}[1]{Fig.~\ref{fig:#1}}
\newcommand{\tab}[1]{Table~\ref{tab:#1}}
\newcommand{\secc}[1]{Section~\ref{sec:#1}}
\def\etal{{\textit{et~al.~}}}
\def\units{{\text{units of }}}
\icmltitlerunning{Visualizing and Understanding Convolutional Networks}
\begin{document}
\twocolumn[
\icmltitle{Visualizing and Understanding Convolutional Networks}
\icmlauthor{Matthew D. Zeiler}{zeiler@cs.nyu.edu}
\icmladdress{Dept. of Computer Science, Courant Institute,
New York University}
\icmlauthor{Rob Fergus}{fergus@cs.nyu.edu}
\icmladdress{Dept. of Computer Science, Courant Institute,
New York University}
\icmlkeywords{boring formatting information, machine learning, ICML}
\vskip 0.3in
]
\begin{abstract}
Large Convolutional Network models have recently demonstrated
impressive classification performance on the ImageNet benchmark
\cite{Kriz12}. However there is no clear understanding of
why they perform so well, or how they might be improved. In this
paper we address both issues. We introduce a novel visualization
technique that gives insight into the function of intermediate feature
layers and the operation of the classifier. Used in a diagnostic role, these
visualizations allow us to find model architectures that outperform Krizhevsky
\etal on the ImageNet classification benchmark. We also perform an
ablation study to discover the performance contribution from different
model layers. We show our ImageNet model
generalizes well to other datasets: when the softmax classifier is
retrained, it convincingly beats the current state-of-the-art
results on Caltech-101 and Caltech-256 datasets.
\end{abstract}
\vspace{-4mm}
\section{Introduction}
\vspace{-2mm}
Since their introduction by \cite{Lecun1989} in the early 1990's, Convolutional
Networks (convnets) have demonstrated excellent performance at
tasks such as hand-written digit classification and face
detection. In the last year, several papers have shown that they can also deliver
outstanding performance on more challenging visual classification tasks. \cite{Ciresan12} demonstrate state-of-the-art performance on NORB and
CIFAR-10 datasets. Most notably, \cite{Kriz12} show
record beating performance on the ImageNet 2012 classification
benchmark, with their convnet model achieving an error rate of 16.4\%,
compared to the 2nd place result of 26.1\%. Several factors are
responsible for this renewed interest in convnet models: (i) the
availability of much larger training sets, with millions of labeled
examples; (ii) powerful GPU implementations, making the training of
very large models practical and (iii) better model regularization
strategies, such as Dropout \cite{Hinton12}.
Despite this encouraging progress, there is still little insight into
the internal operation and behavior of these complex models, or how
they achieve such good performance. From a scientific standpoint,
this is deeply unsatisfactory. Without clear understanding of how and
why they work, the development of better models is reduced to
trial-and-error. In this paper we introduce a visualization technique
that reveals the input stimuli that excite individual feature maps at any
layer in the model. It also allows us to observe the evolution of
features during training and to diagnose potential problems with the
model. The visualization technique we propose uses a multi-layered
Deconvolutional Network (deconvnet), as proposed by \cite{Zeiler11},
to project the feature activations back to the input pixel space. We also perform a
sensitivity analysis of the classifier output by occluding portions of
the input image, revealing which parts of the scene are important for
classification.
Using these tools, we start with the architecture of \cite{Kriz12} and
explore different architectures, discovering ones that outperform
their results on ImageNet. We then explore the generalization ability
of the model to other datasets, just retraining the softmax classifier
on top. As such, this is a form of supervised pre-training, which
contrasts with the unsupervised pre-training methods popularized by
\cite{Hinton2006a} and others \cite{Bengio2007,Vincent2008}. The
generalization ability of convnet features is also explored in
concurrent work by \cite{Donahue13}.
\vspace{-0mm}
\subsection{Related Work}
\vspace{-0mm} Visualizing features to gain intuition about the network
is common practice, but mostly limited to the 1st layer where
projections to pixel space are possible. In higher layers this is not
the case, and there are limited methods for interpreting
activity. \cite{Erhan09} find the optimal stimulus for each unit by
performing gradient descent in image space to maximize the unit's
activation. This requires a careful initialization and does not give
any information about the unit's invariances. Motivated by the
latter's short-coming, \cite{Le10} (extending an idea by
\cite{Berkes06}) show how the Hessian of a given unit may be computed
numerically around the optimal response, giving some insight into
invariances. The problem is that for higher layers, the invariances
are extremely complex so are poorly captured by a simple quadratic
approximation. Our approach, by contrast, provides a non-parametric
view of invariance, showing which patterns from the training set
activate the feature map. \cite{Donahue13} show
visualizations that identify patches within a dataset that are
responsible for strong activations at higher layers in the model. Our
visualizations differ in that they are not just crops of input images,
but rather top-down projections that reveal structures within each
patch that stimulate a particular feature map.
\section{Approach}
We use standard fully supervised convnet models throughout
the paper, as defined by \cite{Lecun1989} and \cite{Kriz12}. These models map a color 2D input image $x_i$,
via a series of layers, to a
probability vector $\hat{y_i}$ over the $C$ different classes. Each
layer consists of (i) convolution of the previous layer output (or, in
the case of the 1st layer, the input image) with a set of learned
filters; (ii) passing the responses through a rectified linear
function ({\em $relu(x)=\max(x,0)$}); (iii)
[optionally] max pooling over local neighborhoods and (iv) [optionally]
a local contrast operation that normalizes the responses across
feature maps. For more details of these operations, see \cite{Kriz12} and \cite{Jarrett2009}. The top few
layers of the network are conventional fully-connected networks and
the final layer is a softmax classifier. \fig{arch} shows
the model used in many of our experiments.
We train these models using a large set of $N$ labeled images
$\{x,y\}$, where label $y_i$ is a discrete variable indicating the
true class. A cross-entropy loss function, suitable for image
classification, is used to compare $\hat{y_i}$ and $y_i$. The
parameters of the network (filters in the convolutional layers, weight
matrices in the fully-connected layers and biases) are trained by
back-propagating the derivative of the loss with respect to the
parameters throughout the network, and updating the parameters via
stochastic gradient descent. Full details of training are given in
\secc{training}.
\subsection{Visualization with a Deconvnet}
Understanding the operation of a convnet requires interpreting the
feature activity in intermediate layers. We present a novel way to
{\em map these activities back to the input pixel space}, showing what
input pattern originally caused a given activation in the feature
maps. We perform this mapping with a Deconvolutional Network
(deconvnet) \cite{Zeiler11}. A deconvnet can be thought of as a
convnet model that uses the same components (filtering, pooling) but
in reverse, so instead of mapping pixels to features does the
opposite. In \cite{Zeiler11}, deconvnets were proposed as a way of
performing unsupervised learning. Here, they are not used in any
learning capacity, just as a probe of an already trained convnet.
To examine a convnet, a deconvnet is attached to each of its layers,
as illustrated in \fig{deconv}(top), providing a continuous
path back to image pixels. To start, an input image is
presented to the convnet and features computed throughout the
layers. To examine a given convnet activation, we set all other activations in
the layer to zero and pass the feature maps as input to the attached
deconvnet layer. Then we successively (i) unpool, (ii) rectify and
(iii) filter to reconstruct the activity in the layer beneath that
gave rise to the chosen activation. This is then repeated until input
pixel space is reached.
\noindent {\bf Unpooling:} In the convnet, the max pooling operation
is non-invertible, however we can obtain an approximate inverse by
recording the locations of the maxima within each pooling region in a
set of {\em switch} variables. In the deconvnet, the unpooling
operation uses these switches to place the reconstructions from the
layer above into appropriate locations, preserving the structure of
the stimulus. See \fig{deconv}(bottom) for an illustration of the procedure.
\noindent {\bf Rectification:} The convnet uses {\em relu}
non-linearities, which rectify the feature maps thus ensuring the
feature maps are always positive. To obtain valid feature
reconstructions at each layer (which also should be positive), we pass
the reconstructed signal through a {\em relu} non-linearity.
\noindent {\bf Filtering:} The convnet uses learned filters to
convolve the feature maps from the previous layer. To invert this, the
deconvnet uses transposed versions of the same filters, but applied to the rectified maps,
not the output of the layer beneath. In
practice this means flipping each filter vertically and horizontally.
Projecting down from higher layers uses the switch settings generated
by the max pooling in the convnet on the way up. As these switch
settings are peculiar to a given input image, the reconstruction
obtained from a single activation thus resembles a small piece of the
original input image, with structures weighted according to their
contribution toward to the feature activation. Since the model is
trained discriminatively, they implicitly show which parts of the
input image are discriminative. Note that these projections are {\em
not} samples from the model, since there is no generative process involved.
\begin{figure}[h!]
\vspace{-3mm}
\begin{center}
\includegraphics[width=3.3in]{conv_deconv.pdf}
\includegraphics[width=3.3in]{unpool.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{Top: A deconvnet layer (left) attached to a convnet layer
(right). The deconvnet will reconstruct an approximate version of
the convnet features from the layer beneath. Bottom:
An illustration of the unpooling operation in the deconvnet, using {\em switches}
which record the location of the local max in each pooling region
(colored zones) during pooling in the convnet. }
\label{fig:deconv}
\vspace*{-0.3cm}
\end{figure}
\section{Training Details} \label{sec:training}
We now describe the large convnet model that will be visualized in
\secc{vis}. The architecture, shown in \fig{arch}, is similar to that
used by \cite{Kriz12} for ImageNet
classification. One difference is that the sparse connections used in
Krizhevsky's layers 3,4,5 (due to the model being split across 2
GPUs) are replaced with dense connections in our model. Other
important differences relating to layers 1 and 2 were made following
inspection of the visualizations in \fig{compareAlex}, as described in
\secc{selection}.
The model was trained on the ImageNet 2012 training set (1.3 million
images, spread over 1000 different classes). Each RGB image was
preprocessed by resizing the smallest dimension to 256, cropping the
center 256x256 region, subtracting the per-pixel mean (across all
images) and then using 10 different sub-crops of size 224x224
(corners $+$ center with(out) horizontal flips). Stochastic gradient
descent with a mini-batch size of 128 was used to update the
parameters, starting with a learning rate of $10^{-2}$, in conjunction
with a momentum term of $0.9$. We anneal the
learning rate throughout training manually when the validation error
plateaus. Dropout \cite{Hinton12} is used in the
fully connected layers (6 and 7) with a rate of 0.5. All weights are initialized to $10^{-2}$ and
biases are set to 0.
Visualization of the first layer filters during training reveals that
a few of them dominate, as shown in \fig{compareAlex}(a). To combat
this, we renormalize each filter in the convolutional layers whose RMS
value exceeds a fixed radius of $10^{-1}$ to this fixed radius. This
is crucial, especially in the first layer of the model, where the
input images are roughly in the [-128,128] range. As in
\cite{Kriz12}, we produce multiple different crops and flips of each
training example to boost training set size. We stopped training after
70 epochs, which took around 12 days on a single
GTX580 GPU, using an implementation based on \cite{Kriz12}.
\section{Convnet Visualization}
\label{sec:vis}
Using the model described in \secc{training}, we now use the deconvnet
to visualize the feature activations on the ImageNet validation set.
\noindent {\bf Feature Visualization:} \fig{top9feat} shows feature visualizations from our
model once training is complete. However, instead of showing the
single strongest activation for a given feature map, we show the top 9
activations. Projecting each separately down to pixel space reveals
the different structures that excite a given feature map, hence
showing its invariance to input deformations. Alongside these
visualizations we show the corresponding image patches. These have
greater variation than visualizations as the latter solely focus on
the discriminant structure within each patch. For example, in layer 5,
row 1, col 2, the patches appear to have little in common, but the
visualizations reveal that this particular feature map focuses on the
grass in the background, not the foreground objects.
\onecolumn
\begin{figure*}[h!]
\begin{center}
\includegraphics[width=6.0in]{rob3_new6.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{Visualization of features in a fully trained model. For
layers 2-5 we show the top 9 activations in a random subset of
feature maps across the validation data, projected down to pixel
space using our deconvolutional network approach. Our
reconstructions are {\em not} samples from the model: they are
reconstructed patterns from the validation set that cause high activations in a
given feature map. For each feature map we also show the
corresponding image patches. Note: (i) the the strong grouping
within each feature map, (ii) greater invariance at higher layers
and (iii) exaggeration of discriminative parts of the image,
e.g.~eyes and noses of dogs (layer 4, row 1, cols 1). Best viewed in electronic form. }
\label{fig:top9feat}
\vspace*{-0.3cm}
\end{figure*}
\twocolumn
The projections from each layer show the hierarchical nature of
the features in the network. Layer
2 responds to corners and other edge/color conjunctions. Layer 3 has
more complex invariances, capturing similar textures (e.g.~mesh
patterns (Row 1, Col 1); text (R2,C4)). Layer 4 shows significant
variation, but is more class-specific: dog faces
(R1,C1); bird's legs (R4,C2). Layer 5 shows entire objects
with significant pose variation, e.g.~keyboards (R1,C11) and dogs (R4).
\begin{figure*}[t!]
\vspace*{-0.2cm}
\begin{center}
\includegraphics[width=7in]{arch_old.pdf}
\end{center}
\vspace*{-0.4cm}
\caption{Architecture of our 8 layer convnet model. A 224 by 224 crop of an
image (with 3 color planes) is presented as the input. This is
convolved with 96 different 1st layer filters (red), each of size 7 by 7,
using a stride of 2 in both x and y. The resulting feature maps are
then: (i) passed through a rectified linear function (not shown),
(ii) pooled (max
within 3x3 regions, using stride 2) and (iii) contrast normalized
across feature maps to give 96 different 55 by 55
element feature maps. Similar operations are repeated in layers
2,3,4,5. The last two layers are fully connected, taking features
from the top convolutional layer as input in vector form (6 $\cdot$ 6 $\cdot$
256 = 9216 dimensions). The final layer is a $C$-way softmax
function, $C$ being the number of classes. All filters and feature maps are square in shape.}
\label{fig:arch}
\vspace*{-0.3cm}
\end{figure*}
\noindent {\bf Feature Evolution during Training:} \fig{evolve}
visualizes the progression during training of the strongest activation
(across all training examples) within a given feature map projected
back to pixel space. Sudden jumps in appearance result from a change
in the image from which the strongest activation originates. The lower
layers of the model can be seen to converge within a few
epochs. However, the upper layers only develop develop after a
considerable number of epochs (40-50), demonstrating the need to let
the models train until fully converged.
\noindent {\bf Feature Invariance:} \fig{invariance} shows 5 sample images being translated, rotated and
scaled by varying degrees while looking at the changes in the feature
vectors from the top and bottom layers of the model, relative to the
untransformed feature. Small
transformations have a dramatic effect in the first layer of the model,
but a lesser impact at the top feature layer, being quasi-linear for
translation \& scaling. The network output is stable to translations
and scalings. In general, the output is not invariant to rotation,
except for object with rotational symmetry (e.g.~entertainment center).
\begin{figure*}[t!]
\begin{center}
\includegraphics[width=7in]{rob2_small.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{Evolution of a randomly chosen subset of model features through training. Each layer's
features are displayed in a different block. Within each
block, we show a randomly chosen subset of features at epochs [1,2,5,10,20,30,40,64]. The
visualization shows the strongest activation (across all
training examples) for a given feature map, projected down to pixel
space using our deconvnet approach. Color contrast is artificially
enhanced and the figure is best viewed in
electronic form.}
\label{fig:evolve}
\vspace*{-0.3cm}
\end{figure*}
\begin{figure*}[t!]
\begin{center}
\includegraphics[width=6.5in]{invariance_small.pdf}
\end{center}
\vspace*{-0.6cm}
\caption{Analysis of vertical translation, scale, and rotation invariance within the
model (rows a-c respectively). Col 1: 5 example images undergoing
the transformations. Col 2 \& 3: Euclidean distance between feature
vectors from the original and transformed images in layers 1 and 7
respectively. Col 4: the probability of the true label for each
image, as the image is transformed. }
\label{fig:invariance}
\vspace*{-0.3cm}
\end{figure*}
\subsection{Architecture Selection} \label{sec:selection} While visualization
of a trained model gives insight into its operation, it can also
assist with selecting good architectures in the first place. By
visualizing the first and second layers of Krizhevsky \etal's
architecture (\fig{compareAlex}(b) \& (d)), various problems are
apparent. The first layer filters are a mix of extremely high and low
frequency information, with little coverage of the mid
frequencies. Additionally, the 2nd layer visualization shows aliasing
artifacts caused by the large stride 4 used in the 1st layer
convolutions. To remedy these problems, we (i) reduced the 1st layer
filter size from 11x11 to 7x7 and (ii) made the stride of the
convolution 2, rather than 4. This new architecture retains much more
information in the 1st and 2nd layer features, as shown in
\fig{compareAlex}(c) \& (e). More importantly, it also improves the classification
performance as shown in \secc{modelsizes}.
\begin{figure*}[t!]
\begin{center}
\includegraphics[width=6.3in]{rob4_small.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{(a): 1st layer features without feature scale clipping. Note
that one feature dominates. (b): 1st layer features from
\cite{Kriz12}. (c): Our 1st layer features. The smaller stride (2 vs
4) and filter size (7x7 vs 11x11) results in more distinctive
features and fewer ``dead'' features. (d): Visualizations of 2nd
layer features from \cite{Kriz12}. (e):
Visualizations of our 2nd layer features. These are cleaner, with no
aliasing artifacts that are visible in (d). }
\label{fig:compareAlex}
\vspace*{-0.3cm}
\end{figure*}
\subsection{Occlusion Sensitivity}
With image classification approaches, a natural question is if the
model is truly identifying the location of the object in the image, or
just using the surrounding context. \fig{block_expt} attempts to
answer this question by systematically occluding different portions of
the input image with a grey square, and monitoring the output of the
classifier. The examples clearly show the model is localizing the
objects within the scene, as the probability of the correct class
drops significantly when the object is occluded. \fig{block_expt} also
shows visualizations from the strongest feature map of the top
convolution layer, in addition to
activity in this map (summed over spatial locations) as a function of occluder position. When the
occluder covers the image region that appears in the visualization, we
see a strong drop in activity in the feature map. This shows that the
visualization genuinely corresponds to the image structure that
stimulates that feature map, hence validating the other visualizations
shown in \fig{evolve} and \fig{top9feat}.
\begin{figure*}[t!]
\begin{center}
\includegraphics[width=7.0in]{rob1_3.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{Three test examples where we systematically cover up different portions
of the scene with a gray square (1st column) and see how the top
(layer 5) feature maps ((b) \& (c)) and classifier output ((d) \& (e)) changes. (b): for each
position of the gray scale, we record the total activation in one
layer 5 feature map (the one with the strongest response in the
unoccluded image). (c): a visualization of this feature map projected
down into the input image (black square), along with visualizations of
this map from other images. The first row example shows the strongest feature
to be the dog's face. When this is covered-up the activity in the
feature map decreases (blue area in (b)). (d): a map of correct class
probability, as a function of the position of the gray
square. E.g.~when the dog's face is obscured, the probability for
``pomeranian'' drops significantly. (e): the most probable label as a
function of occluder position. E.g.~in the 1st row, for most locations
it is ``pomeranian'', but if the dog's face is obscured but not the
ball, then it predicts ``tennis ball''. In the 2nd example, text on
the car is the strongest feature in layer 5, but the classifier is
most sensitive to the wheel. The 3rd example contains multiple
objects. The strongest feature in layer 5 picks out the faces, but the
classifier is sensitive to the dog (blue region in (d)),
since it uses multiple feature maps.}
\label{fig:block_expt}
\vspace*{-0.3cm}
\end{figure*}
\subsection{Correspondence Analysis}
\vspace{-2mm}
Deep models differ from many existing recognition approaches in that
there is no explicit mechanism for establishing correspondence between
specific object parts in different images (e.g. faces have a particular
spatial configuration of the eyes and nose). However, an intriguing possibility is that deep models might
be {\em implicitly} computing them. To explore this, we take 5
randomly drawn dog images with frontal pose and systematically mask out
the same part of the face in each image (e.g.~all left eyes, see \fig{corr_ims}). For each
image $i$, we then compute: $\epsilon^l_i = x^l_i - \tilde{x}^l_i$,
where $x^l_i$ and $\tilde{x}^l_i$ are the feature vectors at layer $l$
for the original and occluded images respectively. We then measure the
consistency of this difference vector $\epsilon$ between all related image
pairs $(i,j)$: $\Delta_l = \sum_{i,j=1, i \neq j}^{5} \mathcal{H}(
\text{sign}(\epsilon^l_i),\text{sign}(\epsilon^l_j))$, where
$\mathcal{H}$ is Hamming distance. A lower value indicates
greater consistency in the change resulting from the masking
operation, hence tighter correspondence between the same object parts
in different images (i.e.~blocking the left eye changes the
feature representation in a consistent way). In \tab{corr} we compare the $\Delta$ score for
three parts of the face (left eye, right eye and nose) to random parts
of the object, using features from layer $l=5$ and $l=7$. The lower score
for these parts, relative to random object regions, for the layer $5$ features
show the model does establish some degree of correspondence.
\begin{figure}[t!]
\vspace{-0mm}
\begin{center}
\includegraphics[width=3.2in]{corr_ims.pdf}
\end{center}
\vspace*{-0.4cm}
\caption{Images used for correspondence experiments. Col 1: Original
image. Col 2,3,4: Occlusion of the right eye, left eye, and nose
respectively. Other columns show examples of random occlusions.}
\label{fig:corr_ims}
\vspace*{-0.0cm}
\end{figure}
\begin{table}[!h]
\vspace*{-0mm}
\begin{center}
\small
\begin{tabular}{|l||c|c|}
\hline
& Mean Feature & Mean Feature \\ & Sign Change & Sign Change \\ Occlusion Location & Layer 5 & Layer 7 \\
\hline
Right Eye & $0.067 \pm 0.007 $ & $0.069 \pm 0.015 $\\ \hline
Left Eye & $0.069 \pm 0.007 $ & $0.068 \pm 0.013 $\\ \hline
Nose & $0.079 \pm 0.017 $ & $0.069 \pm 0.011 $\\ \hline \hline
Random & $0.107 \pm 0.017 $ & $0.073 \pm 0.014 $\\ \hline
\end{tabular}
\vspace*{-0mm}
\caption{Measure of correspondence for different object parts in 5
different dog images. The lower scores for the eyes and nose
(compared to random object parts) show
the model implicitly establishing some form of correspondence
of parts at layer 5 in the model. At layer 7, the scores are more
similar, perhaps due to upper layers trying to discriminate between the
different breeds of dog. }
\label{tab:corr}
\vspace*{-5mm}
\end{center}
\end{table}
\section{Experiments}
\subsection{ImageNet 2012}
\label{ImageNet}
This dataset consists of 1.3M/50k/100k training/validation/test
examples, spread over 1000 categories. \tab{ImageNet} shows our results
on this dataset.
Using the exact architecture specified in \cite{Kriz12}, we attempt to replicate their result on the validation set. We achieve an
error rate within $0.1\%$ of their reported value on the
ImageNet 2012 validation set.
Next we analyze the performance of our model with the architectural
changes outlined in \secc{selection} ($7\times7$ filters in layer 1 and
stride $2$ convolutions in layers 1 \& 2). This model, shown in
\fig{arch}, significantly outperforms the architecture of
\cite{Kriz12}, beating their single model result by $1.7\%$ (test top-5). When we combine
multiple models, we obtain a test error of {\bf ${\bf 14.8\%}$, the best
published performance on this dataset}\footnote{This performance has
been surpassed in the recent Imagenet 2013 competition (\url{http://www.image-net.org/challenges/LSVRC/2013/results.php}).} (despite only using the 2012 training
set). We note that this error is almost half that of the top non-convnet
entry in the ImageNet 2012 classification challenge, which obtained
$26.2\%$ error \cite{ISI}.
\begin{table}[h!]
\scriptsize
\vspace*{-3mm}
\begin{center}
\begin{tabular}{|l|l|l|l|}
\hline
& Val & Val & Test \\
Error \% & Top-1 & Top-5 & Top-5 \\
\hline \hline
\cite{ISI} & - & - & $26.2$ \\ \hline
\cite{Kriz12}, 1 convnet & $40.7$ & $18.2$ & $--$ \\
\cite{Kriz12}, 5 convnets & $38.1$ & $16.4$ & $16.4$ \\
\cite{Kriz12}$^*$, 1 convnets & $39.0$ & $16.6$ & $--$ \\
\cite{Kriz12}$^*$, 7 convnets & $36.7$ & $15.4$ & $15.3$ \\
\hline \hline Our replication of & & & \\
\cite{Kriz12}, 1 convnet & $40.5$ & $18.1$ & $--$ \\
\hline 1 convnet as per \fig{arch} & $38.4$ & $16.5$ & $--$ \\
\hline 5 convnets as per \fig{arch} -- (a) & $36.7$ & $15.3$ & $15.3$ \\
\hline 1 convnet as per \fig{arch} but with & & & \\ layers 3,4,5: 512,1024,512
maps -- (b)& $37.5 $ & $16.0$ & $16.1$ \\
\hline 6 convnets, (a) \& (b) combined & $\bf{36.0} $ & $\bf{14.7}$ & $\bf{14.8}$ \\
\hline
\end{tabular}
\vspace*{-2mm}
\caption{ImageNet 2012 classification error rates. The $*$ indicates
models that were trained on both ImageNet 2011 and 2012 training sets.}
\label{tab:ImageNet}
\vspace*{-3mm}
\end{center}
\end{table}
\noindent {\bf Varying ImageNet Model Sizes:}
\label{sec:modelsizes} In \tab{modelSizes}, we first explore the
architecture of \cite{Kriz12} by adjusting the size of layers, or
removing them entirely. In each case, the model is trained from
scratch with the revised architecture. Removing the fully connected
layers (6,7) only gives a slight increase in error. This is surprising, given that they
contain the majority of model parameters. Removing two of
the middle convolutional layers also makes a relatively small
different to the error rate. However, removing both the middle
convolution layers and the fully connected layers yields a model with
only 4 layers whose performance is dramatically worse. This would
suggest that the overall depth of the model is important for
obtaining good performance. In \tab{modelSizes}, we modify our
model, shown in \fig{arch}. Changing the size of the fully connected
layers makes little difference to performance (same for model of
\cite{Kriz12}). However, increasing the size of the middle convolution layers
goes give a useful gain in performance. But increasing these, while also
enlarging the fully connected layers results in over-fitting.
\begin{table}[h!]
\scriptsize
\vspace*{0mm}
\begin{center}
\begin{tabular}{|l|l|l|l|}
\hline
& Train & Val & Val \\
Error \% & Top-1 & Top-1 & Top-5 \\ \hline
\hline Our replication of & & & \\
\cite{Kriz12}, 1 convnet & $35.1$ & $40.5$ & $18.1$ \\
\hline Removed layers 3,4 & $41.8 $ & $45.4 $ & $22.1 $ \\
\hline Removed layer 7 & $27.4 $ & $40.0$ & $18.4 $ \\
\hline Removed layers 6,7 & $27.4 $ & $44.8 $ & $22.4 $ \\
\hline
Removed layer 3,4,6,7 & $71.1$ & $71.3$ & $50.1$ \\
\hline Adjust layers 6,7: 2048 units & $40.3 $ & $41.7 $ & $18.8 $ \\
\hline Adjust layers 6,7: 8192 units & $26.8 $ & $40.0$ & $18.1$ \\
\hline \hline \hline Our Model (as per \fig{arch}) & $33.1 $ & $38.4 $ & $16.5 $ \\
\hline Adjust layers 6,7: 2048 units & $38.2 $ & $40.2 $ & $17.6 $ \\
\hline Adjust layers 6,7: 8192 units & $22.0 $ & $38.8 $ & $17.0 $ \\
\hline Adjust layers 3,4,5: 512,1024,512 maps & $18.8 $ & $\bf{37.5} $ & $\bf{16.0} $ \\
\hline Adjust layers 6,7: 8192 units and & & & \\ Layers 3,4,5: 512,1024,512 maps & $\bf{10.0} $ & $38.3 $ & $16.9 $ \\
\hline
\end{tabular}
\vspace*{-2mm}
\caption{ImageNet 2012 classification error rates with various
architectural changes to the model of \cite{Kriz12} and our model
(see \fig{arch}). }
\label{tab:modelSizes}
\vspace*{-8mm}
\end{center}
\end{table}
\subsection{Feature Generalization}
\vspace*{-2mm}
The experiments above show the importance of the convolutional part of
our ImageNet model in obtaining state-of-the-art performance.
This is supported by the visualizations of \fig{top9feat} which show
the complex invariances learned in the convolutional layers. We now
explore the ability of these feature extraction layers to generalize
to other datasets, namely Caltech-101 \cite{caltech101}, Caltech-256
\cite{caltech256} and PASCAL
VOC 2012. To do this, we keep layers 1-7 of our ImageNet-trained model
fixed and train a new softmax classifier on top (for the appropriate
number of classes) using the training images of the new dataset. Since
the softmax contains relatively few parameters, it can be trained
quickly from a relatively small number of examples, as is the case for
certain datasets.
The classifiers used by our model (a softmax) and other approaches
(typically a linear SVM) are of similar complexity, thus the
experiments compare our feature representation, learned from ImageNet,
with the hand-crafted features used by other methods. It is important
to note that {\em both} our feature representation and the
hand-crafted features are designed using images beyond the Caltech and
PASCAL training sets. For example, the hyper-parameters in HOG
descriptors were determined through systematic experiments on a
pedestrian dataset \cite{Dalal05}.
We also try a second strategy of training a model from scratch,
i.e.~resetting layers 1-7 to random values and train them, as well as
the softmax, on the training images of the dataset.
One complication is that some of the Caltech datasets have some images that are
also in the ImageNet training data. Using normalized correlation, we
identified these few ``overlap'' images\footnote{ For
Caltech-101, we found 44 images in common (out of 9,144 total images),
with a maximum overlap of 10 for any given class. For Caltech-256, we
found 243 images in common (out of 30,607 total images), with a
maximum overlap of 18 for any given class.} and removed them from our
Imagenet training set and then retrained our Imagenet models, so avoiding the
possibility of train/test contamination.
\vspace{3mm}
\noindent {\bf Caltech-101:} We follow the procedure of
\cite{caltech101} and randomly select 15 or 30 images per class for
training and test on up to 50 images per class reporting the average
of the per-class accuracies in \tab{caltech101}, using 5 train/test
folds. Training took 17 minutes for 30 images/class. The
pre-trained model beats the best reported result for 30 images/class
from \cite{Bo13} by 2.2\%. The convnet model trained from
scratch however does terribly, only achieving 46.5\%.
\begin{table}[h!]
\small
\vspace*{-3mm}
\begin{center}
\begin{tabular}{|l|l|l|}
\hline
& Acc \% & Acc \% \\
\# Train& 15/class & 30/class \\
\hline
\cite{Bo13} & $-$ & $81.4 \pm 0.33$ \\
\cite{Yang09} & $73.2$ & $84.3 $ \\
\hline \hline \hline Non-pretrained convnet & $22.8 \pm 1.5$ & $46.5 \pm 1.7$ \\
\hline ImageNet-pretrained convnet & $\bf{83.8 \pm 0.5}$ & $\bf{86.5 \pm 0.5}$ \\
\hline
\end{tabular}
\vspace*{-3mm}
\caption{Caltech-101 classification accuracy for our convnet models,
against two leading alternate approaches.}
\label{tab:caltech101}
\vspace*{-3mm}
\end{center}
\end{table}
\noindent {\bf Caltech-256:} We follow the procedure of
\cite{caltech256}, selecting 15, 30, 45, or 60 training images per
class, reporting the average of the per-class accuracies in
\tab{caltech256}. Our ImageNet-pretrained model beats the current
state-of-the-art results obtained by Bo \etal \cite{Bo13} by a
significant margin: 74.2\% vs 55.2\% for 60 training
images/class. However, as with Caltech-101, the model trained from
scratch does poorly. In \fig{256plot}, we explore the ``one-shot
learning'' \cite{caltech101} regime. With our pre-trained model, just
6 Caltech-256 training images are needed to beat the leading method
using 10 times as many images. This shows the power of the ImageNet
feature extractor.
\begin{table}[h!]
\small
\vspace*{0mm}
\begin{center}
\tabcolsep=0.07cm
\begin{tabular}{|l|l|l|l|l|}
\hline
& Acc \% & Acc \% & Acc \% & Acc \%\\
\# Train & 15/class & 30/class & 45/class & 60/class \\
\hline
\cite{Sohn11} & $35.1$ & $42.1 $ & $45.7$ & $47.9$ \\
\cite{Bo13} & $40.5 \pm 0.4$ & $48.0 \pm 0.2$ & $51.9 \pm 0.2$ & $55.2 \pm 0.3$ \\
\hline \hline
\hline Non-pretr. & $9.0 \pm 1.4$ & $22.5 \pm 0.7$ & $31.2 \pm 0.5$ & $38.8 \pm 1.4$ \\
\hline ImageNet-pretr. & $\bf{65.7 \pm 0.2}$ & $\bf{70.6 \pm 0.2}$ & $\bf{72.7 \pm 0.4}$ & $\bf{74.2 \pm 0.3}$ \\
\hline
\end{tabular}
\vspace*{-3mm}
\caption{Caltech 256 classification accuracies.}
\label{tab:caltech256}
\vspace*{-4mm}
\end{center}
\end{table}
\begin{figure}[t!]
\begin{center}
\includegraphics[width=2.5in]{Caltech256Plot.pdf}
\end{center}
\vspace*{-0.3cm}
\caption{Caltech-256 classification performance as the number of
training images per class is varied. Using only 6 training examples
per class with our pre-trained feature extractor, we surpass best
reported result by \cite{Bo13}. }
\label{fig:256plot}
\vspace*{-0.4cm}
\end{figure}
\vspace{2mm}
\noindent {\bf PASCAL 2012:} We used the standard training and validation
images to train a 20-way softmax on top of the ImageNet-pretrained
convnet. This is not ideal, as PASCAL images can contain multiple
objects and our model just provides a single exclusive prediction for each
image. \tab{pascal} shows the results on the test set. The PASCAL
and ImageNet images are quite different in nature, the former being
full scenes unlike the latter. This may explain our
mean performance being $3.2\%$ lower than the leading \cite{nus}
result, however we do beat them on 5 classes, sometimes by large margins.
\begin{table}[!h]
\tiny
\begin{center}
\begin{tabular}{|l|l|l|l||l|l|l|l|}
\hline
Acc \% & [A] & [B] & Ours & Acc \% & [A] & [B] & Ours \\ \hline
\hline
Airplane& 92.0 & {\bf 97.3} & 96.0 &Dining tab &63.2 & {\bf 77.8}&67.7 \\ \hline
Bicycle & 74.2 & {\bf 84.2} & 77.1 &Dog& 68.9 & 83.0 & {\bf 87.8} \\ \hline
Bird& 73.0 & 80.8 & {\bf 88.4}& Horse & 78.2 & {\bf 87.5} & 86.0 \\ \hline
Boat & 77.5 & 85.3 & {\bf 85.5}&Motorbike & 81.0 & {\bf 90.1} & 85.1 \\ \hline
Bottle& 54.3 & {\bf 60.8} & 55.8&Person & 91.6 & {\bf 95.0} & 90.9 \\ \hline
Bus& 85.2 & {\bf 89.9} & 85.8& Potted pl & 55.9 & {\bf 57.8} & 52.2\\ \hline
Car&81.9&{\bf 86.8} & 78.6&Sheep & 69.4 & 79.2 & {\bf 83.6} \\ \hline
Cat&76.4 &89.3 & {\bf 91.2}& Sofa & 65.4 & {\bf 73.4} & 61.1 \\ \hline
Chair&65.2&{\bf 75.4}& 65.0&Train & 86.7 & {\bf 94.5} & 91.8 \\ \hline
Cow&63.2&{\bf 77.8}& 74.4& Tv & 77.4 & {\bf 80.7} & 76.1 \\ \hline \hline
Mean & 74.3 & {\bf 82.2} & 79.0 & \# won & 0 & {\bf 15} &5 \\ \hline
\end{tabular}
\vspace*{-3mm}
\caption{PASCAL 2012 classification results, comparing our
Imagenet-pretrained convnet against the leading two methods ([A]=
\cite{cvc} and [B] = \cite{nus}). }
\label{tab:pascal}
\vspace*{-4mm}
\end{center}
\end{table}
\subsection{Feature Analysis} \label{sec:pretrain} We explore how
discriminative the features in each layer of our Imagenet-pretrained
model are. We do this by varying the number of layers retained from
the ImageNet model and place either a linear SVM or softmax classifier
on top. \tab{supretrain} shows results on Caltech-101 and
Caltech-256. For both datasets, a steady improvement can be seen as we
ascend the model, with best results being obtained by using all
layers. This supports the premise that as the feature hierarchies
become deeper, they learn increasingly powerful features.
\begin{table}[h!]
\small
\vspace*{-2mm}
\begin{center}
\tabcolsep=0.11cm
\begin{tabular}{|l|l|l|}
\hline
& Cal-101 & Cal-256 \\
& (30/class) & (60/class) \\
\hline SVM (1) & $44.8 \pm 0.7$ & $24.6 \pm 0.4$ \\
\hline SVM (2) & $66.2 \pm 0.5$ & $39.6 \pm 0.3$ \\
\hline SVM (3) & $72.3 \pm 0.4$ & $46.0 \pm 0.3$ \\
\hline SVM (4) & $76.6 \pm 0.4$ & $51.3 \pm 0.1$ \\
\hline SVM (5) & $\bf{86.2 \pm 0.8}$ & $65.6 \pm 0.3$ \\
\hline SVM (7) & $\bf{85.5 \pm 0.4}$ & $\bf{71.7 \pm 0.2}$ \\
\hline Softmax (5) & $82.9 \pm 0.4$ & $65.7 \pm 0.5$ \\
\hline Softmax (7) & $\bf{85.4 \pm 0.4}$ & $\bf{72.6 \pm 0.1}$ \\
\hline
\end{tabular}
\vspace*{0mm}
\caption{Analysis of the discriminative information contained in each
layer of feature maps within our ImageNet-pretrained convnet. We
train either a linear SVM or softmax on features from different
layers (as indicated in brackets) from the convnet. Higher layers generally
produce more discriminative features.}
\label{tab:supretrain}
\vspace*{-8mm}
\end{center}
\end{table}
\section{Discussion}
We explored large convolutional neural network models, trained for
image classification, in a number ways. First, we presented a novel
way to visualize the activity within the model. This reveals the
features to be far from random, uninterpretable patterns. Rather, they
show many intuitively desirable properties such as compositionality,
increasing invariance and class discrimination as we ascend the
layers. We also showed how these visualization can be used to debug
problems with the model to obtain better results, for example
improving on Krizhevsky \etal's \cite{Kriz12} impressive ImageNet 2012
result. We then demonstrated through a series of occlusion experiments that
the model, while trained for classification, is highly sensitive to
local structure in the image and is not just using broad scene
context. An ablation study on the model revealed that having a minimum
depth to the network, rather than any individual section, is vital to
the model's performance.
Finally, we showed how the ImageNet trained model can generalize
well to other datasets. For Caltech-101 and Caltech-256,
the datasets are similar enough that we can beat the best reported
results, in the latter case by a significant margin.
This result brings into question to utility of benchmarks with small
(i.e.~$<10^4$) training sets. Our convnet model generalized less well
to the PASCAL data, perhaps suffering from dataset bias
\cite{Torralba11}, although it was still within $3.2\%$ of the best reported
result, despite no tuning for the task. For example, our performance
might improve if a different loss function was used that permitted
multiple objects per image. This would naturally enable the networks
to tackle the object detection as well.
\section*{Acknowledgments}
The authors are very grateful for support by NSF grant IIS-1116923, Microsoft Research
and a Sloan Fellowship.
\bibliography{demystify}
\bibliographystyle{icml2013}
\end{document}