File size: 44,199 Bytes
8cf8144 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 |
\documentclass{article}
\usepackage{times}
\usepackage{graphicx} \usepackage{subfigure}
\usepackage{amsmath}
\usepackage{amsfonts}
\usepackage{stfloats}
\usepackage{url}
\usepackage{natbib}
\usepackage{algorithm}
\usepackage{algorithmic}
\usepackage{hyperref}
\newcommand{\theHalgorithm}{\arabic{algorithm}}
\usepackage[accepted]{arXiv}
\graphicspath{ {./figures/} }
\newcommand{\obs}{\text{obs}}
\newcommand{\mis}{\text{mis}}
\newcommand{\qt}[1]{\left<#1\right>}
\newcommand{\ql}[1]{\left[#1\right]}
\newcommand{\hess}{\mathbf{H}}
\newcommand{\jacob}{\mathbf{J}}
\newcommand{\hl}{HL}
\newcommand{\cost}{\mathcal{L}}
\newcommand{\lout}{\mathbf{r}}
\newcommand{\louti}{r}
\newcommand{\outi}{y}
\newcommand{\out}{\mathbf{y}}
\newcommand{\gauss}{\mathbf{G_N}}
\newcommand{\eye}{\mathbf{I}}
\newcommand{\softmax}{\phi}
\newcommand{\targ}{\mathbf{t}}
\newcommand{\metric}{\mathbf{G}}
\newcommand{\sample}{\mathbf{z}}
\newcommand{\bmx}[0]{\begin{bmatrix}}
\newcommand{\emx}[0]{\end{bmatrix}}
\newcommand{\qexp}[1]{\left<#1\right>}
\newcommand{\vect}[1]{\mathbf{#1}}
\newcommand{\vects}[1]{\boldsymbol{#1}}
\newcommand{\matr}[1]{\mathbf{#1}}
\newcommand{\var}[0]{\operatorname{Var}}
\newcommand{\std}[0]{\operatorname{std}}
\newcommand{\cov}[0]{\operatorname{Cov}}
\newcommand{\diag}[0]{\operatorname{diag}}
\newcommand{\matrs}[1]{\boldsymbol{#1}}
\newcommand{\va}[0]{\vect{a}}
\newcommand{\vb}[0]{\vect{b}}
\newcommand{\vc}[0]{\vect{c}}
\newcommand{\ve}[0]{\vect{e}}
\newcommand{\vh}[0]{\vect{h}}
\newcommand{\vv}[0]{\vect{v}}
\newcommand{\vx}[0]{\vect{x}}
\newcommand{\vn}[0]{\vect{n}}
\newcommand{\vz}[0]{\vect{z}}
\newcommand{\vw}[0]{\vect{w}}
\newcommand{\vs}[0]{\vect{s}}
\newcommand{\vf}[0]{\vect{f}}
\newcommand{\vi}[0]{\vect{i}}
\newcommand{\vo}[0]{\vect{o}}
\newcommand{\vy}[0]{\vect{y}}
\newcommand{\vg}[0]{\vect{g}}
\newcommand{\vm}[0]{\vect{m}}
\newcommand{\vu}[0]{\vect{u}}
\newcommand{\vL}[0]{\vect{L}}
\newcommand{\vr}[0]{\vect{r}}
\newcommand{\mW}[0]{\matr{W}}
\newcommand{\mG}[0]{\matr{G}}
\newcommand{\mX}[0]{\matr{X}}
\newcommand{\mQ}[0]{\matr{Q}}
\newcommand{\mU}[0]{\matr{U}}
\newcommand{\mV}[0]{\matr{V}}
\newcommand{\vE}[0]{\matr{E}}
\newcommand{\mA}{\matr{A}}
\newcommand{\mD}{\matr{D}}
\newcommand{\mS}{\matr{S}}
\newcommand{\mI}{\matr{I}}
\newcommand{\td}[0]{\text{d}}
\newcommand{\TT}[0]{\vects{\theta}}
\newcommand{\vsig}[0]{\vects{\sigma}}
\newcommand{\valpha}[0]{\vects{\alpha}}
\newcommand{\vmu}[0]{\vects{\mu}}
\newcommand{\vzero}[0]{\vect{0}}
\newcommand{\tf}[0]{\text{m}}
\newcommand{\tdf}[0]{\text{dm}}
\newcommand{\grad}[0]{\nabla}
\newcommand{\alert}[1]{\textcolor{red}{#1}}
\newcommand{\N}[0]{\mathcal{N}}
\newcommand{\LL}[0]{\mathcal{L}}
\newcommand{\HH}[0]{\mathcal{H}}
\newcommand{\RR}[0]{\mathbb{R}}
\newcommand{\II}[0]{\mathbb{I}}
\newcommand{\Scal}[0]{\mathcal{S}}
\newcommand{\sigmoid}{\sigma}
\newcommand{\E}[0]{\mathbb{E}}
\newcommand{\enabla}[0]{\ensuremath{\overset{\raisebox{-0.3ex}[0.5ex][0ex]{\ensuremath{\scriptscriptstyle e}}}{\nabla}}}
\newcommand{\enhnabla}[0]{\nabla_{\hspace{-0.5mm}e}\,}
\newcommand{\tred}[1]{\textcolor{red}{#1}}
\newcommand{\tgreen}[1]{\textcolor{green}{#1}}
\newcommand{\tblue}[1]{\textcolor{blue}{#1}}
\newcommand{\todo}[1]{{\Large\textcolor{red}{#1}}}
\newcommand{\done}[1]{{\Large\textcolor{green}{#1}}}
\newcommand{\dd}[1]{\ensuremath{\mbox{d}#1}}
\DeclareMathOperator*{\argmax}{\arg \max}
\DeclareMathOperator*{\argmin}{\arg \min}
\newcommand{\newln}{\\&\quad\quad{}}
\newcommand{\Ax}{\mathcal{A}_x}
\newcommand{\Ay}{\mathcal{A}_y}
\newcommand{\ola}{\overleftarrow}
\newcommand{\ora}{\overrightarrow}
\newcommand{\ov}{\overline}
\newcommand{\ts}{\rule{0pt}{2.6ex}} \newcommand{\ms}{\rule{0pt}{0ex}} \newcommand{\bs}{\rule[-1.2ex]{0pt}{0pt}} \newcommand{\specialcell}[2][c]{\begin{tabular}[#1]{@{}c@{}}#2\end{tabular}}
\newcommand{\bx}{\textbf{x}}
\newcommand{\by}{\textbf{y}}
\newcommand{\bW}{\textbf{W}}
\icmltitlerunning{Neural Image Caption Generation with Visual Attention}
\begin{document}
\twocolumn[
\icmltitle{Show, Attend and Tell: Neural Image Caption\\Generation with Visual Attention}
\icmlauthor{Kelvin Xu}{kelvin.xu@umontreal.ca}
\icmlauthor{Jimmy Lei Ba}{jimmy@psi.utoronto.ca}
\icmlauthor{Ryan Kiros}{rkiros@cs.toronto.edu}
\icmlauthor{Kyunghyun Cho}{kyunghyun.cho@umontreal.ca}
\icmlauthor{Aaron Courville}{aaron.courville@umontreal.ca}
\icmlauthor{Ruslan Salakhutdinov}{rsalakhu@cs.toronto.edu}
\icmlauthor{Richard S. Zemel}{zemel@cs.toronto.edu}
\icmlauthor{Yoshua Bengio}{find-me@the.web}
\vskip 0.3in
]
\begin{abstract}
Inspired by recent work in machine translation and object detection, we
introduce an attention based model that automatically learns to describe the
content of images. We describe how we can train this model in a deterministic
manner using standard backpropagation techniques and stochastically by
maximizing a variational lower bound. We also show through visualization how
the model is able to automatically learn to fix its gaze on salient objects
while generating the corresponding words in the output sequence.
We validate the use of attention with state-of-the-art performance on three
benchmark datasets: Flickr8k, Flickr30k and MS COCO.
\end{abstract}
\section{Introduction}
Automatically generating captions of an image is a task very close to the heart
of scene understanding --- one of the primary goals of computer vision. Not only
must caption generation models be powerful enough to solve the computer vision
challenges of determining which objects are in an image, but they must also be
capable of capturing and expressing their relationships in a natural language.
For this reason, caption generation has long been viewed as a difficult
problem. It is a very important challenge for machine learning
algorithms, as it amounts to mimicking the remarkable human ability to compress
huge amounts of salient visual infomation into descriptive language.
Despite the challenging nature of this task, there has been a
recent surge of research interest in attacking the image caption generation
problem. Aided by advances in training neural networks \citep{Krizhevsky2012}
and large classification datasets \citep{Imagenet14}, recent work has
significantly improved the quality of caption generation using a combination of
convolutional neural networks (convnets) to obtain vectorial representation of images and
recurrent neural networks to decode those representations into natural language
sentences (see Sec.~\ref{section:background}).
\begin{figure}[tp]
\label{figure:model_diagram}
\centering
\caption{Our model learns a words/image alignment. The visualized
attentional maps (3) are explained in section \ref{section:model} \& \ref{section:viz}}
\vspace{3mm}
\includegraphics[width=\columnwidth]{model_diagram.pdf}
\vspace{-6mm}
\end{figure}
\begin{figure*}[!tp]
\label{figure:attention_diagram}
\centering
\caption{Attention over time. As the model generates each word, its attention changes to reflect the relevant parts of the image. ``soft''
(top row) vs ``hard'' (bottom row) attention. (Note that both models generated the same captions in this example.) }
\includegraphics[width=6.5in]{runout}
\vspace{-5mm}
\end{figure*}
\begin{figure*}[!tp]
\label{figure:alignments}
\centering
\caption{Examples of attending to the correct object (\textit{white} indicates the attended regions,
\textit{underlines} indicated the corresponding word)}
\includegraphics[width=0.87\textwidth]{good.pdf}
\vspace{-5mm}
\end{figure*}
One of the most curious facets of the human visual system is the presence of
attention \citep{Rensink2000,Corbetta2002}. Rather than compress an entire
image into a static representation, attention allows for salient features to
dynamically come to the forefront as needed. This is especially important when
there is a lot of clutter in an image. Using representations (such as those
from the top layer of a convnet) that distill information in image down to the
most salient objects is one effective solution that has been widely adopted in
previous work. Unfortunately, this has one potential drawback of losing
information which could be useful for richer, more descriptive captions. Using
more low-level representation can help preserve this information. However
working with these features necessitates a powerful mechanism to steer the
model to information important to the task at hand.
In this paper, we describe approaches to caption generation that attempt to
incorporate a form of attention with two variants: a ``hard'' attention
mechanism and a ``soft'' attention mechanism. We also show how one advantage of
including attention is the ability to visualize what the model ``sees''.
Encouraged by recent advances in caption generation and inspired by recent
success in employing attention in machine translation \citep{Bahdanau2014} and
object recognition \citep{Ba2014,Mnih2014}, we investigate models that can
attend to salient part of an image while generating its caption.
The contributions of this paper are the following:
\vspace{- 1mm}
\begin{itemize}
\vspace{-2mm}
\item We introduce two attention-based image caption generators under
a common framework (Sec.~\ref{section:model}): 1) a ``soft'' deterministic attention mechanism trainable by standard
back-propagation methods and 2) a ``hard'' stochastic attention mechanism trainable by
maximizing an approximate variational lower bound or equivalently by
REINFORCE~\citep{Williams92}.
\vspace{-2mm}
\item We show how we can gain insight and interpret the results of
this framework by visualizing ``where'' and ``what'' the attention focused on.
(see Sec.~\ref{section:viz})
\vspace{-2mm}
\item Finally, we quantitatively validate the usefulness of attention in
caption generation with state of the art performance
(Sec.~\ref{section:results}) on three benchmark datasets: Flickr8k
\citep{Hodosh2013} , Flickr30k \citep{Young2014} and the MS COCO dataset
\citep{Lin2014}.
\end{itemize}
\section{Related Work}
\label{section:background}
In this section we provide relevant background on previous work on
image caption generation and attention.
Recently, several methods have been proposed for generating image descriptions.
Many of these methods are based on recurrent neural networks and inspired
by the successful use of sequence to sequence training with neural networks for
machine translation~\citep{Cho2014,Bahdanau2014,Sutskever2014}. One
major reason image caption generation is well suited to the encoder-decoder framework
\citep{Cho2014} of machine translation is because it is analogous
to ``translating'' an image to a sentence.
The first approach to use neural networks for caption generation was
\citet{Kiros2014a}, who proposed a multimodal log-bilinear model that was
biased by features from the image. This work was later followed by
\citet{Kiros2014b} whose method was designed to explicitly allow a natural way
of doing both ranking and generation. \citet{Mao2014} took a similar approach
to generation but replaced a feed-forward neural language model with a
recurrent one. Both \citet{Vinyals2014} and \citet{Donahue2014} use LSTM RNNs
for their models. Unlike \citet{Kiros2014a} and \citet{Mao2014} whose models
see the image at each time step of the output word sequence,
\citet{Vinyals2014} only show the image to the RNN at the beginning. Along
with images, \citet{Donahue2014} also apply LSTMs to videos, allowing
their model to generate video descriptions.
All of these works represent images as a single feature vector from the top
layer of a pre-trained convolutional network. \citet{Karpathy2014} instead
proposed to learn a joint embedding space for ranking and generation whose
model learns to score sentence and image similarity as a function of R-CNN
object detections with outputs of a bidirectional RNN.
\citet{Fang2014} proposed a three-step pipeline for
generation by incorporating
object detections.
Their model first learn detectors for several visual concepts
based on a multi-instance learning framework. A language model trained on
captions was then applied to the detector outputs, followed by rescoring from a
joint image-text embedding space. Unlike these models, our proposed attention
framework does not explicitly use object detectors but instead learns latent
alignments from scratch. This allows our model to go beyond ``objectness'' and
learn to attend to abstract concepts.
Prior to the use of neural networks for generating captions, two main
approaches were dominant. The first involved generating caption templates which
were filled in based on the results of object detections and attribute
discovery (\citet{Kulkarni2013}, \citet{Li2011},
\citet{Yang2011}, \citet{Mitchell2012}, \citet{Elliott2013}). The second
approach was based on first retrieving similar captioned images from a large
database then modifying these retrieved captions to fit the query
\citep{Kuznetsova2012,Kuznetsova2014}. These approaches typically involved an
intermediate ``generalization'' step to remove the specifics of a caption that
are only relevant to the retrieved image, such as the name of a city. Both of
these approaches have since fallen out of favour to the now dominant neural
network methods.
There has been a long line of previous work incorporating attention into neural
networks for vision related tasks. Some that share the same spirit as our work
include \citet{Larochelle2010,Denil2012,Tang2014}. In particular however, our
work directly extends the work of \citet{Bahdanau2014,Mnih2014,Ba2014}.
\section{Image Caption Generation with Attention Mechanism}
\subsection{Model Details}
\label{section:model}
In this section, we describe the two variants of our attention-based model by
first describing their common framework. The main difference is the definition
of the $\phi$ function which we describe in detail in Section
\ref{sec:det_sto}. We denote vectors with bolded font and matrices with
capital letters. In our description below, we suppress bias terms for
readability.
\subsubsection{Encoder: Convolutional Features}
Our model takes a single raw image and generates a caption $\vy$
encoded as a sequence of 1-of-$K$ encoded words.
\[
y = \left\{\vy_1, \ldots, \vy_{C} \right\},\mbox{ } \vy_i \in \RR^{K}
\]
where $K$ is the size of the vocabulary and $C$ is the length of the caption.
We use a convolutional neural network in order to extract a set of feature
vectors which we refer to as annotation vectors. The extractor produces $L$
vectors, each of which is a D-dimensional representation corresponding to a
part of the image.
\begin{align*}
a = \left\{\va_1, \ldots, \va_L \right\},\mbox{ } \va_i \in \RR^{D}
\end{align*}
In order to obtain a correspondence between the feature vectors and portions of
the 2-D image, we extract features from a lower convolutional layer unlike
previous work which instead used a fully connected layer. This allows the
decoder to selectively focus on certain parts of an image by selecting a subset
of all the feature vectors.
\begin{figure}[tp]
\vskip 0.2in
\begin{center}
\centerline{\includegraphics[width=\columnwidth]{lstm_2.pdf}}
\caption{A LSTM cell, lines with bolded squares imply projections with a
learnt weight vector. Each cell learns how to weigh
its input components (input gate), while learning how to modulate that
contribution to the memory (input modulator). It also learns
weights which erase the memory cell (forget gate), and weights
which control how this memory should be emitted (output gate).}
\label{figure:conditional_lstm}
\end{center}
\vskip -0.3 in
\end{figure}
\subsubsection{Decoder: Long Short-Term Memory Network}
We use a long short-term memory (LSTM)
network~\citep{Hochreiter+Schmidhuber-1997} that produces a caption by
generating one word at every time step conditioned on a context vector, the
previous hidden state and the previously generated words. Our implementation of
LSTM closely follows the one used in \citet{Zaremba2014} (see Fig.~\ref{figure:conditional_lstm}). Using $T_{s,t}:
\RR^{s} \rightarrow \RR^{t}$ to denote a simple affine transformation with
parameters that are learned,
\begin{align}
\label{eq:lstm_gates}
\begin{pmatrix}
\vi_t \\
\vf_t \\
\vo_t \\
\vg_t \\
\end{pmatrix} =
&
\begin{pmatrix}
\sigmoid \\
\sigmoid \\
\sigmoid \\
\tanh \\
\end{pmatrix}
T_{D+m+n, n}
\begin{pmatrix}
\vE\vy_{t-1}\\
\vh_{t-1}\\
\hat{\vz_t}\\
\end{pmatrix}
\\
\label{eq:lstm_memory}
\vc_t &= \vf_t \odot \vc_{t-1} + \vi_t \odot \vg_t \\
\label{eq:lstm_hidden}
\vh_t &= \vo_t \odot \tanh (\vc_{t}).
\end{align}
Here, $\vi_t$, $\vf_t$, $\vc_t$, $\vo_t$, $\vh_t$ are the input, forget, memory, output
and hidden state of the LSTM, respectively. The vector $\hat{\vz} \in \RR^{D}$ is the context
vector, capturing the visual information associated with a particular
input location, as explained below. $\vE\in\RR^{m\times K}$ is an embedding matrix. Let $m$ and $n$
denote the embedding and LSTM dimensionality respectively and $\sigma$ and
$\odot$ be the logistic sigmoid activation and element-wise multiplication
respectively.
In simple terms, the context vector $\hat{\vz}_t$ (equations~\eqref{eq:lstm_gates}--\eqref{eq:lstm_hidden})
is a dynamic representation of the relevant part of the image input at time $t$.
We define a mechanism $\phi$ that computes $\hat{\vz}_t$ from the annotation vectors $\va_i, i=1,\ldots,L$
corresponding to the features extracted at different image locations. For
each location $i$, the
mechanism generates a positive weight $\alpha_i$ which can
be interpreted either as the probability that location $i$ is the right place to focus
for producing the next word (the ``hard'' but stochastic attention mechanism),
or as the relative importance to give to location $i$ in blending the $a_i$'s together.
The weight $\alpha_i$ of each annotation vector $a_i$ is computed by an
\emph{attention model} $f_{\mbox{att}}$ for which we use a multilayer perceptron conditioned on the previous hidden state $h_{t-1}$.
The soft version of this attention mechanism was introduced by~\citet{Bahdanau2014}.
For emphasis, we note that the hidden state varies as the output RNN advances in
its output sequence: ``where'' the network looks next depends on the sequence of words that has already been
generated.
\begin{align}
e_{ti} =& f_{\mbox{att}} (\va_i, \vh_{t-1}) \\
\label{eq:alpha}
\alpha_{ti} =& \frac{\exp(e_{ti})}{\sum_{k=1}^L \exp(e_{tk})}.
\end{align}
Once the weights (which sum to one) are computed, the context vector $\hat{z}_t$
is computed by
\begin{align}
\label{eq:context}
\hat{\vz}_t = \phi\left( \left\{ \va_i \right\}, \left\{ \alpha_i \right\}
\right),
\end{align}
where $\phi$ is a function that returns a single vector given the set of
annotation vectors and their corresponding weights. The details of $\phi$ function
are discussed in Sec.~\ref{sec:det_sto}.
The initial memory state and hidden state of the LSTM are predicted by an
average of the annotation vectors fed through two separate MLPs
($\text{init,c}$ and $\text{init,h}$):
\begin{align}
\vc_0 = f_{\text{init,c}} (\frac{1}{L} \sum_i^L \va_i) \nonumber \\
\vh_0 = f_{\text{init,h}} (\frac{1}{L} \sum_i^L \va_i) \nonumber
\end{align}
In this work, we use a deep output layer~\citep{Pascanu2014} to compute the
output word probability given the LSTM state, the context vector and the
previous word:
\begin{gather}
\label{eq:p-out}
p(\vy_t | \va, \vy_1^{t-1}) \propto \exp(\vL_o(\vE\vy_{t-1} + \vL_h\vh_t+ \vL_z \hat{\vz}_t))
\end{gather}
Where $\vL_o\in\RR^{K\times m}$, $\vL_h\in\RR^{m\times n}$, $\vL_z\in\RR^{m\times D}$, and
$\vE$ are learned parameters initialized randomly.
\section{Learning Stochastic ``Hard'' vs Deterministic ``Soft'' Attention}
\label{sec:det_sto}
In this section we discuss two alternative mechanisms for the attention
model $f_{\mbox{att}}$: stochastic attention and deterministic attention.
\begin{figure*}[!tp]
\label{fig:second}
\caption{Examples of mistakes where we can use attention to gain intuition into what the model saw.}
\includegraphics[width=1.02\textwidth]{errors.pdf}
\label{fig:subfigures}
\end{figure*}
\subsection{Stochastic ``Hard'' Attention}
\label{sec:sto_attn}
We represent the location variable $s_t$ as where the model decides to focus attention
when generating the $t^{th}$ word. $s_{t,i}$ is an indicator one-hot variable which is set to 1
if the $i$-th location (out of $L$)
is the one used to extract visual features. By treating
the attention locations as intermediate latent variables, we can
assign a multinoulli distribution parametrized by $\{\alpha_i\}$,
and view $\hat{z}_t$ as a random variable:
\begin{align}
\label{eq:s_dist}
p(&s_{t,i} = 1 \mid s_{j<t}, \va ) = \alpha_{t,i} \\
\label{eq:hard_context}
\hat{\vz}_t &= \sum_i {s}_{t,i}\va_{i}.
\end{align}
We define a new objective function $L_s$ that is a variational lower bound on the marginal log-likelihood
$\log p( \vy \mid \va)$ of observing the sequence of words $\vy$ given image features $\va$.
The learning algorithm for the parameters $W$ of the models can be derived by
directly optimizing $L_s$:
\begin{align}
L_s = & \sum_s p(s \mid \va) \log p(\vy \mid s, \va) \nonumber \\
\leq & \log \sum_{s}p(s \mid \va)p( \vy \mid s, \va) \nonumber \\
= & \log p( \vy \mid \va) \label{eq:mll}
\end{align}
\begin{multline}
\label{eq:lb_gradient}
\frac{\partial L_s}{\partial W} = \sum_s p(s \mid \va) \bigg[ \frac{\partial \log p(\vy \mid s, \va)}{\partial W} + \\
\log p(\vy \mid s, \va) \frac{\partial \log p(s \mid \va)}{\partial W} \bigg].
\end{multline}
Equation \ref{eq:lb_gradient} suggests a Monte Carlo based sampling
approximation of the gradient with respect to the model parameters. This can be
done by sampling the location $s_t$ from a multinouilli distribution defined by
Equation \ref{eq:s_dist}.
\begin{gather}
\tilde{s_t} \sim \text{Multinoulli}_L(\{\alpha_i\}) \nonumber
\end{gather}
\begin{multline}
\frac{\partial L_s}{\partial W} \approx \frac{1}{N} \sum_{n=1}^{N} \bigg[ \frac{\partial \log p(\vy \mid \tilde{s}^n, \va)}{\partial W} + \\
\log p(\vy \mid \tilde{s}^n, \va) \frac{\partial \log p(\tilde{s}^n \mid \va)}{\partial W} \bigg]
\end{multline}
A moving average baseline is used to reduce the variance in the
Monte Carlo estimator of the gradient, following~\citet{Weaver+Tao-UAI2001}.
Similar, but more complicated variance
reduction techniques have previously been used by \citet{Mnih2014} and
\citet{Ba2014}. Upon seeing the $k^{th}$ mini-batch, the moving average baseline
is estimated as an accumulated sum of the previous log likelihoods with exponential decay:
\begin{align}
b_k = 0.9 \times b_{k-1} + 0.1 \times \log p(\vy \mid \tilde{s}_k , \va) \nonumber
\end{align}
To further reduce the estimator variance, an entropy term on the multinouilli
distribution $H[s]$ is added. Also, with probability 0.5 for a given image, we
set the sampled attention location $\tilde{s}$ to its expected value $\alpha$.
Both techniques improve the robustness of the stochastic attention learning algorithm.
The final learning rule for the model is then the following:
\begin{multline}
\frac{\partial L_s}{\partial W} \approx \frac{1}{N} \sum_{n=1}^{N} \bigg[ \frac{\partial \log p(\vy \mid \tilde{s}^n, \va)}{\partial W} + \\
\lambda_r( \log p(\vy \mid \tilde{s}^n, \va) - b) \frac{\partial \log p(\tilde{s}^n \mid \va)}{\partial W} + \lambda_e\frac{\partial H[\tilde{s}^n]}{\partial W} \bigg]
\nonumber
\end{multline}
where, $\lambda_r$ and $\lambda_e$ are two hyper-parameters set by cross-validation.
As pointed out and used in \citet{Ba2014} and \citet{Mnih2014}, this is
formulation is equivalent to the REINFORCE learning rule~\citep{Williams92}, where the
reward for the attention choosing a sequence of actions is a real value proportional to the log likelihood
of the target sentence under the sampled attention trajectory.
In making a hard choice at every point, $\phi\left( \left\{ \va_i \right\},
\left\{ \alpha_i \right\}\right)$ from Equation \ref{eq:context} is a function that
returns a sampled $\va_i$ at every point in time based upon a multinouilli
distribution parameterized by $\alpha$.
\subsection{Deterministic ``Soft'' Attention}
\label{sec:det_attn}
Learning stochastic attention requires sampling the attention location $s_t$
each time, instead we can take the expectation of the context vector
$\hat{\vz}_t$ directly,
\begin{align}
\label{eq:s_dist_soft}
\mathbb{E}_{p(s_t|a)}[\hat{\vz}_t] = \sum_{i=1}^L \alpha_{t,i} \va_{i}
\end{align}
and formulate a deterministic attention model by computing a soft
attention weighted annotation vector $\phi\left( \left\{ \va_i \right\},
\left\{ \alpha_i \right\}\right) = \sum_i^L \alpha_i \va_i$ as introduced
by \citet{Bahdanau2014}. This corresponds to feeding in a
soft $\alpha$ weighted context into the system. The whole model is smooth and
differentiable under the deterministic attention, so learning end-to-end is
trivial by using standard back-propagation.
Learning the deterministic attention can also be understood as approximately
optimizing the marginal likelihood in Equation \ref{eq:mll} under the attention
location random variable $s_t$ from Sec.~\ref{sec:sto_attn}. The hidden activation of
LSTM $\vh_t$ is a linear projection of the stochastic context vector $\hat{\vz}_t$
followed by $\tanh$ non-linearity. To the first order Taylor approximation, the
expected value $\mathbb{E}_{p(s_t|a)}[\vh_t]$ is equal to computing $\vh_t$
using a single forward prop with the expected context vector
$\mathbb{E}_{p(s_t|a)}[\hat{\vz}_t]$. Considering Eq.~\ref{eq:p-out},
let $\vn_t = \vL_o(\vE\vy_{t-1} + \vL_h\vh_t+ \vL_z \hat{\vz}_t)$,
$\vn_{t,i}$ denotes $\vn_t$ computed by setting the random variable $\hat{\vz}$ value to $\va_i$.
We define the normalized weighted geometric mean for the softmax $k^{th}$ word
prediction:
\begin{align}
NWGM[p(y_t=k \mid \va)] &= {\prod_i \exp(n_{t,k,i})^{p(s_{t,i}=1|a)} \over \sum_j \prod_i \exp(n_{t,j,i})^{p(s_{t,i}=1|a)}} \nonumber \\
&= {\exp(\mathbb{E}_{p(s_{t}|a)}[n_{t,k}]) \over \sum_j \exp(\mathbb{E}_{p(s_t|a)}[n_{t,j}])} \nonumber
\end{align}
The equation above shows the normalized weighted geometric mean of the caption
prediction can be approximated well by using the expected context vector, where
$\mathbb{E}[\vn_t] = \vL_o(\vE\vy_{t-1} + \vL_h\mathbb{E}[\vh_t]+ \vL_z \mathbb{E}[\hat{\vz}_t])$.
It shows that the NWGM of a softmax unit is obtained by applying softmax to the
expectations of the underlying linear projections. Also, from the results in
\cite{baldi2014dropout}, $NWGM[p(y_t=k \mid \va)] \approx \mathbb{E}[p(y_t=k \mid \va)]$
under softmax activation. That means the expectation of the outputs over all
possible attention locations induced by random variable $s_t$ is computed by
simple feedforward propagation with expected context vector
$\mathbb{E}[\hat{\vz}_t]$. In other words, the deterministic attention model
is an approximation to the marginal likelihood over the attention locations.
\subsubsection{Doubly Stochastic Attention}
\label{section:ds_attn}
By construction, $\sum_i \alpha_{ti} = 1$ as they are the output of a softmax.
In training the deterministic version of our model we introduce a form of
doubly stochastic regularization, where we also encourage $\sum_t \alpha_{ti} \approx
1$. This can be interpreted as encouraging the model to pay equal attention to
every part of the image over the course of generation. In our experiments, we
observed that this penalty was important quantitatively to improving overall
BLEU score and that qualitatively this leads to more rich and descriptive
captions. In addition, the soft attention model predicts a gating scalar $\beta$
from previous hidden state $\vh_{t-1}$ at each time step $t$, such that, $\phi\left( \left\{ \va_i \right\}, \left\{ \alpha_i \right\}\right) = \beta \sum_i^L \alpha_i \va_i$, where $\beta_t = \sigma(f_{\beta}(\vh_{t-1}))$. We notice
our attention weights put more emphasis on the objects in the images by
including the scalar $\beta$.
\begin{table*}[!tph]
\label{table:results}
\caption{BLEU-{1,2,3,4}/METEOR metrics compared to other methods, $\dagger$ indicates a different split, (---) indicates an unknown metric, $\circ$ indicates the authors
kindly provided missing metrics by personal communication, $\Sigma$ indicates an ensemble, $a$ indicates using
AlexNet}
\vskip 0.15in
\centering
\begin{tabular}{cccccccc}
\cline{3-6}
& \multicolumn{1}{l|}{} & \multicolumn{4}{c|}{\bf BLEU} & & \\ \cline{1-7}
\multicolumn{1}{|c|}{Dataset} & \multicolumn{1}{c|}{Model} & \multicolumn{1}{c|}{BLEU-1} & \multicolumn{1}{c|}{BLEU-2} & \multicolumn{1}{c|}{BLEU-3} & \multicolumn{1}{c|}{BLEU-4} & \multicolumn{1}{c|}{METEOR} & \\ \cline{1-7}
\multicolumn{1}{|c|}{Flickr8k} & \specialcell{ Google NIC\citep{Vinyals2014}$^\dagger$$^\Sigma$ \\ Log Bilinear \citep{Kiros2014a}$^\circ$ \\ Soft-Attention \\ Hard-Attention }
& \specialcell{63 \\ 65.6 \\ \bf 67 \\ \bf 67 }
& \specialcell{41 \\ 42.4 \\ 44.8 \\ \bf 45.7 }
& \specialcell{27 \\ 27.7 \\ 29.9 \\ \bf 31.4 }
& \specialcell{ --- \\ 17.7 \\ 19.5 \\ \bf 21.3 }
& \multicolumn{1}{c|}{\specialcell{ --- \\ 17.31 \\ 18.93 \\ \bf 20.30 }} \ts\\
\cline{1-7}
\multicolumn{1}{|c|}{Flickr30k} & \specialcell{ Google NIC$^\dagger$$^\circ$$^\Sigma$ \\Log Bilinear\\ Soft-Attention \\ Hard-Attention }
& \specialcell{66.3 \\ 60.0 \\ 66.7 \\ \bf 66.9 }
& \specialcell{42.3 \\ 38 \\ 43.4 \\ \bf 43.9 }
& \specialcell{27.7 \\ 25.4 \\ 28.8 \\ \bf 29.6 }
& \specialcell{18.3 \\ 17.1 \\ 19.1 \\ \bf 19.9 }
& \multicolumn{1}{c|}{\specialcell{ --- \\ 16.88 \\ \bf 18.49 \\ 18.46}} \ts\\
\cline{1-7}
\multicolumn{1}{|c|}{COCO} & \specialcell{CMU/MS Research \citep{Chen2014}$^a$ \\ MS Research \citep{Fang2014}$^\dagger$$^a$ \\ BRNN \citep{Karpathy2014}$^\circ$ \\Google NIC$^\dagger$$^\circ$$^\Sigma$ \\ Log Bilinear$^\circ$ \\ Soft-Attention \\ Hard-Attention }
& \specialcell{--- \\ --- \\ 64.2 \\ 66.6 \\ 70.8 \\ 70.7 \\ \bf 71.8 }
& \specialcell{--- \\ --- \\ 45.1 \\ 46.1 \\ 48.9 \\ 49.2 \\ \bf 50.4 }
& \specialcell{--- \\ --- \\ 30.4 \\ 32.9 \\ 34.4 \\ 34.4 \\ \bf 35.7 }
& \specialcell{--- \\ --- \\ 20.3 \\ 24.6 \\ 24.3 \\ 24.3 \\ \bf 25.0 }
& \multicolumn{1}{c|}{\specialcell{ 20.41 \\ 20.71 \\ --- \\ --- \\ 20.03 \\ \bf 23.90 \\ 23.04 }} \ts\\
\cline{1-7}
\end{tabular}
\vskip -0.1in
\end{table*}
Concretely, the model is trained end-to-end by minimizing the following penalized negative
log-likelihood:
\begin{gather}
L_d = -\log(P(\textbf{y}|\textbf{x})) + \lambda \sum_i^{L}(1 - \sum_t^C \alpha_{ti})^2
\end{gather}
\subsection{Training Procedure}
Both variants of our attention model were trained with stochastic gradient
descent using adaptive learning rate algorithms. For the Flickr8k dataset, we
found that RMSProp \citep{Tieleman2012} worked best, while for Flickr30k/MS
COCO dataset we used the recently proposed Adam algorithm~\citep{Kingma2014} .
To create the annotations $a_i$ used by our decoder, we used the Oxford VGGnet
\citep{Simonyan14} pre-trained on ImageNet without finetuning. In principle however,
any encoding function could be used. In addition, with enough data, we could also train
the encoder from scratch (or fine-tune) with the rest of the model.
In our experiments we use the 14$\times$14$\times$512 feature map
of the fourth convolutional layer before max pooling. This means our
decoder operates on the flattened 196 $\times$ 512 (i.e $L\times D$) encoding.
As our implementation requires time proportional to the length of the longest
sentence per update, we found training on a random group of captions to be
computationally wasteful. To mitigate this problem, in preprocessing we build a
dictionary mapping the length of a sentence to the corresponding subset of
captions. Then, during training we randomly sample a length and retrieve a
mini-batch of size 64 of that length. We found that this greatly improved
convergence speed with no noticeable diminishment in performance. On our
largest dataset (MS COCO), our soft attention model took less than 3 days to
train on an NVIDIA Titan Black GPU.
In addition to dropout~\citep{Srivastava14}, the only other regularization
strategy we used was early stopping on BLEU score. We observed a breakdown in
correlation between the validation set log-likelihood and BLEU in the later stages
of training during our experiments. Since BLEU is the most commonly reported
metric, we used BLEU on our validation set for model selection.
In our experiments with
soft attention, we also used Whetlab\footnote{\url{https://www.whetlab.com/}} \citep{Snoek2012,Snoek2014} in our
Flickr8k experiments. Some of the intuitions we gained from hyperparameter
regions it explored were especially important in our Flickr30k and COCO experiments.
We make our code for these models based in Theano \citep{Bergstra2010}
publicly available upon publication to encourage future research in this
area.
\section{Experiments}
We describe our experimental methodology and quantitative results which validate the effectiveness
of our model for caption generation.
\subsection{Data}
We report results on the popular Flickr8k and Flickr30k dataset which has 8,000
and 30,000 images respectively as well as the more challenging Microsoft COCO
dataset which has 82,783 images. The Flickr8k/Flickr30k dataset both come with
5 reference sentences per image, but for the MS COCO dataset, some of the
images have references in excess of 5 which for consistency across
our datasets we discard. We applied only basic tokenization to MS COCO so that it is
consistent with the tokenization present in Flickr8k and Flickr30k. For all our
experiments, we used a fixed vocabulary size of 10,000.
Results for our attention-based architecture are reported in Table
\ref{table:results}. We report results with the frequently used BLEU
metric\footnote{We verified that our BLEU evaluation code matches the authors of
\citet{Vinyals2014}, \citet{Karpathy2014} and \citet{Kiros2014b}. For fairness, we only compare against
results for which we have verified that our BLEU evaluation code is the same. With
the upcoming release of the COCO evaluation server, we will include comparison
results with all other recent image captioning models.} which is the standard in the
caption generation literature. We report BLEU from 1 to 4 without a brevity
penalty. There has been, however, criticism of BLEU, so in addition we report
another common metric METEOR \citep{Denkowski2014}, and compare whenever
possible.
\subsection{Evaluation Procedures}
A few challenges exist for comparison, which we explain here. The first is a
difference in choice of convolutional feature extractor. For identical decoder
architectures, using more recent architectures such as GoogLeNet or Oxford VGG
\citet{Szegedy2014}, \citet{Simonyan14} can give a boost in performance over
using the AlexNet \citep{Krizhevsky2012}. In our evaluation, we compare directly
only with results which use the comparable GoogLeNet/Oxford VGG features, but
for METEOR comparison we note some results that use AlexNet.
The second challenge is a single model versus ensemble comparison. While
other methods have reported performance boosts by using ensembling, in our
results we report a single model performance.
Finally, there is challenge due to differences between dataset splits. In our
reported results, we use the pre-defined splits of Flickr8k. However, one
challenge for the Flickr30k and COCO datasets is the lack of standardized
splits. As a result, we report with the publicly available
splits\footnote{\url{http://cs.stanford.edu/people/karpathy/deepimagesent/}}
used in previous work \citep{Karpathy2014}. In our experience, differences in
splits do not make a substantial difference in overall performance, but
we note the differences where they exist.
\subsection{Quantitative Analysis}
\label{section:results}
In Table \ref{table:results}, we provide a summary of the experiment validating
the quantitative effectiveness of attention. We obtain state of the art
performance on the Flickr8k, Flickr30k and MS COCO. In addition, we note
that in our experiments we are able to significantly improve the state of the
art performance METEOR on MS COCO that we speculate is connected to
some of the regularization techniques we used \ref{section:ds_attn} and our
lower level representation. Finally, we also note that we are
able to obtain this performance using a single model without an ensemble.
\subsection{Qualitative Analysis: Learning to attend}
\label{section:viz}
By visualizing the attention component learned by the model, we are able to add
an extra layer of interpretability to the output of the model
(see Fig. 1). Other systems that have done this rely on object
detection systems to produce candidate alignment targets~\citep{Karpathy2014}.
Our approach is much more flexible, since the model can attend to ``non
object'' salient regions.
The 19-layer OxfordNet uses stacks of 3x3 filters meaning the only time the
feature maps decrease in size are due to the max pooling layers. The input
image is resized so that the shortest side is 256 dimensional with preserved
aspect ratio. The input to the convolutional network is the center cropped
224x224 image. Consequently, with 4 max pooling layers we get an output
dimension of the top convolutional layer of 14x14. Thus in order to visualize
the attention weights for the soft model, we simply upsample the weights by a
factor of $2^4$ = 16 and apply a Gaussian filter. We note that the receptive
fields of each of the 14x14 units are highly overlapping.
As we can see in Figure 2 and 3, the model learns alignments
that correspond very strongly with human intuition. Especially in the examples of
mistakes, we see that it is possible to exploit such visualizations to get an
intuition as to why those mistakes were made. We provide a more extensive list of
visualizations in Appendix \ref{section:appendix} for the reader.
\section{Conclusion}
We propose an attention based approach that gives state of the art performance
on three benchmark datasets using the BLEU and METEOR metric. We also show how the
learned attention can be exploited to give more interpretability into the
models generation process, and demonstrate that the learned alignments
correspond very well to human intuition. We hope that the results of this paper
will encourage future work in using visual attention. We also expect that the
modularity of the encoder-decoder approach combined with attention to have
useful applications in other domains.
\section*{Acknowledgments}
\label{sec:ack}
The authors would like to thank the developers of Theano~\citep{Bergstra2010,
Bastien2012}. We acknowledge the support of the following
organizations for research funding and computing support: the Nuance Foundation, NSERC, Samsung,
Calcul Qu\'{e}bec, Compute Canada, the Canada Research Chairs and CIFAR. The authors
would also like to thank Nitish Srivastava for assistance with his ConvNet
package as well as preparing the Oxford convolutional network and Relu Patrascu
for helping with numerous infrastructure related problems.
\small
\bibliography{capgen}
\bibliographystyle{icml2015}
\newpage
\appendix
\normalsize
\onecolumn
\section{Appendix}
\label{section:appendix}
Visualizations from our ``hard'' (a) and ``soft'' (b) attention
model. \textit{White} indicates the regions where the model roughly attends to (see section \ref{section:viz}).
\vskip -0.14in
\begin{figure*}[!ht]
\begin{center}
\centerline{\includegraphics[width=6.75in]{61_hard}}
\vskip -0.3in
(a) A man and a woman playing frisbee in a field.
\centerline{\includegraphics[width=6.75in]{61}}
\vskip -0.2in
(b) A woman is throwing a frisbee in a park.
\caption{}
\label{figure:im61}
\vskip -3in
\end{center}
\end{figure*}
\begin{figure*}[ht]
\vskip 0.2in
\begin{center}
\centerline{\includegraphics[width=7.3in]{1038_hard}}
\vskip -0.2in
(a) A giraffe standing in the field with trees.
\centerline{\includegraphics[width=7.3in]{1038}}
(b) A large white bird standing in a forest.
\caption{}
\label{figure:im1038}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7in]{570_hard}}
\vskip -0.2in
(a) A dog is laying on a bed with a book.
\centerline{\includegraphics[width=7in]{570}}
(b) A dog is standing on a hardwood floor.
\caption{}
\label{figure:im570}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7in]{352_hard}}
\vskip -0.3in
(a) A woman is holding a donut in his hand.
\centerline{\includegraphics[width=7in]{352}}
\vskip -0.4in
(b) A woman holding a clock in her hand.
\caption{}
\label{figure:im352}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\vskip -0.1in
\centerline{\includegraphics[width=8in]{1304_hard}}
\vskip -0.5in
(a) A stop sign with a stop sign on it.
\centerline{\includegraphics[width=9in]{1304}}
\vskip -0.2in
(b) A stop sign is on a road with a mountain in the background.
\caption{}
\label{figure:im1304}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7.5in]{1066_hard}}
\vskip -0.5in
(a) A man in a suit and a hat holding a remote control.
\centerline{\includegraphics[width=7.5in]{1066}}
\vskip -0.2in
(b) A man wearing a hat and a hat on a skateboard.
\caption{}
\label{figure:im1066}
\end{center}
\vskip -0.7in
\end{figure*}
\begin{figure*}[ht]
\vskip 0.2in
\begin{center}
\centerline{\includegraphics[width=7.5in]{120_hard}}
\vskip -0.4in
(a) A little girl sitting on a couch with a teddy bear.
\centerline{\includegraphics[width=7.5in]{120}}
\vskip -0.2in
(b) A little girl sitting on a bed with a teddy bear.
\label{figure:im120}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7in]{3884_hard}}
\vskip -0.4in
(a) A man is standing on a beach with a surfboard.
\centerline{\includegraphics[width=7in]{861_3884}}
\vskip -0.2in
(b) A person is standing on a beach with a surfboard.
\label{figure:im861_3884}
\end{center}
\vskip -0.8in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7.5in]{1322_hard}}
\vskip -0.4in
(a) A man and a woman riding a boat in the water.
\centerline{\includegraphics[width=7.5in]{1322_1463_3861_4468}}
\vskip -0.4in
(b) A group of people sitting on a boat in the water.
\caption{}
\label{figure:im1322_1463_3861_4468}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=8in]{462_hard}}
\vskip -0.5in
(a) A man is standing in a market with a large amount of food.
\centerline{\includegraphics[width=8in]{462_1252}}
\vskip -0.4in
(b) A woman is sitting at a table with a large pizza.
\caption{}
\label{figure:im462_1252}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7.3in]{4837_hard}}
\vskip -0.3in
(a) A giraffe standing in a field with trees.
\centerline{\includegraphics[width=7.3in]{1300_4837}}
\vskip -0.4in
(b) A giraffe standing in a forest with trees in the background.
\caption{}
\label{figure:im1300_4837}
\end{center}
\vskip -0.4in
\end{figure*}
\begin{figure*}[ht]
\begin{center}
\centerline{\includegraphics[width=7.5in]{1228_hard}}
\vskip -0.2in
(a) A group of people standing next to each other.
\centerline{\includegraphics[width=7.5in]{1228}}
\vskip -0.4in
(b) A man is talking on his cell phone while another man watches.
\caption{}
\label{figure:im1228}
\end{center}
\vskip -0.4in
\end{figure*}
\end{document}
|