File size: 96,787 Bytes
8cf8144
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
\RequirePackage{fix-cm}
\documentclass[twocolumn]{svjour3}

\smartqed  

\parindent=0pt
\parskip=0pt

\usepackage{times}

\usepackage{epsfig}
\usepackage{graphicx}
\usepackage[belowskip=-10pt,aboveskip=2pt,font=small]{caption}
\usepackage{dblfloatfix}

\usepackage{amsmath}
\usepackage{nccmath}
\usepackage{amssymb}

\usepackage{multirow}
\usepackage{rotating}
\usepackage{booktabs}
\usepackage{slashbox} \usepackage{algpseudocode}
\usepackage{algorithm}

\usepackage{enumerate}
\usepackage[olditem,oldenum]{paralist}




\usepackage[pagebackref=true,breaklinks=true,letterpaper=true,colorlinks,bookmarks=false,citecolor=blue,linkcolor=green]{hyperref}

\usepackage{color}

\usepackage[utf8]{inputenc} \usepackage[T1]{fontenc}    \usepackage[pagebackref=true,breaklinks=true,colorlinks,bookmarks=false,citecolor=blue]{hyperref} \usepackage{url}            \usepackage{amsfonts}       \usepackage{nicefrac}       \usepackage{microtype}      \usepackage{subcaption}\usepackage[font=footnotesize,labelfont=footnotesize]{caption}
\captionsetup[subfigure]{skip=0pt}

\usepackage{comment}
\usepackage{mysymbols}
\parskip=3pt
\abovedisplayskip 3.0pt plus2pt minus2pt\belowdisplayskip \abovedisplayskip
\renewcommand{\baselinestretch}{0.98}



\newenvironment{packed_enum}{
\begin{enumerate}
  \setlength{\itemsep}{1pt}
  \setlength{\parskip}{0pt}
  \setlength{\parsep}{0pt}
}
{\end{enumerate}}

\newenvironment{packed_item}{
\begin{itemize}
  \setlength{\itemsep}{1pt}
  \setlength{\parskip}{0pt}
  \setlength{\parsep}{0pt}
}{\end{itemize}}

\newlength{\sectionReduceTop}
\newlength{\sectionReduceBot}
\newlength{\subsectionReduceTop}
\newlength{\subsectionReduceBot}
\newlength{\abstractReduceTop}
\newlength{\abstractReduceBot}
\newlength{\captionReduceTop}
\newlength{\captionReduceBot}
\newlength{\subsubsectionReduceTop}
\newlength{\subsubsectionReduceBot}

\newlength{\eqnReduceTop}
\newlength{\eqnReduceBot}

\newlength{\horSkip}
\newlength{\verSkip}

\newlength{\figureHeight}
\setlength{\figureHeight}{1.7in}

\setlength{\horSkip}{-.09in}
\setlength{\verSkip}{-.1in}


\setlength{\subsectionReduceTop}{-0.12in}
\setlength{\subsectionReduceBot}{-0.09in}
\setlength{\sectionReduceTop}{-0.08in}
\setlength{\sectionReduceBot}{-0.10in}
\setlength{\subsubsectionReduceTop}{-0.06in}
\setlength{\subsubsectionReduceBot}{-0.05in}
\setlength{\abstractReduceTop}{-0.05in}
\setlength{\abstractReduceBot}{-0.10in}

\setlength{\eqnReduceTop}{-0.05in}
\setlength{\eqnReduceBot}{-0.05in}



\setlength{\captionReduceTop}{-0.14in}
\setlength{\captionReduceBot}{-0.12in}
 



\def\httilde{\mbox{\tt\raisebox{-.5ex}{\symbol{126}}}}

\definecolor{orange}{RGB}{225, 90, 0}
\definecolor{teal}{RGB}{5, 210, 150}
\definecolor{yellow}{RGB}{220, 210, 10}
\definecolor{purple}{RGB}{100, 0, 205}

\newcommand{\gb}{Guided Backpropagation}
\newcommand{\cgb}{Guided Grad-CAM}
\newcommand{\dec}{Deconvolution}
\newcommand{\cdec}{Deconvolution Grad-CAM}
\newcommand{\gcam}{Grad-CAM}
\newcommand{\refsec}[1]{Sec.~\ref{#1}}
\newcommand{\secref}[1]{Sec.~\ref{#1}}
\newcommand{\reffig}[1]{Fig.~\ref{#1}}
\newcommand{\figref}[1]{Fig.~\ref{#1}}
\newcommand{\reffigp}[1]{~\ref{#1}}
\newcommand{\reftab}[1]{Table.~\ref{#1}}
\newcommand{\rp}[1]{\textcolor{black}{#1}}
\newcommand{\rpi}[1]{\textcolor{black}{#1}}

\newcommand{\iccv}[1]{\textcolor{black}{#1}}
\newcommand{\ijcv}[1]{\textcolor{black}{#1}}
\newcommand{\ramaijcv}[1]{\textcolor{black}{#1}}

\newcommand{\rpr}[1]{\textcolor{black}{#1}}\newcommand{\fb}[1]{\textcolor{black}{#1}}\newcommand{\ad}[1]{\textcolor{black}{#1}}
\newcommand{\review}[1]{\textcolor{black}{#1}}
\newcommand{\mac}[1]{\textcolor{black}{#1}} \newcommand{\rama}[1]{\textcolor{black}{#1}}
\newcommand{\viceversa}{\textit{vice versa}\/}
\newcommand{\what}[1]{\emph{what}}
\newcommand{\where}[1]{\emph{where}}
\newcommand{\whatcent}[1]{\emph{what}-centric}
\newcommand{\wherecent}[1]{\emph{where}-centric}
\newcommand{\Whatcent}[1]{\emph{What}-centric}
\newcommand{\Wherecent}[1]{\emph{Where}-centric}
\newcommand{\para}[1]{\textbf{#1}.}
\usepackage[toc,page]{appendix}
\renewcommand\appendixtocname{Appendix}
\renewcommand\appendixpagename{Appendix}
\newcommand{\dummyfig}[1]{
  \centering
  \fbox{
    \begin{minipage}[c][0.2\textheight][c]{0.9\columnwidth}
      \centering{#1}
    \end{minipage}
  }
}


\setcounter{page}{1}
\begin{document}
\twocolumn
\title{Grad-CAM: Visual Explanations from Deep Networks\\
via Gradient-based Localization
}

\author{Ramprasaath R. Selvaraju
        \and Michael Cogswell
        \and Abhishek Das
        \and Ramakrishna Vedantam
        \and Devi Parikh
        \and Dhruv Batra
}



\institute{Ramprasaath R. Selvaraju \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
\email{ramprs@gatech.edu}           \and
           Michael Cogswell \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
              \email{cogswell@gatech.edu}
              \and
           Abhishek Das \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
              \email{abhshkdz@gatech.edu}
              \and
           Ramakrishna Vedantam \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
              \email{vrama@gatech.edu}
              \and
           Devi Parikh \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
              Facebook AI Research, Menlo Park, CA, USA\\
              \email{parikh@gatech.edu}
              \and
           Dhruv Batra \at
              Georgia Institute of Technology, Atlanta, GA, USA\\
              Facebook AI Research, Menlo Park, CA, USA\\
              \email{dbatra@gatech.edu}
}



\date{}






\maketitle



\makeatletter
\def\blfootnote{\xdef\@thefnmark{}\@footnotetext}
\makeatother
\begin{abstract}







\label{sec:abstract}
We propose a technique for producing `visual explanations' for decisions from a large class of Convolutional Neural Network (CNN)-based models, making them more transparent and explainable.

Our approach -- Gradient-weighted Class Activation Mapping (Grad-CAM), uses the gradients of any target concept (say `dog' in a classification network or a sequence of words in captioning network) flowing into the final convolutional layer to produce a coarse localization map highlighting the important regions in the image for predicting the concept.

Unlike previous approaches, \gcam{} is applicable to a wide variety of CNN model-families: 
(1) CNNs with fully-connected layers (\eg VGG), 
(2) CNNs used for structured outputs (\eg captioning), 
(3) CNNs used in tasks with multi-modal inputs (\eg visual question answering) or reinforcement learning, all \emph{without architectural changes or re-training}.
We combine \gcam{} with existing fine-grained visualizations to create a high-resolution class-discriminative visualization, \cgb{},
and apply it to image classification,
image captioning, and visual question answering (VQA) models, including ResNet-based architectures.

In the context of image classification models, our visualizations (a) lend insights into failure modes of these models
(showing that seemingly unreasonable predictions have reasonable explanations),
(b) \mac{outperform previous methods on the ILSVRC-15 weakly-supervised localization task,}
(c) are robust to adversarial perturbations,
(d) are more faithful to the underlying model,
and (e) help achieve model generalization by identifying dataset bias.

For image captioning and VQA, our visualizations show that
even non-attention based models learn to localize discriminative regions of input image.

We devise a way to identify important neurons through \gcam{}
and combine it with neuron names~\cite{netdissect} to provide textual explanations for model decisions.
Finally, we design and conduct human studies to measure if \gcam{} explanations help users
establish appropriate trust in predictions from deep networks and show that \gcam{} helps untrained users
successfully discern a `stronger' deep network from a `weaker' one even when both make identical predictions.
\iccv{Our code is available at {\small \url{https://github.com/ramprs/grad-cam/}}, along with a demo
on CloudCV~\cite{agrawal2015cloudcv}\footnote{\url{http://gradcam.cloudcv.org}}, and a video at \footnotesize{\url{youtu.be/COjUB9Izk6E}}.}
 \end{abstract}

\vspace{-15pt}
\section{Introduction}
Deep neural models based on Convolutional Neural Networks (CNNs)
have enabled unprecedented breakthroughs in a variety of computer vision tasks, from image
classification~\cite{krizhevsky_nips12,he_cvpr15}, object detection~\cite{girshick2014rcnn}, semantic segmentation~\cite{long2015fcn} to image captioning~\cite{vinyals_cvpr15,chen2015microsoft,fang2015captions,johnson_cvpr16}, visual question answering~\cite{antol2015vqa,gao2015you,malinowski_iccv15,ren_nips15}
and more recently, visual dialog~\cite{visdial,guesswhat,visdial_rl} and
embodied question answering~\cite{embodiedqa,gordon2017iqa}.
While these models enable superior performance, their lack of decomposability into
\emph{individually intuitive} components makes them hard to interpret~\cite{lipton_arxiv16}.
Consequently, when today's intelligent systems fail, they often fail spectacularly disgracefully
without warning or explanation, leaving a user staring at an incoherent output, wondering why the system did what it did.  


\emph{Interpretability matters.}
In order to build trust in intelligent systems and move towards their meaningful
integration into our everyday lives,
it is clear that we must build `transparent' models that have the ability to
explain \emph{why they predict what they \rama{predict}}.
Broadly speaking, this transparency and ability to explain is useful at three different stages of Artificial Intelligence (AI) evolution.
First, when AI is significantly weaker than humans and not yet reliably deployable (\eg visual question
answering ~\cite{antol2015vqa}), the goal of transparency and explanations is to identify the failure
modes~\cite{agrawal2016analyzing,hoiem2012diagnosing}, thereby helping researchers focus their
efforts on the most fruitful research directions.
Second, when AI is on par with humans and reliably deployable (\eg, image classification~\cite{karpathy_imagenet}
trained on sufficient data), the goal
is to establish appropriate trust \rama{and confidence in} users.
Third, when AI is significantly stronger than humans (\eg chess or Go~\cite{silver2016mastering}),
the goal of explanations is \rama{in} machine teaching~\cite{JohnsCVPR2015} -- \ie, a
machine teaching a human about how to make better decisions.




\begin{figure*}[t!]
	\centering
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=\textwidth]{figures/teaser/original.jpg}
        \caption{\scriptsize{Original Image}}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=\textwidth]{figures/teaser/gb.jpg}
        \caption{\scriptsize{Guided Backprop `Cat'}}
        \label{fig:teaser_gb_cat}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=\textwidth]{figures/teaser/newgcam_heatmap_overlaid_283_cat_dog_with_margin_small.jpg}
        \caption{\scriptsize{Grad-CAM `Cat'}}
        \label{fig:teaser_gcam_cat}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=0.995\textwidth]{figures/teaser/gbgcam_cat_margin.jpg}
        \caption{\hspace{-1.5pt}\scriptsize{Guided Grad-CAM `Cat'}}
        \label{fig:teaser_gbgcam_cat}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=\textwidth]{figures/teaser/occlusion.jpg}
        \caption{\scriptsize{Occlusion map `Cat'}}
        \label{fig:teaser_cat_occlusion}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=0.995\textwidth]{figures/gcam_res18_283.jpg}
		\caption{\scriptsize{ResNet \gcam{} `Cat'}}
        \label{fig:teaser_cat_occlusion}
	\end{subfigure}\\
    \vspace*{10pt}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
        \includegraphics[width=\textwidth]{figures/teaser/original.jpg}
        \caption{\scriptsize{Original Image}\hspace{-12pt}}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
        \includegraphics[width=\textwidth]{figures/teaser/dog_gb.jpg}
		\caption{\scriptsize{Guided Backprop `Dog'}\hspace{-12pt}}
        \label{fig:teaser_gb_dog}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
        \includegraphics[width=\textwidth]{figures/teaser/newgcam_heatmap_overlaid_243_cat_dog_with_margin_small.jpg}
        \caption{\scriptsize{Grad-CAM `Dog'}\hspace{-12pt}}
        \label{fig:teaser_gcam_dog}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
        \includegraphics[width=1.01\textwidth]{figures/teaser/gbgcam_dog_margin.jpg}
        \caption{\hspace{-1.9pt}\scriptsize{Guided Grad-CAM `Dog'}\hspace{-12pt}}
        \label{fig:teaser_gbgcam_dog}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
        \includegraphics[width=\textwidth]{figures/teaser/dog_occlusion.jpg}
        \caption{\scriptsize{Occlusion map `Dog'}\hspace{-12pt}}
        \label{fig:teaser_dog_occlusion}
	\end{subfigure}
	\begin{subfigure}[t]{0.161\textwidth}
		\centering
		\includegraphics[width=0.995\textwidth]{figures/gcam_res18_243.jpg}
		\caption{\hspace{-1.9pt}\scriptsize{ResNet \gcam{} `Dog'}}
        \label{fig:teaser_gcam_resnet}
	\end{subfigure}
    \vspace{14pt}
	\caption{
		(a) Original image with a cat and a dog. (b-f) Support for the cat category according to various visualizations \rp{for VGG-16 and ResNet}.
		(b) Guided Backpropagation~\cite{springenberg_arxiv14}: highlights all contributing features. (c, f) \gcam{} (Ours): localizes class-discriminative regions, (d) Combining (b) and (c) gives Guided Grad-CAM, which gives high-resolution class-discriminative visualizations. Interestingly, the localizations achieved by our \gcam{} technique, (c) are very similar to results from occlusion sensitivity (e), while being orders of magnitude cheaper to compute. (f, l) are \gcam{} visualizations for ResNet-18 layer.
		Note that in (c, f, i, l), red regions corresponds to high score for class, while in (e, k), blue corresponds to evidence for the class.
		Figure best viewed in color.
	  }
    \label{fig:teaser}
\end{figure*}

There typically exists a trade-off between accuracy and simplicity or interpretability.
Classical rule-based or expert systems~\cite{jackson_expertsys} are highly interpretable but not very accurate (or robust).
Decomposable pipelines where each stage is hand-designed are thought to be more interpretable as each individual component assumes a natural intuitive explanation.
By using deep models, we sacrifice interpretable modules for uninterpretable ones that achieve greater performance through greater abstraction (more layers) and tighter integration (end-to-end training). Recently introduced deep residual networks (ResNets)~\cite{he_cvpr15} are over 200-layers deep and have shown state-of-the-art performance in several challenging tasks. \rama{
Such complexity makes these models hard to interpret.}
As such, deep models are beginning to explore the spectrum between interpretability and accuracy.

\rpr{Zhou~\etal~\cite{zhou_cvpr16} recently proposed a technique called Class Activation Mapping (CAM) for identifying discriminative regions used by a restricted class of image classification CNNs which do not contain any fully-connected layers.
  In essence, this work trades off model complexity and performance for more transparency into the working of the model.
In contrast, we make existing state-of-the-art deep models interpretable without altering their architecture, thus \rama{avoiding the interpretability \vs accuracy trade-off}.
Our approach is a generalization of CAM~\cite{zhou_cvpr16} and is applicable to a significantly broader range of CNN model families: (1) CNNs with fully-connected layers (\eg VGG), (2) CNNs used for structured outputs (\eg captioning), (3) CNNs used in tasks with multi-modal inputs (\eg VQA) or reinforcement learning, without requiring architectural changes or re-training. }






\noindent \textbf{What makes a good visual explanation?}
Consider image classification~\cite{imagenet_cvpr09} -- a `good’ visual explanation from the model for justifying any target category should be (a) class-discriminative (\ie localize the category in the image) and (b) high-resolution (\ie capture fine-grained detail).

Fig.~\hyperlink{page.2}{1} shows outputs from a number of visualizations for the `tiger cat' class (top) and `boxer' (dog) class (bottom).
\mac{Pixel-space gradient visualizations such as \gb{}~\cite{springenberg_arxiv14}  and \dec{}~\cite{zeiler_eccv14}
are high-resolution and highlight fine-grained details in the image, but are not class-discriminative
(\figref{fig:teaser_gb_cat} and \figref{fig:teaser_gb_dog} are very similar).}



\mac{In contrast, localization approaches like CAM or our proposed method Gradient-weighted Class Activation Mapping (\gcam{}),
are highly class-discriminative
(the `cat' explanation exclusively highlights the `cat' regions but not `dog' regions in \figref{fig:teaser_gcam_cat}, and \viceversa{} in \figref{fig:teaser_gcam_dog}).
}

In order to combine the best of both worlds, we show that it is possible to fuse existing pixel-space gradient visualizations with \gcam{} to create \cgb{} visualizations that are both high-resolution and class-discriminative. As a result, important regions of the image which correspond to any decision of interest are visualized in high-resolution detail even if the image contains \rp{evidence for multiple possible} \rama{concepts}, as shown in Figures \hyperlink{page.2}{1d} and \hyperlink{page.2}{1j}.
When visualized for `tiger cat', \cgb{} not only highlights the cat regions, but also highlights the stripes on the cat, which is important for predicting that particular variety of cat.
















\noindent To summarize, our contributions are as follows:

\noindent \textbf{(1)}
\mac{We introduce \gcam{}, a class-discriminative localization technique that generates
visual explanations for \emph{any} CNN-based network without requiring architectural changes or re-training.
We evaluate \gcam{} for localization (\secref{sec:localization}), and faithfulness to model (\secref{sec:occ}), where it outperforms baselines.}

\noindent \textbf{(2)}
\mac{We apply \gcam{} to existing top-performing classification, captioning (\secref{sec:nic}), and VQA (\secref{sec:vqa}) models.}
For image classification, our visualizations lend insight into failures of current CNNs (\secref{sec:diagnose}), showing that seemingly unreasonable predictions have reasonable explanations.
For captioning and VQA, our visualizations expose
that common \mac{CNN + LSTM models are often surprisingly good at}
localizing discriminative image regions despite not being trained on
grounded image-text pairs.

\noindent \textbf{(3)}
We show a proof-of-concept of how interpretable \gcam{} visualizations
help in diagnosing failure modes by uncovering biases in datasets.
This is important not just for generalization, but also for fair and
bias-free outcomes as more and more decisions are made by algorithms in society.

\noindent \textbf{(4)}
We present Grad-CAM visualizations for ResNets~\cite{he_cvpr15} applied to
image classification and VQA (\secref{sec:vqa}).

\ijcv{\noindent \textbf{(5)}
We use neuron importance from \gcam{} and neuron names from~\cite{netdissect} and obtain textual explanations for model decisions (\secref{sec:text_exp}).}

\noindent \textbf{(6)}
We conduct human studies (\secref{sec:human_evaluation}) \mac{that show \cgb{} explanations are class-discriminative and not only help humans establish trust,
but also help untrained users successfully discern a `stronger' network from a `weaker' one,
\emph{even when both make identical predictions.}}




\rpi{\noindent \textbf{Paper Organization: }
The rest of the paper is organized as follows. 
In section 3 we propose our approach \gcam{} and \cgb{}. 
In sections 4 and 5 we evaluate the localization ability, class-discriminativeness, trustworthyness and faithfulness of \gcam{}. 
In section 6 we show certain use cases of \gcam{} such as diagnosing image classification CNNs and identifying biases in datasets. 
In section 7 we provide a way to obtain textual explanations with \gcam{}. 
In section 8 we show how \gcam{} can be applied to vision and language models -- image captioning and Visual Question Answering (VQA).}


\begin{figure*}
    \begin{center}
  \centering
  \includegraphics[width=\textwidth]{figures/Grad-CAM_approach.png}
	\caption{\scriptsize  \gcam{} overview: Given an image \mac{ and a class of interest (\eg, `tiger cat' or any other type of differentiable output) as input, we forward propagate the image through the CNN part of the model and then through task-specific computations to obtain a raw score for the category.}
        The gradients are set to zero for all classes except the desired class (tiger cat), which is set to 1.
		This signal is then backpropagated to the rectified convolutional \mac{feature maps of interest, which we combine to compute} the coarse \gcam{} localization (blue heatmap) \rp{which represents where the model has to look to make the particular decision}.
	Finally, we pointwise multiply the heatmap with guided backpropagation to get \cgb{} visualizations which are both high-resolution and concept-specific.
}
	\label{fig:approach}
\end{center}
\end{figure*}

 \vspace{-15pt}
\section{Related Work}

Our work draws on recent work in CNN visualizations, model trust assessment, and weakly-supervised localization.



\noindent \para{Visualizing CNNs}
A number of previous works~\cite{simonyan_arxiv13,springenberg_arxiv14,zeiler_eccv14,Gan_2015_CVPR} have visualized CNN predictions by highlighting `important' pixels (\ie change in intensities of these pixels have the most impact on the prediction score).
Specifically, Simonyan \etal~\cite{simonyan_arxiv13} visualize partial derivatives of predicted class scores \wrt pixel intensities, while \gb{}~\cite{springenberg_arxiv14} and \dec{}~\cite{zeiler_eccv14} make modifications to `raw' gradients that result in qualitative improvements.
\mac{These approaches are compared in ~\cite{mahendran16eccv}.}
Despite producing fine-grained visualizations, these methods are not class-discriminative. Visualizations with respect to different classes are nearly identical (see Figures \hyperlink{page.2}{1b} and \hyperlink{page.2}{1h}).


Other visualization methods synthesize images to maximally activate a network unit~\cite{simonyan_arxiv13,erhan2009visualizing} or invert a latent representation~\cite{mahendran2016visualizing,dosovitskiy_cvpr16}.
Although these can be high-resolution and class-discriminative,
they are not specific
to a single input image and visualize a model overall.



\noindent\para{Assessing Model Trust}
Motivated by notions of interpretability~\cite{lipton_arxiv16} and assessing trust in models~\cite{lime_sigkdd16}, we evaluate \gcam{} visualizations in a manner similar to ~\cite{lime_sigkdd16} via human studies to show that they can be important tools for users to evaluate and place trust in automated systems.

\noindent\para{Aligning Gradient-based Importances}
 Selvaraju \etal \cite{niwt} proposed an approach that uses the gradient-based neuron importances introduced in our work, and maps it to class-specific domain knowledge from humans in order to learn classifiers for novel classes. In future work, Selvaraju \etal \cite{hint} proposed an approach to align gradient-based importances to human attention maps in order to ground vision and language models. 

























\noindent \para{Weakly-supervised localization}
Another relevant line of work is weakly-supervised localization in the context of CNNs, where the task is to localize objects in images using holistic image class labels only~\cite{cinbis2016weakly,oquab_cvpr14,oquab_cvpr15,zhou_cvpr16}.








Most relevant to our approach is the Class Activation Mapping (CAM) approach to localization~\cite{zhou_cvpr16}.
This approach modifies image classification CNN architectures replacing fully-connected layers with convolutional layers and global average pooling~\cite{lin2013network}, thus achieving class-specific feature maps. Others have investigated similar methods using global max pooling~\cite{oquab_cvpr15} and log-sum-exp pooling~\cite{pinheiro2015image}.



A drawback of CAM is that it requires feature maps to directly precede softmax layers, so it is only applicable to a particular kind of CNN architectures performing global average pooling over convolutional maps immediately prior to prediction (\ie conv feature maps $\rightarrow$ global average pooling $\rightarrow$ softmax layer).
Such architectures may achieve inferior accuracies compared to general networks on some tasks (\eg image classification) or may simply be inapplicable to any other tasks (\eg image captioning or VQA).
We introduce a new way of combining feature maps using the gradient signal that does not require \emph{any} modification in the network architecture.
This allows our approach to be applied to off-the-shelf CNN-based architectures,
including those for image captioning and visual question answering.
For a fully-convolutional architecture, CAM is a special case of \gcam{}.





Other methods approach localization by classifying perturbations of the input image.
Zeiler and Fergus~\cite{zeiler_eccv14} perturb inputs by occluding patches and classifying the occluded image, typically resulting in lower classification scores for relevant objects when those objects are occluded.
This principle is applied for localization in \cite{bazzani2016self}. Oquab \etal~\cite{oquab_cvpr14} classify many patches containing a pixel then average these patch-wise
scores to provide the pixel's class-wise score.
Unlike these, our approach achieves localization in one shot; it only requires a single forward and a partial backward pass per image and thus is typically an order of magnitude more efficient.
In recent work, Zhang~\etal~\cite{zhang2016top} introduce contrastive Marginal Winning Probability (c-MWP), a probabilistic Winner-Take-All formulation for modelling the top-down attention for neural classification models which can highlight discriminative regions. 
This is computationally more expensive than \gcam{} and only works for
image classification CNNs. 
Moreover, \gcam{} outperforms c-MWP in quantitative and qualitative evaluations (see \refsec{sec:localization} and \refsec{sec:sup_pointing}).






 \vspace{-10pt}
\section{\gcam{}}\label{sec:approach}
A number of previous works have asserted that deeper representations in a CNN capture higher-level visual constructs~\cite{bengio2013representation,mahendran2016visualizing}.
Furthermore, convolutional layers naturally retain spatial information which is lost in fully-connected layers, so we can expect the last convolutional layers to have the best compromise between high-level semantics and detailed spatial information.
The neurons in these layers look for semantic class-specific information in the image (say object parts).
\gcam{} uses the gradient information flowing into the last convolutional layer
of the CNN to assign importance values to each neuron for a particular decision of interest.
\rp{
    Although our technique is fairly general in that it can be used to explain
    activations in any layer of a deep network, in this work, we focus on explaining
    output layer decisions only.
}

As shown in \reffig{fig:approach}, in order to obtain the class-discriminative localization map \gcam{}
$L_{\text{\gcam{}}}^c$ $\in \mathbb{R}^{u \times v}$ of width $u$ and height $v$ for any class $c$,
we first compute the gradient
of the score for class $c$, $y^c$ \mac{(before the softmax)}, with respect to feature map
activations $A^k$ of a convolutional layer, \ie $\frac{\del y^c}{\del A^k}$.
These gradients flowing back are global-average-pooled~\footnote{\rpi{Empirically we found global-average-pooling to work better than global-max-pooling as can be found in the Appendix. }}
\ad{
    over the width and height dimensions (indexed by $i$ and $j$ respectively)}
to obtain the neuron importance weights $\alpha{}_{k}^c$:
\begin{ceqn}
\begin{equation} \label{eq:alpha1}
    \alpha{}_{k}^c =
    \overbrace{
        \frac{1}{Z}\sum_{i}\sum_{j}
    }^{\text{global average pooling}}
    \hspace{-17pt}
    \underbrace{
        \vphantom{\sum_{i}\sum_{j}} \frac{\partial y^c}{\partial A_{ij}^{k}}
    }_{\text{gradients via backprop}}
\end{equation}
\end{ceqn}
\rpi{During computation of $\alpha{}_{k}^c$ while backpropagating gradients with respect to activations, the exact computation amounts to successive matrix products of the weight matrices and the gradient with respect to activation functions till the final convolution layer that the gradients are being propagated to. }
Hence, this weight $\alpha{}_{k}^c$ represents a \emph{partial linearization} of the deep network downstream from A,
and captures the `importance' of feature map $k$ for a target class $c$. 

We perform a weighted combination of forward activation maps, and follow it by a ReLU to obtain,
\begin{ceqn}
\begin{equation} \label{eq:gcam}
    L_{\text{\gcam{}}}^{c} = ReLU \underbrace{\left(\sum_k \alpha{}_{k}^{c} A^{k}\right)}_{\text{linear combination}}
\end{equation}
\end{ceqn}
Notice that this results in a coarse heatmap of the same size as the convolutional feature
maps ($14 \times 14$ in the case of last convolutional layers of VGG \cite{simonyan_arxiv14} and AlexNet \cite{krizhevsky_nips12} networks) \footnote{\rpi{We find that \gcam{} maps become progressively worse as we move to earlier convolutional layers as they have smaller receptive fields and only focus on less semantic local features.}}.
\mac{We apply a ReLU to the linear combination of maps because we are only interested in the features that have a \emph{positive} influence on the class of interest, \ie
pixels whose intensity should be \emph{increased} in order to increase $y^c$.
Negative pixels are likely to belong to other categories in the image.
As expected, without this ReLU, localization maps sometimes highlight more than
just the desired class and perform worse at localization.}
Figures \hyperlink{page.2}{1c, 1f} and \hyperlink{page.2}{1i, 1l} show \gcam{} visualizations for `tiger cat'
and `boxer (dog)' respectively. 
Ablation studies are available in ~\secref{sec:ablation}. 


\mac{
    In general, $y^c$ need not be the class score produced by an image classification
    CNN. It could be any differentiable activation including words from a caption or
    answer to a question.
}

\vspace{-10pt}
\subsection{\gcam{} generalizes CAM}
\label{sec:generalization}
\ijcv{In this section, we \ad{discuss the connections between \gcam{} and Class Activation Mapping (CAM)~\cite{zhou_cvpr16},
and formally  prove that \gcam{} generalizes CAM} for a wide variety of CNN-based architectures.}
\noindent Recall that CAM produces a localization map for an image classification
CNN with a specific kind of architecture where global average pooled convolutional feature maps
are fed directly into softmax.
Specifically, let the penultimate layer produce $K$ feature maps,
$A^k \in \mathbb{R}^{u \times v}$, with each element indexed by $i,j$.
So $A_{ij}^k$ refers to the activation at location $(i,j)$ of the feature map $A^k$.
These feature maps are then spatially pooled using Global Average Pooling (GAP)
and linearly transformed to produce a score $Y^c$ for each class $c$,
\begin{ceqn}
\begin{equation} \label{eq:scores}
    Y^c = \sum_{k}
	\hspace{-20pt}
    \underbrace{
        \vphantom{\frac{1}{Z}\sum_i \sum_j} w^c_k
    }_{\vspace{-3pt}
	\text{class feature weights}}
    \hspace{-27pt}
    \overbrace{
        \frac{1}{Z}\sum_i \sum_j
    }^{\text{global average pooling}}
    \hspace{-15pt}
    \underbrace{
        \vphantom{\frac{1}{Z}\sum_i \sum_j} A^k_{ij}
    }_{\vspace{-5pt}\text{feature map}}
\end{equation}
\end{ceqn}




Let us define $F^k$ to be the global average pooled output,
\begin{ceqn}
\begin{equation}\label{eq:cam_gap}
    F^{k} = \frac{1}{Z} \sum_{i} \sum_{j} A_{ij}^k
\end{equation}
\end{ceqn}

CAM computes the final scores by,
\begin{ceqn}
\begin{equation}\label{eq:cam_scores}
    Y^c = \sum_{k} w_{k}^c \cdot F^{k}
\end{equation}
\end{ceqn}
where $w_{k}^c$ is the weight connecting the $k^{th}$ feature map with the $c^{th}$ class.
Taking the gradient of the score for class c ($Y^c$)  with respect to the feature map $F^k$ we get,
\begin{ceqn}
\begin{equation}\label{eq:cam_grad}
    \frac{\partial Y^c}{\partial F^k} = \frac{\frac{\partial Y^c}{\partial A_{ij}^k}}{\frac{\partial F^k}{\partial A_{ij}^k}}
\end{equation}
\end{ceqn}

Taking partial derivative of  \eqref{eq:cam_gap} \wrt $A_{ij}^k$, we can see that $\frac{\del F^k}{\del A^k_{ij}} = \frac{1}{Z}$. Substituting this in \eqref{eq:cam_grad}, we get,
\begin{ceqn}
\begin{equation}
    \frac{\partial Y^c}{\partial F^k} = \frac{\partial Y^c}{\partial A_{ij}^k} \cdot Z\\
\end{equation}
\end{ceqn}

From \eqref{eq:cam_scores} we get that, $ \frac{\partial Y^c}{\partial F^k} = w_{k}^c$. Hence,

\begin{ceqn}
\begin{equation}\label{eq:cam_weights}
    w_{k}^c = Z \cdot \frac{\partial Y^c}{\partial A_{ij}^k}\\
\end{equation}
\end{ceqn}

Summing both sides of \eqref{eq:cam_weights} over all pixels $(i,j)$,

\begin{ceqn}
\begin{equation}
    \sum_i \sum_j w^c_k  = \sum_i \sum_j Z \cdot \frac{\del Y^c}{\del A^k_{ij} }
\end{equation}
\end{ceqn}

Since $Z$ and $w^c_k$ do not depend on $(i,j)$, rewriting this as

\begin{ceqn}
\begin{equation}
    Z w^c_k  = Z \sum_i \sum_j \frac{\del Y^c}{\del A^k_{ij}}
\end{equation}
\end{ceqn}

Note that $Z$ is the number of pixels in the feature map (or $Z = \sum_i \sum_j \mathbf{1} $).
Thus, we can re-order terms and see that

\begin{ceqn}
\begin{equation}
    w^c_k  = \sum_i \sum_j \frac{\del Y^c}{\del A^k_{ij} }\\
\end{equation}
\end{ceqn}

Up to a proportionality constant ($1/Z$) that gets normalized-out during
visualization, the expression for $w^c_k$ is identical to $\alpha^c_k$ used by \gcam{} \eqref{eq:alpha1}.
Thus, \gcam{} is a strict generalization of CAM.
This generalization allows us to generate visual explanations from CNN-based
models that cascade convolutional layers with much more complex interactions,
such as those for image captioning and VQA (\refsec{sec:vqa}).


\vspace{-5pt}
\subsection{\cgb{}}
While \gcam{} is class-discriminative and localizes relevant image regions,
it lacks the ability to highlight fine-grained details like pixel-space gradient
visualization methods (\gb{}~\cite{springenberg_arxiv14}, \dec{}~\cite{zeiler_eccv14}).
\ad{\gb{} visualizes gradients with respect to the image where
    negative gradients are suppressed when backpropagating through ReLU layers.
    Intuitively, this aims to capture pixels detected by neurons, not
    the ones that suppress neurons.}
See Figure \hyperlink{page.2}{1c}, where
\gcam{} can easily localize the cat; however, it is unclear from the coarse
heatmap why the network predicts this particular instance as `tiger cat'.
In order to combine the best aspects of both, we fuse \gb{} and \gcam{} visualizations
via element-wise multiplication
($L_{\text{\gcam{}}}^c$ is first upsampled to the input image resolution using bilinear interpolation).
\figref{fig:approach} bottom-left illustrates this fusion.
This visualization is both high-resolution (when the class of interest is `tiger cat',
it identifies important `tiger cat' features like stripes, pointy ears and eyes) and
class-discriminative (it highlights the `tiger cat' but not the `boxer (dog)').
Replacing \gb{} with \dec{} gives similar results, but we found \dec{}
visualizations to have artifacts and \gb{} to be generally less noisy.

\subsection{Counterfactual Explanations}


Using a slight modification to \gcam{}, we can obtain explanations
that highlight support for regions that would make the network change its prediction. 
As a consequence, removing concepts occurring in those regions would make the
model more confident about its prediction. 
We refer to this explanation modality as counterfactual explanations.

Specifically, we negate the gradient of $y^c$ (score for class $c$) with respect to feature maps $A$ of a convolutional layer.
Thus the importance weights $\alpha{}_{k}^{c}$ now become
\begin{ceqn}
\begin{equation} \label{eq:alpha_neg}
    \alpha{}_{k}^c =
    \overbrace{
        \frac{1}{Z}\sum_{i}\sum_{j}
    }^{\text{global average pooling}}
    \hspace{-17pt}
    \underbrace{
        \vphantom{\sum_{i}\sum_{j}} -\frac{\partial y^c}{\partial A_{ij}^{k}}
    }_{\text{Negative gradients}}
\end{equation}
\end{ceqn}
As in \eqref{eq:gcam}, we take a weighted sum of the forward activation maps, $A$, with weights $\alpha{}_{k}^c$, and follow it by a ReLU to obtain counterfactual explanations as shown in \reffig{fig:negexp}.

\begin{figure}[ht!]
    \centering \begin{subfigure}[t]{0.158\textwidth} \centering
        \includegraphics[width=\textwidth]{figures/teaser/original.jpg}
        \caption{\scriptsize{Original Image}}
	\end{subfigure}
	\begin{subfigure}[t]{0.158\textwidth}
        \centering
        \includegraphics[width=\textwidth]{figures/cat_neg_exp.jpg}
        \caption{\scriptsize{Cat Counterfactual exp}}
		\label{fig:neg_exp_cat}
	\end{subfigure}
    \centering
	\begin{subfigure}[t]{0.158\textwidth}
        \centering
        \includegraphics[width=\textwidth]{figures/dog_neg_exp.jpg}
        \caption{\ijcv{\scriptsize{\hspace{-1pt}Dog Counterfactual exp}}}\label{fig:neg_exp_dog}
	\end{subfigure}
    \vspace{10pt}
    \caption{Counterfactual Explanations with \gcam{}}
    \label{fig:negexp}
\end{figure}






 \vspace{-10pt}
\section{Evaluating Localization Ability of \gcam{}}
\subsection{Weakly-supervised Localization}\label{sec:localization}
In this section, we evaluate the localization capability of \gcam{} in the context of image classification.
The ImageNet localization challenge~\cite{imagenet_cvpr09} requires
approaches to provide bounding boxes in addition to classification labels.
Similar to classification, evaluation is performed for both the top-1 and top-5 predicted categories.

Given an image, we first obtain class predictions from our network and
then generate \gcam{} maps for each of the predicted classes and binarize them with a
threshold of 15\% of the max intensity.
This results in connected segments of pixels and we draw a bounding box around
the single largest segment.
Note that this is weakly-supervised localization -- the models were never exposed
to bounding box annotations during training.

We evaluate \gcam{} localization with off-the-shelf pretrained VGG-16~\cite{simonyan_arxiv14} \rpi{, AlexNet ~\cite{krizhevsky_nips12} and GoogleNet \cite{szegedy2016rethinking} 
(obtained from the Caffe~\cite{jia2014caffe} Zoo). }
Following ILSVRC-15 evaluation, we report both top-1 and top-5 localization errors
on the val set in \reftab{table:locres}.
\gcam{} localization errors are significantly better than those achieved by
c-MWP~\cite{zhang2016top} and Simonyan~\etal~\cite{simonyan_arxiv13}, which use grab-cut to
post-process image space gradients into heat maps.
\gcam{} for VGG-16 also achieves better top-1 localization error than CAM~\cite{zhou_cvpr16}, which requires a change
in the model architecture, necessitates re-training and thereby achieves worse classification errors (2.98\%
worse top-1), while \gcam{} does not compromise on classification performance.



\begin{table}[h!]
\vspace{-10pt}
\centering
    \resizebox{1.00\columnwidth}{!}{
        \begin{tabular}{l l l c c c c c c}
            & & \multicolumn{2}{c}{\textbf{Classification}} &~~~ & \multicolumn{2}{c}{\textbf{Localization}}\\
            \cmidrule{3-4}\cmidrule{6-7}
            & & \textbf{Top-$1$} & \textbf{Top-$5$} & & \textbf{Top-$1$} & \textbf{Top-$5$} \\
            \midrule
            \multirow{4}{*}{\rotatebox{90}{\tiny{\centering VGG-16}}} 

 &           Backprop~\cite{simonyan_arxiv13}   & $30.38$ & $10.89$ & & $61.12$ & $51.46$        \\
 &           c-MWP~\cite{zhang2016top}          & $30.38$ & $10.89$ & & $70.92$ & $63.04$        \\
&			\gcam{} (ours)                     & $30.38$ & $10.89$ & & $\mathbf{56.51}$ & $46.41$        \\
             \cmidrule{2-7}
&			CAM~\cite{zhou_cvpr16} & $33.40$ & $12.20$ & & $57.20$ & $\mathbf{45.14}$        \\
\midrule
\multirow{2}{*}{\rotatebox{90}{\tiny{\centering AlexNet}}} 
&           c-MWP~\cite{zhang2016top}          & $44.2$ & $20.8$ & & $92.6$ & $89.2$        \\
&			\gcam{} (ours)                     & $44.2$ & $20.8$ & & $68.3$ & $56.6$        \\
\midrule
\multirow{2}{*}{\rotatebox{90}{\tiny{\centering GoogleNet}}} 
&			\gcam{} (ours)                     & $31.9$ & $11.3$ & & $60.09$ & $49.34$        \\
&			CAM~\cite{zhou_cvpr16} 			   & $31.9$ & $11.3$ & & $60.09$ & $49.34$        \\
            \bottomrule
        \end{tabular}
    }
    \vspace{5pt}
    \caption{Classification and localization error \% on ILSVRC-15 val (lower is better) for VGG-16, AlexNet and GoogleNet. We see that \gcam{} achieves superior localization errors without compromising on classification performance.}
    \label{table:locres}
\end{table}

\vspace{-20pt}
\subsection{Weakly-supervised Segmentation}\label{sec:segmentation}

Semantic segmentation involves the task of assigning each pixel in the image an object class (or background class).
Being a challenging task, this requires expensive pixel-level annotation. 
The task of weakly-supervised segmentation involves segmenting objects with just image-level annotation, which can be obtained relatively cheaply from image classification datasets.
In recent work, Kolesnikov~\etal \cite{seed_eccv16} introduced a new loss function for training weakly-supervised image segmentation models.
Their loss function is based on three principles --
1) to seed with weak localization cues, encouraging segmentation network to match these cues,
2) to expand object seeds to regions of reasonable size based on information about which classes can occur in an image,
3) to constrain segmentations to object boundaries that alleviates the problem of imprecise boundaries already at training time.
They showed that their proposed loss function, consisting of the above three losses leads to better segmentation. 

However, their algorithm is sensitive to the choice of weak localization seed,
without which the network fails to localize objects correctly.
In their work, they used CAM maps from a VGG-16 based network which are used as object seeds for weakly localizing foreground classes.
We replaced the CAM maps with \gcam{} obtained from a standard VGG-16 network and obtain a Intersection over Union (IoU) score of 49.6 (compared to 44.6 obtained with CAM) on the PASCAL VOC 2012 segmentation task. \reffig{fig:segmentation_qual} shows some qualitative results.


\begin{figure}[htp]
\vspace{-17pt}
 \centering
 \includegraphics[width=1\linewidth]{figures/segmentation_qualitative.jpg}
 \caption{PASCAL VOC 2012 Segmentation results with \gcam{} as seed for SEC~\cite{seed_eccv16}.}
 \label{fig:segmentation_qual}
 \vspace{-10pt}
\end{figure}




\begin{figure*}[ht]
    \centering
   \begin{subfigure}[b]{0.25\textwidth}
       \centering
       \includegraphics[width=0.55\linewidth]{figures/heval_1.jpg}
       \vspace{0.72in}
       \caption{\scriptsize{Raw input image. Note that this is not a part of the tasks (b) and (c)}}
		\label{fig:hs_clsdisc}
	\end{subfigure}
    \unskip\ \vrule\
   \hspace{0.05in}
    \begin{subfigure}[b]{0.25\textwidth}
        \centering
        \includegraphics[width=0.90\linewidth]{figures/heval_2.jpg}
        \vspace{0.183in}
        \caption{\scriptsize{AMT interface for evaluating the class-discriminative property}}
		\label{fig:hs_clsdisc}
	\end{subfigure}
    \unskip\ \vrule\ \hspace{0.05in}
	\begin{subfigure}[b]{0.45\textwidth}
        \centering
        \includegraphics[width=0.95\linewidth]{figures/heval_3.jpg}
        \vspace{0.01in}
        \caption{\scriptsize{AMT interface for evaluating if our visualizations instill trust in an end user}}\label{fig:hs_trust}
	\end{subfigure}
    \vspace{12pt}
	\caption{\rp{AMT interfaces for evaluating different visualizations for class discrimination (b) and trustworthiness (c). \cgb{} outperforms baseline approaches (Guided-backprop and \dec) showing that our visualizations are more class-discriminative and help humans place trust in a more accurate classifier.}}
\label{fig:human_studies}
\end{figure*}

\vspace{-15pt}
\subsection{Pointing Game}\label{sec:pointing_game}
\vspace{-2pt}

Zhang \etal~\cite{zhang2016top} introduced the Pointing Game experiment to evaluate the discriminativeness of
different visualization methods for localizing target objects in scenes.
Their evaluation protocol first cues each visualization technique with the ground-truth object label
and extracts the maximally activated point on the generated heatmap. 
It then evaluates if the point lies within one of the annotated instances of the target object category,
thereby counting it as a hit or a miss.

The localization accuracy is then calculated as \\$Acc = \frac{\#Hits}{\#Hits+\#Misses}$.
However, this evaluation only measures precision of the visualization technique.
We modify the protocol to also measure recall --
we compute localization maps for top-5 class predictions from the
CNN classifiers\footnote{We use GoogLeNet finetuned on COCO, as provided by ~\cite{zhang2016top}.}
and evaluate them using the pointing game setup with an additional option to
reject any of the top-5 predictions from the model if the maximally activated
point in the map is below a threshold,
\rp{\ie if the visualization correctly rejects the predictions which are absent
from the ground-truth categories, it gets that as a hit.}
We find that \gcam{} outperforms c-MWP~\cite{zhang2016top} by a significant
margin (70.58\% \vs 60.30\%).
Qualitative examples comparing c-MWP~\cite{zhang2016top} and \gcam{} on can be found in \secref{sec:sup_pointing}\footnote{
c-MWP~\cite{zhang2016top} highlights arbitrary regions for predicted but non-existent categories, unlike \gcam{} maps which typically do not.}.

 \vspace{-10pt}
\section{Evaluating Visualizations}\label{sec:human_evaluation}

\ad{In this section, we describe the human studies and experiments we conducted to understand
    the interpretability \vs faithfulness tradeoff of our approach to model predictions.}
Our first human study evaluates the main premise of our approach -- are \gcam{} visualizations more
class discriminative than previous techniques?
\mac{Having established that, we turn to understanding whether it can lead an end user to trust
    the visualized models appropriately.
    For these experiments, we compare VGG-16 and AlexNet finetuned on PASCAL VOC 2007
    \texttt{train} and visualizations evaluated on \texttt{val}.}

\vspace{-15pt}
\subsection{Evaluating Class Discrimination} \label{sec:class_disc}

\mac{In order to measure whether \gcam{} helps distinguish between classes,
we select images from the PASCAL VOC 2007 val set, which contain exactly $2$ annotated categories
and create visualizations for each one of them.}
For both VGG-16 and AlexNet CNNs, we obtain \rp{category-specific} visualizations using four techniques:
\dec{}, \gb{}, and \gcam{} versions of each of these methods (\cdec{} and \cgb{}).
We show these visualizations to 43 workers on Amazon Mechanical Turk (AMT) and ask
them ``Which of the two object categories is depicted in the image?''
(shown in \figref{fig:human_studies}).





Intuitively, a good prediction explanation is one that produces discriminative visualizations for the class of interest.
The experiment was conducted using all 4 visualizations for 90 image-category pairs (\ie 360 visualizations); 9 ratings were collected for each image,
evaluated against the ground truth and averaged to obtain the accuracy in \reftab{tab:eval_vis}.
When viewing \cgb{}, human subjects can correctly identify the category being visualized in $61.23$\%
of cases (compared to $44.44$\% for \gb{}; thus, \gcam{} improves human performance by $16.79$\%).
Similarly, we also find that \gcam{} helps make \dec{} more class-discriminative
(from $53.33$\% $\rightarrow$ $60.37$\%). \cgb{} performs the best among all methods.
Interestingly, our results indicate that \dec{} is more class-discriminative
than \gb{} ($53.33$\% \vs $44.44$\%),
although \gb{} is more aesthetically pleasing.
To the best of our knowledge, our evaluations are the first to quantify \mac{this subtle difference.}



\begin{table}[h!]
\vspace{-15pt}
\centering
\resizebox{.48\textwidth}{!}{
    \begin{tabular}{c p{3.0cm} p{1.7cm} p{3.1cm}}\toprule
        \textbf{Method} & \textbf{Human Classification Accuracy} & \textbf{Relative Reliability} & \textbf{Rank Correlation \;\;\; w/ Occlusion} \\
        \midrule
        \gb{}  & 44.44 & +1.00 & 0.168 \\
        \cgb{} & 61.23 & +1.27 & 0.261 \\
        \bottomrule
    \end{tabular}
}
    \vspace{2pt}
    \caption{\review{Quantitative Visualization Evaluation.
\cgb{} enables humans to differentiate between visualizations of different classes (Human Classification Accuracy) and 
pick more reliable models (Relative Reliability). It also accurately reflects the behavior of the model (Rank Correlation w/ Occlusion).}}
\label{tab:eval_vis}
\end{table}

\vspace{-15pt}
\subsection{Evaluating Trust}
Given two prediction explanations, we evaluate which seems more trustworthy.
We use AlexNet and VGG-16 to compare \gb{} and \cgb{} visualizations, noting
that VGG-16 is known to be more reliable than AlexNet with an accuracy of $79.09$ mAP
(\vs $69.20$ mAP) on PASCAL classification.
In order to tease apart the efficacy of the visualization from the accuracy of the model being visualized, we consider only those instances where \emph{both} models made the same prediction as ground truth. Given a visualization from AlexNet and one from VGG-16, and the predicted object category, \rp{54 AMT workers} were instructed to rate the reliability of the models relative to each other on a scale of clearly more/less reliable (+/-$2$), slightly more/less reliable (+/-$1$), and equally reliable ($0$).
This interface is shown in \figref{fig:human_studies}.
To eliminate any biases, VGG-16 and AlexNet were assigned to be `model-1' with approximately equal probability.
Remarkably, as can be seen in \reftab{tab:eval_vis}, we find that human subjects are able to identify the more accurate
classifier (VGG-16 over AlexNet) \emph{\iccv{simply from the prediction explanations,
despite both models making identical predictions.}}
With \gb{}, humans assign VGG-16 an average score of $1.00$ which means that it is
slightly more reliable than AlexNet, while \cgb{} achieves a higher score of $1.27$
which is closer to saying that VGG-16 is clearly more reliable.
Thus, our visualizations can help users place trust in a model that generalizes better,
just based on individual prediction explanations.



\vspace{-20pt}
\subsection{Faithfulness \vs Interpretability}\label{sec:occ}

Faithfulness of a visualization to a model is its ability to accurately explain the function learned by the model.
Naturally, there exists a trade-off between the interpretability and faithfulness of a visualization --
a more faithful visualization is typically less interpretable and \viceversa.
In fact, one could argue that a fully faithful explanation is the entire description of the model, which in the case of deep models is not interpretable/easy to visualize.
We have verified in previous sections that our visualizations are reasonably interpretable.
We now evaluate how faithful they are to the underlying model.
One expectation is that our explanations should be locally accurate, \ie in the vicinity of the input data point, our explanation should be faithful to the model~\cite{lime_sigkdd16}.


For comparison, we need a reference explanation with high local-faithfulness.
One obvious choice for such a visualization is image occlusion~\cite{zeiler_eccv14}, where we measure the difference in CNN scores when patches of the input image are masked. \rp{Interestingly, patches which change the CNN score are also patches to which \gcam{} and \cgb{} assign high intensity, achieving rank correlation $0.254$ and $0.261$ (\vs $0.168$, $0.220$
and $0.208$ achieved by \gb{}, c-MWP and CAM respectively) averaged over 2510
images in the PASCAL 2007 val set. This shows that \gcam{} is more faithful to
the original model compared to prior methods.
Through localization experiments and human studies, we
see that \gcam{} visualizations are \emph{more interpretable}, and through
correlation with occlusion maps, we see that \gcam{} is \emph{more faithful} to the model.}

 \vspace{-75pt}
\section{Diagnosing image classification CNNs with \gcam{}}\label{sec:benefits}

\ijcv{In this section we further demonstrate the use of \gcam{} in analyzing failure modes of image classification CNNs, understanding the effect of adversarial noise, and identifying and removing biases in datasets, in the context of VGG-16 pretrained on imagenet.}

\vspace{-8pt}
\subsection{Analyzing failure modes for VGG-16}\label{sec:diagnose}
\vspace{-5pt}

\begin{figure}[ht!]
    \begin{center}
    \begin{subfigure}[b]{0.23\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/failure_62.jpg}
        \caption{}
        \label{fig:failure_volcano}
    \end{subfigure}
    ~
    \begin{subfigure}[b]{0.23\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/failure_11.jpg}
        \caption{}
        \label{fig:failure_mirror}
    \end{subfigure}
    \begin{subfigure}[b]{0.23\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/failure_60.jpg}
        \caption{}
        \label{fig:failure_syringe}
    \end{subfigure}
    \begin{subfigure}[b]{0.23\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/failure_78.jpg}
        \caption{}
        \label{fig:failure_snake}
    \end{subfigure}
    \vspace{14pt}
     \caption{
In these cases the model (VGG-16) failed to predict the correct class in its top 1 (a and d) and top 5 (b and c) predictions. Humans would find it hard to explain some of these predictions without looking at the visualization for the predicted class. But with  \gcam{}, these mistakes seem justifiable.
    }
    \label{fig:failures}
    \end{center}
    \vspace{-5pt}
\end{figure}

In order to see what mistakes a network is making, we first get a list of examples
that the network (VGG-16) fails to classify correctly.
For these misclassified examples, we use \cgb{} to visualize both the correct and the predicted class.
As seen in \figref{fig:failures}, some failures are due to ambiguities inherent
in ImageNet classification. We can also see that \emph{seemingly unreasonable
predictions have reasonable explanations}, an
observation also made in HOGgles~\cite{vondrick_iccv13}.
    A major advantage of \cgb{} visualizations over other methods is that due to
    its high-resolution and ability to be class-discriminative, it readily enables these analyses.

\vspace{-8pt}
\subsection{Effect of adversarial noise on VGG-16}\label{sec:adversarial_noise}

Goodfellow \etal~\cite{goodfellow2015explaining} demonstrated the
vulnerability of current deep networks to adversarial examples, which
are slight imperceptible perturbations of input images that fool the network into
misclassifying them with high confidence. We generate adversarial
images for an ImageNet-pretrained VGG-16 model such that it assigns high probability
($>0.9999$) to a category that is not present in the image and low probabilities
to categories that are present. We then compute
\gcam{} visualizations for the categories that are present.
As shown in \reffig{fig:adversarial_main}, despite the network
being certain about the absence of these categories (`tiger cat' and `boxer'),
\gcam{} visualizations can correctly localize them.
This shows that \gcam{} is fairly robust to adversarial noise.
\begin{figure}[ht!]
\vspace{-2pt}
  \begin{center}
    \begin{subfigure}[b]{0.32\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/caffe2_figs/vgg_orig_cat_dog.jpg}
        \tiny{Boxer: 0.4   Cat: 0.2} \caption{\tiny{Original image}}

        \label{fig:orig_cat_dog}
    \end{subfigure}
    \begin{subfigure}[b]{0.32\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/caffe2_figs/pos_vgg_405_4_cat_dog.jpg}
        \tiny{Airliner: 0.9999\\
        }
		\caption{\tiny{Adversarial image}}
        \label{fig:adv_cat_dog_airliner}
    \end{subfigure}
    \begin{subfigure}[b]{0.32\linewidth}
        \centering
        \includegraphics[width=1\linewidth]{figures/gcam_pos_vgg_405_cat_dog_243.jpg}
        \tiny{Boxer: 1.1e-20}
		\caption{\tiny{\gcam{} ``Dog''}}
        \label{fig:gcam_dog_airliner}
    \end{subfigure}
    
    \begin{subfigure}[b]{0.32\linewidth}
    \vspace{10pt}
        \centering
        \includegraphics[width=1\linewidth]{figures/gcam_pos_vgg_405_4_cat_dog_283.jpg}
        \tiny{Tiger Cat: 6.5e-17\\}
		\caption{\tiny{\gcam{} ``Cat''}}
        \label{fig:gcam_cat_airliner}
	\end{subfigure}
    \begin{subfigure}[b]{0.33\linewidth}
    \vspace{10pt}
        \centering
        \includegraphics[width=1\linewidth]{figures/caffe2_figs/gcam_pos_vgg_airliner_405.jpg}
        \tiny{Airliner: 0.9999\\}
		\caption{\tiny{\gcam{} ``Airliner''}}
        \label{fig:gcam_airliner_airliner}
	\end{subfigure}
    \begin{subfigure}[b]{0.33\linewidth}
    \vspace{10pt} 
        \centering
        \includegraphics[width=1\linewidth]{figures/caffe2_figs/gcam_pos_vgg_airliner_813.jpg}
        \tiny{Space shuttle: 1e-5\\}
		\caption{\tiny{\gcam{} ``Space Shuttle''}}
        \label{fig:gcam_spaceshuttle_airliner}
	\end{subfigure}
	\vspace{14pt}
	\caption{\rp{
	  (a-b) Original image and the generated adversarial image for
	  category ``airliner''. (c-d) \gcam{} visualizations for the
	  original categories ``tiger cat'' and ``boxer (dog)'' along with their confidence. Despite the network being completely fooled into predicting the dominant category label of
	  ``airliner'' with high confidence (>0.9999), \gcam{} can localize the original categories accurately. \rpi{(e-f) \gcam{} for the top-2 predicted classes ``airliner'' and ``space shuttle'' seems to highlight the background.}}
    }
    \label{fig:adversarial_main}
    \end{center}
\end{figure}

\vspace{-20pt}
\subsection{Identifying bias in dataset}\label{sec:bias}
In this section, we demonstrate another use of \gcam{}: identifying and
reducing bias in training datasets. Models trained on biased datasets may not
generalize to real-world scenarios, or worse, may perpetuate biases and stereotypes
(\wrt gender, race, age, \etc). We finetune an ImageNet-pretrained VGG-16 model for a ``doctor'' \vs ``nurse''
binary classification task. We built our training and validation splits using
the top $250$ relevant images (for each class) from a popular image search engine.
And the test set was controlled to be balanced in its distribution
of genders across the two classes. Although the trained model achieves good validation
accuracy, it does not generalize well ($82$\% test accuracy).


\gcam{} visualizations of the model predictions (see the red box\footnote{The green and red boxes are drawn manually to highlight correct and incorrect focus of the model.} regions in the middle column of \reffig{fig:bias_qual})
revealed that the model had learned to look at the person's face / hairstyle to distinguish nurses from doctors, thus learning a gender stereotype.
Indeed, the model was misclassifying several female doctors to be a nurse and male nurses to be a doctor. Clearly, this is problematic. Turns out the image search results were gender-biased (78\% of images for doctors were men, and 93\% images for nurses were women).

\begin{figure}[htp]
 \centering
 \includegraphics[width=1\linewidth]{figures/bias_qualitative.jpg}
 \caption{In the first row, we can see that even though both models made the right decision, the biased model (model1) was looking at the face of the person to decide if the person was a nurse, whereas the unbiased model was looking at the short sleeves to make the decision. For the example image in the second row, the biased model made the wrong prediction (misclassifying a doctor as a nurse) by looking at the face and the hairstyle, whereas the unbiased model made the right prediction looking at the white coat, and the stethoscope.
 }
 \label{fig:bias_qual}
\end{figure}

Through these intuitions gained from \gcam{} visualizations,
we reduced bias in the training set by adding in images of male nurses and female doctors,
while maintaining the same number of images per class as before.
The re-trained model not only generalizes better ($90$\% test accuracy),
but also looks at the right regions (last column of \reffig{fig:bias_qual}).
This experiment demonstrates a proof-of-concept that \gcam{} can help detect and
remove biases in datasets, which is important not just for better generalization,
but also for fair and ethical outcomes as more algorithmic decisions are made in society.





 \vspace{-10pt}
\section{Textual Explanations with \gcam{}}\label{sec:text_exp}

\begin{figure*}[h]
\begin{center}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/1_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/2_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/3_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/4_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/1_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/2_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
  \caption{\ijcv{Examples showing visual explanations and textual explanations for VGG-16 trained on Places365 dataset~\cite{zhou2017places}. For textual explanations we provide the most important neurons for the predicted class along with their names. Important neurons can be either be persuasive (positive importance) or inhibitive (negative importance). The first 2 rows show success cases, and the last row shows 2 failure cases. We see that in (a), the important neurons computed by \eqref{eq:alpha1} look for concepts such as book and shelf which are indicative of class `Book-store' which is fairly intuitive. }}
  \label{fig:text_explanations}
\end{center}
\end{figure*}



Equation. \eqref{eq:alpha1} gives a way to obtain neuron-importance, $\alpha{}$, for each neuron in a convolutional layer for a particular class.
There have been hypotheses presented in the literature \cite{Zhou2014ObjectDE,zeiler_eccv14} that neurons act as concept `detectors'.
Higher positive values of the neuron importance indicate that the presence of that concept leads to an increase in the class score,
whereas higher negative values indicate that its absence leads to an increase in the score for the class.

Given this intuition, let's examine a way to generate textual explanations.
In recent work, Bau~\etal~\cite{netdissect} proposed an approach to automatically name
neurons in any convolutional layer of a trained network.
These names indicate concepts that the neuron looks for in an image.
Using their approach. we first obtain neuron names for the last convolutional layer. Next, we sort and obtain the top-5 and bottom-5 neurons based on their class-specific importance scores, $\alpha_k$.
The names for these neurons can be used as text explanations.

\reffig{fig:text_explanations} shows some examples of visual and textual explanations for the
image classification model (VGG-16) trained on the Places365 dataset~\cite{zhou2017places}.
In (a), the positively important neurons computed by \eqref{eq:alpha1} look for intuitive
concepts such as book and shelf that are indicative of the class `Book-store'.
Also note that the negatively important neurons look for concepts such as sky, road, water and car which don't occur in `Book-store' images.
In (b), for predicting `waterfall', both visual and textual explanations highlight
`water' and `stratified' which are descriptive of `waterfall' images.
(e) is a failure case due to misclassification as the network predicted
`rope-bridge' when there is no rope, but still the important concepts (water
and bridge) are indicative of the predicted class.
In (f), while \gcam{} correctly looks at the door and the staircase on the
paper to predict `Elevator door', the neurons detecting doors did not pass the
IoU threshold\footnote{Area of overlap between ground truth concept annotation and neuron activation over area of their union. More details of this metric can be found in \cite{netdissect}} of 0.05 (chosen in order to suppress the noise in the neuron names),
and hence are not part of the textual explanations. More qualitative examples
can be found in the \refsec{sec:sup_text_exp}.


 \begin{figure*}[ht!]
    \centering
  \begin{subfigure}[t]{0.49\textwidth}
    \includegraphics[width=\textwidth]{figures/captioning_main.jpg}
    \caption{\scriptsize{Image captioning explanations}}
  \label{fig:captioning}
  \end{subfigure}
  ~
    \begin{subfigure}[t]{0.49\textwidth}
    \includegraphics[width=\textwidth]{figures/densecap_main.jpg}
    \caption{\scriptsize{Comparison to DenseCap}}
   \label{fig:densecap}
   \end{subfigure}
   \vspace{14pt}
    \caption{Interpreting image captioning models: We use our class-discriminative localization technique, \gcam{} to find spatial support regions for captions in images. \reffig{fig:captioning} Visual explanations from image captioning model~\cite{karpathy2015deep} highlighting image regions considered to be important for producing the captions. \reffig{fig:densecap} \gcam{} localizations of a \emph{global} or \emph{holistic} captioning model for captions generated by a dense
    captioning model~\cite{johnson_cvpr16} for the three bounding box proposals marked on the left. We can see that we get back \gcam{} localizations (right) that agree with those bounding boxes -- even though the captioning model and Grad-CAM techniques do not use any bounding box annotations.}
\end{figure*}

\begin{figure*}[h]
\begin{center}
\begin{subfigure}[t]{\columnwidth}
\hspace{-20pt}
\includegraphics[scale=0.25]{figures/bike_bus_bench_show_tell.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.25]{figures/people_bicycles_bird_show_tell.jpg}\caption{}
\vspace{10pt}

\end{subfigure}
\caption{\rpi{Qualitative Results for our word-level captioning experiments: (a) Given the image on the left and the caption, we visualize \gcam{} maps for the visual words ``bike", ``bench" and ``bus". Note how well the \gcam{} maps correlate with the COCO segmentation maps on the right column. (b) shows a similar example where we visualize \gcam{} maps for the visual words ``people", ``bicycle" and ``bird".}}
\label{fig:captioning_word}
\end{center}
\end{figure*}

\vspace{-10pt}
\section{\gcam{} for Image Captioning and VQA}
Finally, we apply \gcam{} to vision \& language tasks such as image captioning~\cite{chen2015microsoft,johnson_cvpr16,vinyals_cvpr15}
and Visual Question Answering (VQA)~\cite{antol2015vqa,gao2015you,malinowski_iccv15,ren_nips15}.
We find that \gcam{} leads to interpretable visual explanations for these tasks
as compared to baseline visualizations which do not change noticeably across changing predictions.
Note that existing visualization techniques either are not class-discriminative
(\gb{}, \dec{}), or simply cannot be used for these tasks/architectures,
or both (CAM, c-MWP).

\vspace{-10pt}
\subsection{Image Captioning}\label{sec:nic}

  In this section, we visualize spatial support for an image captioning model using \gcam{}.
  We build \gcam{} on top of the publicly available neuraltalk2\footnote{\url{https://github.com/karpathy/neuraltalk2}} implementation~\cite{karpathy2015deep} that uses a finetuned VGG-16 CNN for images and an LSTM-based language model. Note that this model does not have an explicit attention mechanism.
  Given a caption, we compute the gradient of its log probability~\wrt units in the last convolutional layer of the CNN ($conv5\_3$ for VGG-16) and generate \gcam{} visualizations as described in \secref{sec:approach}.
See \reffig{fig:captioning}.
In the first example, \gcam{} maps for the generated caption localize
every occurrence of both the kites and people despite their relatively small size.
In the next example, \gcam{} correctly highlights the pizza and the man, but ignores the woman nearby, since `woman' is not mentioned in the caption. More examples are in \refsec{sec:sup_experiments}.

\noindent \para{Comparison to dense captioning}
Johnson~\etal~\cite{johnson_cvpr16} recently introduced the Dense Captioning (DenseCap) task that requires a system to jointly localize and caption salient regions in a given image.
Their model consists of a Fully Convolutional Localization Network (FCLN)
that produces bounding boxes for regions of interest
and an LSTM-based language model that generates associated captions, all in a
single forward pass.
Using DenseCap, we generate \iccv{5 region-specific captions per image with associated ground truth bounding boxes.
\gcam{} for a whole-image captioning model (neuraltalk2) should localize the bounding box
the region-caption was generated for, which is shown in \reffig{fig:densecap}.
We quantify this by computing the ratio of mean activation inside \vs
outside the box. Higher ratios are better because they indicate stronger
attention to the region the caption was generated for.
Uniformly highlighting the whole image results in a baseline ratio of $1.0$ whereas
\gcam{} achieves $3.27$ $\pm$ $0.18$. Adding high-resolution detail gives an improved
baseline of $2.32$ $\pm$ 0.08 (\gb{}) and the best localization at $6.38$ $\pm$ 0.99 (\cgb{}).
Thus, \gcam{} is able to localize regions in the image that the
DenseCap model describes, even though the holistic captioning model was never
trained with bounding-box annotations.}

\vspace{-30pt}
\rpi{\subsubsection{Grad-CAM for individual words of caption}}

\rpi{In our experiment we use the Show and Tell model \cite{vinyals_cvpr15} pre-trained on MSCOCO without fine-tuning through the visual representation obtained from Inception \cite{szegedy2016rethinking} architecture. 
In order to obtain \gcam{} map for individual words in the ground-truth caption we one-hot encode each of the visual words at the corresponding time-steps and compute the neuron importance score using Eq. \eqref{eq:alpha1} and combine with the convolution feature maps using Eq. \eqref{eq:gcam}. 
}

\vspace{5pt}
\noindent
\rpi{\textbf{Comparison to Human Attention}~
We manually created an object category to word mapping that maps object categories
like $<$person$>$ to a list of potential fine-grained labels like
[“child”, “man”, "woman", ...]. 
We map a total of 830 visual words existing in COCO captions to 80 COCO categories.
We then use the segmentation annotations for the 80 categories as human attention for this subset of matching words.
}

\rpi{We then use the pointing evaluation from \cite{zhang2016top}. 
For each visual word from the caption, we generate the \gcam{} map and then extract the maximally activated point. 
We then evaluate if the point lies within the human attention mapsegmentation for the corresponding COCO category, thereby counting it as a hit or a miss. 
The pointing accuracy is then calculated as \\$Acc = \frac{\#Hits}{\#Hits+\#Misses}$. 
We perform this experiment on 1000 randomly sampled images from COCO dataset and obtain an accuracy of 30.0\%. 
Some qualitative examples can be found in \reffig{fig:captioning_word}. 
}










\vspace{10pt}
\subsection{Visual Question Answering}\label{sec:vqa}

Typical VQA pipelines ~\cite{antol2015vqa,gao2015you,malinowski_iccv15,ren_nips15}
consist of a CNN to process images and an RNN language model for questions.
The image and the question representations are fused to predict the answer,
typically with a $1000$-way classification ($1000$ being the size of the answer space).
Since this is a classification problem, we pick an answer (the score $y^c$ in \eqref{eq:scores})
and use its score to compute \gcam{} visualizations over the image to explain the answer.
Despite the complexity of the task, involving both visual and textual components,
the explanations (of the VQA model from Lu~\etal~\cite{Lu2015}) described in \reffig{fig:vqa} are surprisingly intuitive and informative.
\iccv{We quantify the performance of \gcam{} via correlation with occlusion maps,
as in \secref{sec:occ}. \gcam{} achieves a rank correlation (with occlusion maps)
of 0.60 $\pm$ 0.038 whereas \gb{} achieves 0.42 $\pm$ 0.038, indicating higher
faithfulness of our \gcam{} visualization.}




\begin{figure}[t]
    \begin{subfigure}[t]{\columnwidth}
    \includegraphics[scale=0.315]{figures/vqa_main_fig.jpg}
    \vspace{5pt}
    \caption{\scriptsize{Visualizing VQA model from \cite{Lu2015}}}
    \vspace{35pt}
    \label{fig:vqa_main_fig}
    \end{subfigure}
    \begin{subfigure}[t]{\columnwidth}
    \includegraphics[scale=0.225]{figures/vqa_residual_main.jpg}
    \vspace{10pt}
    \caption{\scriptsize{Visualizing ResNet based Hierarchical co-attention VQA model from \cite{Lu2016}}}
    
        \label{fig:vqa_residual_main}
    \end{subfigure}
	\vspace{14pt}
    \caption{Qualitative Results for our VQA experiments: (a) Given the image on the left and the question ``What color is the firehydrant?'', we visualize \gcam{}s and \cgb{}s for the answers ``red", ``yellow" and ``yellow and red".
        \gcam{} visualizations are highly interpretable and help explain any target prediction -- for ``red'', the model focuses on the bottom red part of the firehydrant; when forced to answer ``yellow'', the model concentrates on it`s top yellow cap, and when forced to answer ``yellow and red", it looks at the whole firehydrant! \rp{ (b) Our approach is capable of providing interpretable explanations even for complex models.}
    }
	\label{fig:vqa}
\end{figure}
\noindent \para{Comparison to Human Attention}
Das~\etal~\cite{vqahat} collected human attention maps for a subset of the VQA dataset~\cite{antol2015vqa}.
These maps have high intensity where humans looked in the image in order to answer a visual question.
Human attention maps are compared to \gcam{} visualizations for the VQA model from~\cite{Lu2015}
on 1374 val question-image (QI) pairs from ~\cite{antol2015vqa} using the rank correlation evaluation protocol as in ~\cite{vqahat}.
\gcam{} and human attention maps have a correlation of 0.136, which is higher than chance or random attention maps (zero correlation).
This shows that despite not being trained on grounded image-text pairs, even non-attention based
CNN + LSTM based VQA models are surprisingly good at localizing regions for predicting a particular answer.

\noindent \para{Visualizing ResNet-based VQA model with co-attention}
    Lu~\etal~\cite{Lu2016} use a 200 layer ResNet~\cite{he_cvpr15} to encode the image, and jointly learn a hierarchical attention mechanism on the question and image. \reffig{fig:vqa_residual_main} shows \gcam{} visualizations for this network.
    \mac{As we visualize deeper layers of the ResNet, we see small changes in \gcam{}
    for most adjacent layers and larger changes between layers that involve dimensionality reduction.}
    \rpi{More visualizations for ResNets can be found in \refsec{sec:sup_resnet_analysis}}.
    To the best of our knowledge, we are the first to visualize decisions from ResNet-based models.





    
 \vspace{-10pt}
\section{Conclusion}
\ad{
    In this work, we proposed a novel class-discriminative localization technique --
    Gradient-weighted Class Activation Mapping (\gcam{}) -- for making \emph{any}
    CNN-based model more transparent by producing visual explanations.
    Further, we combined \gcam{} localizations with existing high-resolution visualization
    techniques to obtain the best of both worlds -- high-resolution and class-discriminative
    \cgb{} visualizations.
}
Our visualizations outperform existing approaches on both axes -- interpretability
and faithfulness to original model. Extensive human studies reveal that our
visualizations can discriminate between classes more accurately, better expose
the trustworthiness of a classifier, and help identify biases in datasets.
Further, we devise a way to identify important neurons through Grad-CAM and provide a way to obtain textual explanations for model decisions.
Finally, we show the broad applicability of \gcam{} to various off-the-shelf
architectures for tasks such as image classification, image captioning and
visual question answering. 
We believe that a true AI system should not only be intelligent, but also be able
to reason about its beliefs and actions for humans to trust and use it.
Future work includes explaining decisions made by deep networks in domains such
as reinforcement learning, natural language processing and video applications.



\vspace{-10pt}
\section{Acknowledgements}

This work was funded in part by NSF CAREER awards to
DB and DP, DARPA XAI grant to DB and DP, ONR YIP awards to DP and DB, ONR Grant
N00014-14-1-0679 to DB, a Sloan Fellowship to DP, ARO
YIP awards to DB and DP, an Allen Distinguished Investigator
award to DP from the Paul G. Allen Family Foundation,
ICTAS Junior Faculty awards to DB and DP, Google
Faculty Research Awards to DP and DB, Amazon Academic
Research Awards to DP and DB, AWS in Education
Research grant to DB, and NVIDIA GPU donations to DB. The
views and conclusions contained herein are those of the authors
and should not be interpreted as necessarily representing
the official policies or endorsements, either expressed or
implied, of the U.S. Government, or any sponsor.
 \vspace{-10pt}
\begin{appendices}



\section{Appendix Overview}

In the appendix, we provide:
 \begin{enumerate}[I]
\setlength{\itemsep}{1pt}
  \setlength{\parskip}{0pt}
  \setlength{\parsep}{0pt}
 \item - Ablation studies evaluating our design choices
 \item - More qualitative examples for image classification, captioning and VQA
 \item - More details of Pointing Game evaluation technique
 \item - Qualitative comparison to existing visualization techniques
 \item - More qualitative examples of textual explanations
 \end{enumerate}



 

\vspace{-15pt}
\section{Ablation studies} \label{sec:ablation}

We perform several ablation studies to explore and validate our design choices for computing \gcam{} visualizations. 
This includes visualizing different layers in the network, understanding importance of ReLU in \eqref{eq:gcam}, analyzing different types of gradients (for ReLU backward pass), and different gradient pooling strategies.




\vspace{-8pt}
\subsection*{1. \gcam{} for different layers}

We show \gcam{} visualizations for the ``tiger-cat'' class at different convolutional
layers in AlexNet and VGG-16.
As expected, the results from \reffig{fig:conv_layers} show that localization
becomes progressively worse as we move to earlier convolutional layers.
This is because later convolutional layers better capture high-level semantic information
while retaining spatial information than earlier layers, that have smaller
receptive fields and only focus on local features.

\begin{figure*}[t!]
	\centering
    \includegraphics[width=1\textwidth]{figures/vgg16_conv_layers_new.jpg}\\
    \vspace{-5pt}
\label{fig:vgg16conv}
        \vspace{5pt}
		\caption{ \gcam{} at different convolutional layers for the `tiger cat' class.
            This figure analyzes how localizations change qualitatively as we perform \gcam{} with respect to different feature maps in a CNN (VGG16~\cite{simonyan_arxiv14}).
	We find that the best looking visualizations are often obtained after the deepest convolutional layer in the network, and localizations get progressively worse at shallower layers.
This is consistent with our intuition described in Section 3 of main paper, that deeper convolutional layer capture more semantic concepts. }
	\label{fig:conv_layers}
\end{figure*}

\begin{figure*}
 \centering
 \includegraphics[width=1\linewidth]{figures/alex_conv_layers_new.jpg}
 \caption{\gcam{} localizations for ``tiger cat'' category for different rectified convolutional layer feature maps for AlexNet.}
 \label{fig:alex_conv}
\end{figure*}

\vspace{-5pt}
\subsection*{2. Design choices}

\vspace{-5pt}
\begin{table}[h!]
\centering
\captionsetup{belowskip=0pt,aboveskip=4pt}
    {
        \begin{tabular}{c@{\hskip 0.6in} c}\toprule
            \textbf{Method} & \textbf{Top-1 Loc error} \\
            \midrule

            \gcam{}                                                  & 59.65          \\
            \midrule
            \gcam{} without ReLU in Eq.1                             & 74.98          \\
            \gcam{} with Absolute gradients                          & 58.19          \\
            \gcam{} with GMP gradients                               & 59.96          \\
            \gcam{} with Deconv ReLU                                 & 83.95          \\
            \gcam{} with Guided ReLU                                 & 59.14          \\
            \bottomrule
        \end{tabular}

    }


    \caption{Localization results on ILSVRC-15 val for the ablations. Note that
    this evaluation is over 10 crops, while visualizations are single crop.}

    \label{table:locresablation}
\end{table}

We evaluate different design choices via top-1 localization errors on the ILSVRC-15
val set~\cite{imagenet_cvpr09}. See \reftab{table:locresablation}.
\subsubsection*{2.1. Importance of ReLU in \eqref{eq:scores}}

Removing ReLU (\eqref{eq:scores}) increases error by 15.3\%.
Negative values in \gcam{} indicate confusion between multiple occurring classes.






\subsubsection*{2.2. Global Average Pooling \vs Global Max Pooling}

Instead of Global Average Pooling (GAP) the incoming gradients to the convolutional layer,
we tried Global Max Pooling (GMP). We observe that using GMP lowers the localization ability
of \gcam{}. An example can be found in \reffig{fig:gmp} below. This may be due to the fact
that max is statistically less robust to noise compared to the averaged gradient.

\begin{figure}[h]

     \centering
     \includegraphics[width=1\linewidth]{figures/gcam_ablation_gap_gmp.jpg}
     \caption{\gcam{} visualizations for ``tiger cat'' category with Global Average Pooling and Global Max Pooling.}
     \label{fig:gmp}

   \end{figure}

\subsubsection*{2.3. Effect of different ReLU on \gcam{}}

We experiment with Guided-ReLU~\cite{springenberg_arxiv14} and Deconv-ReLU~\cite{zeiler_eccv14}
as modifications to the backward pass of ReLU.

\textbf{Guided-ReLU:} Springenberg~\etal\cite{springenberg_arxiv14} introduced Guided Backprop,
where the backward pass of ReLU is modified to only pass positive gradients
to regions of positive activations.
Applying this change to the computation of \gcam{} introduces a drop in the
class-discriminative ability as can be seen in \figref{fig:relu},
but it marginally improves localization performance as can be seen in \reftab{table:locresablation}.

\textbf{Deconv-ReLU:} In \dec{}~\cite{zeiler_eccv14}, Zeiler and Fergus introduced a
modification to the backward pass of ReLU to only pass positive gradients.
Applying this modification to the computation of \gcam{} leads to worse results
(\figref{fig:relu}). This indicates that negative gradients also carry
important information for class-discriminativeness.



\begin{figure}[h]

     \centering
     \includegraphics[width=1\linewidth]{figures/gcam_ablation_guided_deconv_ReLU.jpg}
     \caption{\gcam{} visualizations for ``tiger cat'' category for different modifications to the ReLU backward pass. The best results are obtained when we use the actual gradients during the computation of \gcam{}.}
     \label{fig:relu}
   \end{figure}


 \vspace{-5pt}
\section{Qualitative results for vision and language tasks}\label{sec:sup_experiments}

In this section we provide more qualitative results for \gcam{} and \cgb{} applied to the task of image classification, image captioning and VQA.
\vspace{-5pt}
\subsection*{1. Image Classification}

\begin{figure*}
 \centering
 \includegraphics[width=1\linewidth]{figures/classification_supp.jpg}
 \caption{Visualizations for randomly sampled images from the COCO validation dataset.
 Predicted classes are mentioned at the top of each column.}
 \label{fig:classification}
\end{figure*}


We use \gcam{} and \cgb{} to visualize the regions of the image that provide support for a particular prediction.
The results reported in \reffig{fig:classification} correspond to the VGG-16 \cite{simonyan_arxiv14} network trained on ImageNet.


\reffig{fig:classification} shows randomly sampled examples from COCO~\cite{Lin_2014} validation set.
COCO images typically have multiple objects per image and \gcam{} visualizations show precise localization to support the model's prediction.

\cgb{} can even localize tiny objects. For example our approach correctly localizes the predicted class ``torch'' (\reffig{fig:classification}.a) inspite of its size and odd location in the image. Our method is also class-discriminative -- it places attention \emph{only} on the ``toilet seat'' even when a popular ImageNet category ``dog'' exists in the image (\reffig{fig:classification}.e).


We also visualized Grad-CAM, Guided Backpropagation (GB), Deconvolution (DC), GB + \gcam{} (\cgb{}), DC + \gcam{} (\cdec{}) for images from the ILSVRC13 detection val set that have at least 2 unique object categories each.
The visualizations for the mentioned class can be found in the following links.

``computer keyboard, keypad'' class: \\http://i.imgur.com/QMhsRzf.jpg

``sunglasses, dark glasses, shades'' class: \\ http://i.imgur.com/a1C7DGh.jpg

\vspace{-10pt}
\subsection*{2. Image Captioning}

\begin{figure*}
     \centering
     \includegraphics[width=0.83\linewidth]{figures/captioning_supp_new.jpg}
     \caption{\gb{}, \gcam{}  and \cgb{}  visualizations for the captions produced by the Neuraltalk2 image captioning model.}
     \label{fig:sup_captioning}
\end{figure*}

\begin{figure*}
     \centering
     \includegraphics[scale=0.30]{figures/vqa_sup.jpg}
     \caption{\gb{}, \gcam{} and \cgb{}  visualizations for the answers from a VQA model.
     For each image-question pair, we show visualizations for AlexNet, VGG-16 and VGG-19.
     Notice how the attention changes in row 3, as we change the answer from \emph{Yellow} to \emph{Green}.}
     \label{fig:vqa_supp}
\end{figure*}

\begin{figure*}[htp]
 \centering
 \includegraphics[width=0.65\linewidth]{figures/gcam_pascal.jpg}
 \caption{Visualizations for ground-truth categories (shown below each image) for images sampled from the PASCAL~\cite{pascal-voc-2007} validation set.}
 \label{fig:gcam_pascal}
\end{figure*}


We use the publicly available Neuraltalk2 code and model\footnote{\url{https://github.com/karpathy/neuraltalk2}} for our image captioning experiments.
The model uses VGG-16 to encode the image.
The image representation is passed as input at the first time step to an LSTM that generates a caption for the image.
The model is trained end-to-end along with CNN finetuning using the COCO~\cite{Lin_2014} Captioning dataset.
We feedforward the image to the image captioning model to obtain a caption. \rp{We use \gcam{} to get a coarse localization and combine it with \gb{} to get a high-resolution visualization that highlights regions in the image that provide support for the generated caption.} 


\subsection*{3. Visual Question Answering (VQA)}

\rp{We use \gcam{} and \cgb{} to explain why \rp{a} publicly available VQA model \cite{Lu2015} answered what it answered.}

The VQA model by Lu~\etal uses a standard CNN followed by a fully connected layer to transform the image to 1024-dim to match the \rp{LSTM} embeddings of the question. Then the transformed image and LSTM embeddings are pointwise multiplied to get a combined representation of the image and question and a multi-layer perceptron is trained on top to predict one among 1000 answers.
We show visualizations for the VQA model trained with 3 different CNNs - AlexNet~\cite{krizhevsky_nips12}, VGG-16 and VGG-19~\cite{simonyan_arxiv14}.
Even though the CNNs were not finetuned for the task of VQA, it is interesting to see how our approach can serve as a tool to understand these networks better by providing a localized high-resolution visualization of the regions the model is looking at. Note that these networks were trained with no explicit attention mechanism enforced.

Notice in the first row of \reffig{fig:vqa_supp}, for the question, ``\emph{Is the person riding the waves?}'', the VQA model with AlexNet and VGG-16 answered ``No'', as they concentrated on the person mainly, and not the waves.
On the other hand, VGG-19 correctly answered ``Yes'', and it looked at the regions around the man in order to answer the question.
In the second row, for the question, ``\emph{What is the person hitting?}'', the VQA model trained with AlexNet answered ``Tennis ball'' just based on context without looking at the ball.
Such a model might be risky when employed in real-life scenarios.
It is difficult to determine the trustworthiness of a model just based on the predicted answer.
Our visualizations provide an accurate way to explain the model's predictions and help in determining which model to trust, without making any architectural changes or sacrificing accuracy.
Notice in the last row of \reffig{fig:vqa_supp}, for the question, ``\emph{Is this a whole orange?}'', the model looks for regions around the orange to answer ``No''.

\begin{figure*}[h]
\begin{center}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/5_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/6_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/7_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/8_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/9_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/10_pos.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/3_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/4_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/5_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
\begin{subfigure}[t]{\columnwidth}
\includegraphics[scale=0.135]{figures/6_neg.jpg}\caption{}
\vspace{10pt}
\end{subfigure}
  \caption{More Qualitative examples showing visual explanations and textual explanations for VGG-16 trained on Places365 dataset (\cite{zhou2017places}). For textual explanations we provide the most important neurons for the predicted class along with their names. Important neurons can be either be persuasive (positively important) or inhibitive (negatively important). 
  The first 3 rows show positive examples, and the last 2 rows show failure cases.}
  \label{fig:sup_text_explanations}
\end{center}
\end{figure*}

 \vspace{-10pt}
\section{More details of Pointing Game}\label{sec:sup_pointing}

In \cite{zhang2016top}, the pointing game was setup to evaluate the discriminativeness of different attention maps for localizing ground-truth categories. In a sense, this evaluates the precision of a visualization, \ie how often does the attention map intersect the segmentation map of the ground-truth category.
This does not evaluate how often the visualization technique produces maps which do not correspond to the category of interest. 

Hence we propose a modification to the pointing game to evaluate visualizations of the top-5 predicted category.
In this case the visualizations are given an additional option to reject any of the top-5 predictions from the CNN classifiers.
For each of the two visualizations, \gcam{} and c-MWP, we choose a threshold on the max value of the visualization, that can be used to determine if the category being visualized exists in the image.

We compute the maps for the top-5 categories, and based on the maximum value in the map, we try to classify if the map is of the GT label or a category that is absent in the image.
As mentioned in Section 4.2 of the main paper, we find that our approach \gcam{} outperforms c-MWP by a significant margin (70.58\% vs 60.30\% on VGG-16).





 

\vspace{-10pt}
\section{Qualitative comparison to Excitation Backprop (c-MWP) and CAM}\label{sec:sup_comparison}
\vspace{-5pt}
In this section we provide more qualitative results comparing \gcam{} with CAM~\cite{zhou_cvpr16} and c-MWP~\cite{zhang2016top} on Pascal~\cite{pascal-voc-2007}. 




We compare \gcam{}, CAM and c-MWP visualizations from ImageNet trained VGG-16 models finetuned on PASCAL VOC 2012 dataset.
While \gcam{} and c-MWP visualizations can be directly obtained from existing models, CAM requires an architectural change, and requires re-training, which leads to loss in accuracy.
Also, unlike \gcam{}, c-MWP and CAM can only be applied for image classification networks.
Visualizations for the ground-truth categories can be found in \reffig{fig:gcam_pascal}. 






 
\begin{figure*}
     \centering
        \begin{subfigure}[t]{0.49\textwidth}
        \begin{center}
     \includegraphics[width=\linewidth]{figures/resnet_200_analysis.jpg}
     \caption{\gcam{}  visualizations for the ResNet-200 layer architecture for 'tiger cat'(left) and 'boxer'(right) category.}
 \end{center}
    \end{subfigure}
    \begin{subfigure}[t]{0.49\textwidth}
        \begin{center}
     \includegraphics[width=\linewidth]{figures/resnet_200_analysis-2.jpg}
     \caption{\gcam{}  visualizations for the ResNet-200 layer architecture for 'tabby cat'(left) and 'boxer'(right) category.}
        \end{center}
    \end{subfigure}
    \vspace{18pt}
    \caption{We observe that the discriminative ability of \gcam{} significantly reduces as we encounter the downsampling layer.}
     \label{fig:sup_resnet}
\end{figure*}


\vspace{-10pt}
\section{Visual and Textual explanations for Places dataset}\label{sec:sup_text_exp}
\vspace{-15pt}
\reffig{fig:sup_text_explanations} shows more examples of visual and textual explanations (\refsec{sec:text_exp}) for the image classification model (VGG-16) trained on Places365 dataset (\cite{zhou2017places}).



 \vspace{-18pt}
\section{Analyzing Residual Networks}\label{sec:sup_resnet_analysis}
\vspace{-5pt}


In this section, we perform Grad-CAM on Residual Networks (ResNets). In particular, we analyze the 200-layer architecture trained on ImageNet\footnote{We use the 200-layer ResNet architecture from \url{https://github.com/facebook/fb.resnet.torch.}}.

Current ResNets~\cite{he_cvpr15} typically consist of residual blocks.
One set of blocks use identity skip connections (shortcut connections between two layers having identical output dimensions).
These sets of residual blocks are interspersed with downsampling modules that alter dimensions of propagating signal.
As can be seen in \reffig{fig:sup_resnet} our visualizations applied on the last convolutional layer can correctly localize the cat and the dog.
\gcam{} can also visualize the cat and dog correctly in the residual blocks of the last set.
However, as we go towards earlier sets of residual blocks with different spatial resolution, we see that \gcam{} fails to localize the category of interest (see last row of \reffig{fig:sup_resnet}).
We observe similar trends for other ResNet architectures (18 and 50-layer).



 
\end{appendices} 
\vspace{-5pt}
{\footnotesize
\bibliographystyle{ieee}
\bibliography{strings,main}
}











\end{document}