{
 "cells": [
  {
   "cell_type": "markdown",
   "metadata": {},
   "source": [
    "一、项目背景\n",
    "-\n",
    "通过飞桨提供的高层API，调用飞桨封装好的ResNet模型。学习ResNet模型的特点"
   ]
  },
  {
   "cell_type": "markdown",
   "metadata": {},
   "source": [
    "二、数据介绍\n",
    "-\n",
    "使用Cifar10数据集,CIFAR-10数据集由10个类中的60000个32x32彩色图像组成，每个类有6000个图像。有50000个训练图像和10000个测试图像。"
   ]
  },
  {
   "cell_type": "markdown",
   "metadata": {},
   "source": [
    "三、模型介绍\n",
    "-\n",
    "ResNet是2015年ImageNet比赛的冠军，将识别错误率降低到了3.6%，这个结果甚至超出了正常人眼识别的精度。\n",
    "随着深度学习的不断发展，模型的层数越来越多，网络结构也越来越复杂。那么是否加深网络结构，就一定会得到更好的效果呢？从理论上来说，假设新增加的层都是恒等映射，只要原有的层学出跟原模型一样的参数，那么深模型结构就能达到原模型结构的效果。换句话说，原模型的解只是新模型的解的子空间，在新模型解的空间里应该能找到比原模型解对应的子空间更好的结果。但是实践表明，增加网络的层数之后，训练误差往往不降反升。\n",
    "\n",
    "Kaiming He等人提出了残差网络ResNet来解决上述问题，其基本思想如 图6所示。\n",
    "\n",
    "图1(a)：表示增加网络的时候，将xxx映射成y=F(x)y=F(x)y=F(x)输出。\n",
    "图1(b)：对图1(a)作了改进，输出y=F(x)+xy=F(x) + xy=F(x)+x。这时不是直接学习输出特征yyy的表示，而是学习y−xy-xy−x。\n",
    "如果想学习出原模型的表示，只需将F(x)F(x)F(x)的参数全部设置为0，则y=xy=xy=x是恒等映射。\n",
    "F(x)=y−xF(x) = y - xF(x)=y−x也叫做残差项，如果x→yx\\rightarrow yx→y的映射接近恒等映射，图1(b)中通过学习残差项也比图6(a)学习完整映射形式更加容易。\n",
    "![](https://ai-studio-static-online.cdn.bcebos.com/b8a83393f2a64c28a23718b7f1246981efdf1b7f949a4cad8dff008a2e02ce04)\n",
    "\n",
    "\n",
    "                                           图1：残差设计思想\n",
    "\n",
    "图1(b)的结构是残差网络的基础，这种结构也叫做残差块（Residual block）。输入xxx通过跨层连接，能更快的向前传播数据，或者向后传播梯度。通俗的比喻，在火热的电视节目《王牌对王牌》上有一个“传声筒”的游戏，排在队首的嘉宾把看到的影视片段表演给后面一个嘉宾看，经过四五个嘉宾后，最后一个嘉宾如果能表演出更多原剧的内容，就能取得高分。我们常常会发现刚开始的嘉宾往往表演出最多的信息（类似于Loss），而随着表演的传递，有效的表演信息越来越少（类似于梯度弥散）。如果每个嘉宾都能看到原始的影视片段，那么相信传声筒的效果会好很多。类似的，由于ResNet每层都存在直连的旁路，相当于每一层都和最终的损失有“直接对话”的机会，自然可以更好的解决梯度弥散的问题。残差块的具体设计方案如 图2 所示，这种设计方案也常称作瓶颈结构（BottleNeck）。1*1的卷积核可以非常方便的调整中间层的通道数，在进入3*3的卷积层之前减少通道数（256->64），经过该卷积层后再恢复通道数(64->256)，可以显著减少网络的参数量。这个结构（256->64->256）像一个中间细，两头粗的瓶颈，所以被称为“BottleNeck”。\n",
    "\n",
    "![](https://ai-studio-static-online.cdn.bcebos.com/51b4659e4e2d42d6bc085b742f58d317857b63937e244f5882c815e7055d09cf)\n",
    "                                     \n",
    "                                            图2：残差块结构示意图\n",
    "                     \n",
    "下图表示出了ResNet-50的结构，一共包含49层卷积和1层全连接，所以被称为ResNet-50。\n",
    "![](https://ai-studio-static-online.cdn.bcebos.com/ff22fb48e71a471f93a12e30df5271faaec2c0e8a73c4f5c9b141b43a97ab698)\n",
    "\n",
    "                                        图3：ResNet-50模型网络结构示意图\n",
    "\n"
   ]
  },
  {
   "cell_type": "markdown",
   "metadata": {},
   "source": [
    "四、程序实现\n",
    "-\n"
   ]
  },
  {
   "cell_type": "code",
   "execution_count": 1,
   "metadata": {
    "execution": {
     "iopub.execute_input": "2022-02-23T14:59:34.446137Z",
     "iopub.status.busy": "2022-02-23T14:59:34.445288Z",
     "iopub.status.idle": "2022-02-23T14:59:35.816659Z",
     "shell.execute_reply": "2022-02-23T14:59:35.815852Z",
     "shell.execute_reply.started": "2022-02-23T14:59:34.446101Z"
    },
    "scrolled": true
   },
   "outputs": [],
   "source": [
    "# 从paddle.vision.models 模块中import 残差网络，VGG网络，LeNet网络\n",
    "import paddle\n",
    "from paddle.vision.models import resnet50, vgg16, LeNet\n",
    "from paddle.vision.datasets import Cifar10\n",
    "from paddle.optimizer import Momentum\n",
    "from paddle.regularizer import L2Decay\n",
    "from paddle.nn import CrossEntropyLoss\n",
    "from paddle.metric import Accuracy\n",
    "from paddle.vision.transforms import Transpose"
   ]
  },
  {
   "cell_type": "code",
   "execution_count": 4,
   "metadata": {
    "execution": {
     "iopub.execute_input": "2022-02-23T15:09:03.155221Z",
     "iopub.status.busy": "2022-02-23T15:09:03.154292Z",
     "iopub.status.idle": "2022-02-23T15:09:11.868122Z",
     "shell.execute_reply": "2022-02-23T15:09:11.867346Z",
     "shell.execute_reply.started": "2022-02-23T15:09:03.155171Z"
    },
    "scrolled": true
   },
   "outputs": [],
   "source": [
    "import numpy as np\n",
    "import matplotlib.pyplot as plt\n",
    "\n",
    "\n",
    "# 使用Cifar10数据集\n",
    "train_dataset = Cifar10(mode='train', transform=Transpose())\n",
    "val_dataset = Cifar10(mode='test', transform=Transpose())\n",
    "\n",
    "\n"
   ]
  },
  {
   "cell_type": "code",
   "execution_count": 5,
   "metadata": {
    "execution": {
     "iopub.execute_input": "2022-02-23T15:09:28.188873Z",
     "iopub.status.busy": "2022-02-23T15:09:28.188226Z",
     "iopub.status.idle": "2022-02-23T15:19:27.070479Z",
     "shell.execute_reply": "2022-02-23T15:19:27.068176Z",
     "shell.execute_reply.started": "2022-02-23T15:09:28.188833Z"
    },
    "scrolled": true
   },
   "outputs": [
    {
     "name": "stdout",
     "output_type": "stream",
     "text": [
      "The loss value printed in the log is the current step, and the metric is the average value of previous steps.\n",
      "Epoch 1/50\n"
     ]
    },
    {
     "name": "stderr",
     "output_type": "stream",
     "text": [
      "/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:653: UserWarning: When training, we now always track global mean and variance.\n",
      "  \"When training, we now always track global mean and variance.\")\n"
     ]
    },
    {
     "name": "stdout",
     "output_type": "stream",
     "text": [
      "step  10/782 - loss: 5.6795 - acc_top1: 0.1344 - acc_top5: 0.5219 - 152ms/step\n",
      "step  20/782 - loss: 4.6656 - acc_top1: 0.1562 - acc_top5: 0.5680 - 95ms/step\n",
      "step  30/782 - loss: 3.3192 - acc_top1: 0.1609 - acc_top5: 0.5901 - 78ms/step\n",
      "step  40/782 - loss: 2.9309 - acc_top1: 0.1707 - acc_top5: 0.6145 - 73ms/step\n",
      "step  50/782 - loss: 3.4723 - acc_top1: 0.1741 - acc_top5: 0.6297 - 71ms/step\n",
      "step  60/782 - loss: 3.3919 - acc_top1: 0.1773 - acc_top5: 0.6388 - 70ms/step\n",
      "step  70/782 - loss: 5.8369 - acc_top1: 0.1837 - acc_top5: 0.6507 - 68ms/step\n",
      "step  80/782 - loss: 8.3766 - acc_top1: 0.1893 - acc_top5: 0.6553 - 64ms/step\n",
      "step  90/782 - loss: 2.6302 - acc_top1: 0.1910 - acc_top5: 0.6585 - 62ms/step\n",
      "step 100/782 - loss: 2.5693 - acc_top1: 0.1889 - acc_top5: 0.6573 - 59ms/step\n",
      "step 110/782 - loss: 5.8991 - acc_top1: 0.1884 - acc_top5: 0.6585 - 58ms/step\n",
      "step 120/782 - loss: 3.0208 - acc_top1: 0.1904 - acc_top5: 0.6641 - 56ms/step\n",
      "step 130/782 - loss: 4.7586 - acc_top1: 0.1906 - acc_top5: 0.6696 - 55ms/step\n",
      "step 140/782 - loss: 2.2792 - acc_top1: 0.1929 - acc_top5: 0.6741 - 55ms/step\n",
      "step 150/782 - loss: 2.4172 - acc_top1: 0.1906 - acc_top5: 0.6710 - 55ms/step\n",
      "step 160/782 - loss: 3.1234 - acc_top1: 0.1907 - acc_top5: 0.6699 - 54ms/step\n",
      "step 170/782 - loss: 2.0904 - acc_top1: 0.1905 - acc_top5: 0.6710 - 53ms/step\n",
      "step 180/782 - loss: 3.9016 - acc_top1: 0.1907 - acc_top5: 0.6720 - 52ms/step\n",
      "step 190/782 - loss: 4.3320 - acc_top1: 0.1917 - acc_top5: 0.6744 - 52ms/step\n",
      "step 200/782 - loss: 3.3967 - acc_top1: 0.1953 - acc_top5: 0.6772 - 51ms/step\n",
      "step 210/782 - loss: 5.1925 - acc_top1: 0.1969 - acc_top5: 0.6781 - 50ms/step\n",
      "step 220/782 - loss: 6.8681 - acc_top1: 0.1994 - acc_top5: 0.6815 - 50ms/step\n",
      "step 230/782 - loss: 2.5929 - acc_top1: 0.1995 - acc_top5: 0.6836 - 49ms/step\n",
      "step 240/782 - loss: 2.5464 - acc_top1: 0.2026 - acc_top5: 0.6876 - 49ms/step\n",
      "step 250/782 - loss: 4.3838 - acc_top1: 0.2045 - acc_top5: 0.6893 - 48ms/step\n",
      "step 260/782 - loss: 3.2058 - acc_top1: 0.2046 - acc_top5: 0.6904 - 48ms/step\n",
      "step 270/782 - loss: 3.0605 - acc_top1: 0.2048 - acc_top5: 0.6900 - 48ms/step\n",
      "step 280/782 - loss: 4.2023 - acc_top1: 0.2055 - acc_top5: 0.6919 - 48ms/step\n",
      "step 290/782 - loss: 3.1417 - acc_top1: 0.2066 - acc_top5: 0.6942 - 48ms/step\n",
      "step 300/782 - loss: 7.3257 - acc_top1: 0.2069 - acc_top5: 0.6948 - 47ms/step\n",
      "step 310/782 - loss: 5.3402 - acc_top1: 0.2080 - acc_top5: 0.6960 - 47ms/step\n",
      "step 320/782 - loss: 4.6829 - acc_top1: 0.2090 - acc_top5: 0.6983 - 47ms/step\n",
      "step 330/782 - loss: 3.6497 - acc_top1: 0.2097 - acc_top5: 0.7001 - 47ms/step\n",
      "step 340/782 - loss: 3.1487 - acc_top1: 0.2107 - acc_top5: 0.7010 - 46ms/step\n",
      "step 350/782 - loss: 2.6579 - acc_top1: 0.2108 - acc_top5: 0.7021 - 46ms/step\n",
      "step 360/782 - loss: 3.7326 - acc_top1: 0.2118 - acc_top5: 0.7042 - 46ms/step\n",
      "step 370/782 - loss: 3.5145 - acc_top1: 0.2120 - acc_top5: 0.7048 - 46ms/step\n",
      "step 380/782 - loss: 2.9521 - acc_top1: 0.2111 - acc_top5: 0.7043 - 46ms/step\n",
      "step 390/782 - loss: 4.1713 - acc_top1: 0.2112 - acc_top5: 0.7052 - 45ms/step\n",
      "step 400/782 - loss: 2.1432 - acc_top1: 0.2116 - acc_top5: 0.7059 - 45ms/step\n",
      "step 410/782 - loss: 2.7896 - acc_top1: 0.2107 - acc_top5: 0.7054 - 45ms/step\n",
      "step 420/782 - loss: 2.4999 - acc_top1: 0.2099 - acc_top5: 0.7049 - 45ms/step\n",
      "step 430/782 - loss: 2.4027 - acc_top1: 0.2093 - acc_top5: 0.7051 - 45ms/step\n",
      "step 440/782 - loss: 2.9068 - acc_top1: 0.2098 - acc_top5: 0.7061 - 44ms/step\n",
      "step 450/782 - loss: 2.0958 - acc_top1: 0.2094 - acc_top5: 0.7073 - 44ms/step\n",
      "step 460/782 - loss: 2.4070 - acc_top1: 0.2102 - acc_top5: 0.7091 - 45ms/step\n",
      "step 470/782 - loss: 2.0297 - acc_top1: 0.2105 - acc_top5: 0.7106 - 45ms/step\n",
      "step 480/782 - loss: 3.3655 - acc_top1: 0.2116 - acc_top5: 0.7122 - 45ms/step\n",
      "step 490/782 - loss: 3.5381 - acc_top1: 0.2124 - acc_top5: 0.7136 - 45ms/step\n",
      "step 500/782 - loss: 5.4106 - acc_top1: 0.2131 - acc_top5: 0.7147 - 45ms/step\n",
      "step 510/782 - loss: 5.2342 - acc_top1: 0.2142 - acc_top5: 0.7159 - 44ms/step\n",
      "step 520/782 - loss: 5.0450 - acc_top1: 0.2149 - acc_top5: 0.7165 - 44ms/step\n",
      "step 530/782 - loss: 2.1638 - acc_top1: 0.2157 - acc_top5: 0.7175 - 44ms/step\n",
      "step 540/782 - loss: 2.8996 - acc_top1: 0.2166 - acc_top5: 0.7188 - 44ms/step\n",
      "step 550/782 - loss: 2.1775 - acc_top1: 0.2178 - acc_top5: 0.7206 - 44ms/step\n",
      "step 560/782 - loss: 2.2152 - acc_top1: 0.2189 - acc_top5: 0.7222 - 44ms/step\n",
      "step 570/782 - loss: 2.5439 - acc_top1: 0.2197 - acc_top5: 0.7235 - 44ms/step\n",
      "step 580/782 - loss: 2.0781 - acc_top1: 0.2214 - acc_top5: 0.7250 - 44ms/step\n",
      "step 590/782 - loss: 1.8310 - acc_top1: 0.2222 - acc_top5: 0.7268 - 43ms/step\n",
      "step 600/782 - loss: 4.1566 - acc_top1: 0.2235 - acc_top5: 0.7280 - 43ms/step\n",
      "step 610/782 - loss: 1.8589 - acc_top1: 0.2248 - acc_top5: 0.7294 - 43ms/step\n",
      "step 620/782 - loss: 2.9738 - acc_top1: 0.2260 - acc_top5: 0.7307 - 43ms/step\n",
      "step 630/782 - loss: 1.9902 - acc_top1: 0.2275 - acc_top5: 0.7323 - 43ms/step\n",
      "step 640/782 - loss: 3.1473 - acc_top1: 0.2297 - acc_top5: 0.7340 - 44ms/step\n",
      "step 650/782 - loss: 4.2280 - acc_top1: 0.2307 - acc_top5: 0.7350 - 44ms/step\n",
      "step 660/782 - loss: 3.1643 - acc_top1: 0.2322 - acc_top5: 0.7366 - 44ms/step\n",
      "step 670/782 - loss: 1.8322 - acc_top1: 0.2334 - acc_top5: 0.7378 - 44ms/step\n",
      "step 680/782 - loss: 2.0853 - acc_top1: 0.2351 - acc_top5: 0.7394 - 44ms/step\n",
      "step 690/782 - loss: 1.7796 - acc_top1: 0.2366 - acc_top5: 0.7413 - 43ms/step\n",
      "step 700/782 - loss: 2.1500 - acc_top1: 0.2381 - acc_top5: 0.7425 - 43ms/step\n",
      "step 710/782 - loss: 1.8981 - acc_top1: 0.2392 - acc_top5: 0.7437 - 43ms/step\n",
      "step 720/782 - loss: 1.8118 - acc_top1: 0.2406 - acc_top5: 0.7445 - 43ms/step\n",
      "step 730/782 - loss: 3.3562 - acc_top1: 0.2417 - acc_top5: 0.7454 - 43ms/step\n",
      "step 740/782 - loss: 3.4929 - acc_top1: 0.2429 - acc_top5: 0.7466 - 43ms/step\n",
      "step 750/782 - loss: 2.6125 - acc_top1: 0.2440 - acc_top5: 0.7477 - 43ms/step\n",
      "step 760/782 - loss: 1.9044 - acc_top1: 0.2451 - acc_top5: 0.7490 - 43ms/step\n",
      "step 770/782 - loss: 5.1057 - acc_top1: 0.2463 - acc_top5: 0.7501 - 43ms/step\n",
      "step 780/782 - loss: 2.2067 - acc_top1: 0.2471 - acc_top5: 0.7508 - 43ms/step\n",
      "step 782/782 - loss: 8.8959 - acc_top1: 0.2470 - acc_top5: 0.7509 - 44ms/step\n",
      "save checkpoint at /home/aistudio/output/0\n",
      "Eval begin...\n",
      "step  10/157 - loss: 25.9702 - acc_top1: 0.2859 - acc_top5: 0.7672 - 131ms/step\n",
      "step  20/157 - loss: 7.2651 - acc_top1: 0.2859 - acc_top5: 0.7711 - 75ms/step\n",
      "step  30/157 - loss: 90.5822 - acc_top1: 0.2797 - acc_top5: 0.7672 - 56ms/step\n",
      "step  40/157 - loss: 26.3394 - acc_top1: 0.2766 - acc_top5: 0.7637 - 46ms/step\n",
      "step  50/157 - loss: 2.6096 - acc_top1: 0.2812 - acc_top5: 0.7681 - 41ms/step\n",
      "step  60/157 - loss: 31.3984 - acc_top1: 0.2826 - acc_top5: 0.7711 - 37ms/step\n",
      "step  70/157 - loss: 97.8104 - acc_top1: 0.2819 - acc_top5: 0.7679 - 34ms/step\n",
      "step  80/157 - loss: 6.4152 - acc_top1: 0.2816 - acc_top5: 0.7729 - 32ms/step\n",
      "step  90/157 - loss: 11.0218 - acc_top1: 0.2748 - acc_top5: 0.7720 - 31ms/step\n",
      "step 100/157 - loss: 1.7503 - acc_top1: 0.2772 - acc_top5: 0.7759 - 29ms/step\n",
      "step 110/157 - loss: 65.1060 - acc_top1: 0.2750 - acc_top5: 0.7722 - 28ms/step\n",
      "step 120/157 - loss: 1.7679 - acc_top1: 0.2763 - acc_top5: 0.7725 - 27ms/step\n",
      "step 130/157 - loss: 105.5897 - acc_top1: 0.2755 - acc_top5: 0.7718 - 27ms/step\n",
      "step 140/157 - loss: 24.5136 - acc_top1: 0.2760 - acc_top5: 0.7725 - 26ms/step\n",
      "step 150/157 - loss: 66.6106 - acc_top1: 0.2764 - acc_top5: 0.7740 - 25ms/step\n",
      "step 157/157 - loss: 1.8677 - acc_top1: 0.2773 - acc_top5: 0.7748 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 2/50\n",
      "step  10/782 - loss: 1.8949 - acc_top1: 0.2844 - acc_top5: 0.7906 - 160ms/step\n",
      "step  20/782 - loss: 3.5874 - acc_top1: 0.2930 - acc_top5: 0.7977 - 101ms/step\n",
      "step  30/782 - loss: 2.1395 - acc_top1: 0.2979 - acc_top5: 0.8021 - 80ms/step\n",
      "step  40/782 - loss: 2.7155 - acc_top1: 0.2949 - acc_top5: 0.8102 - 70ms/step\n",
      "step  50/782 - loss: 2.4429 - acc_top1: 0.3050 - acc_top5: 0.8213 - 64ms/step\n",
      "step  60/782 - loss: 1.6468 - acc_top1: 0.3164 - acc_top5: 0.8299 - 60ms/step\n",
      "step  70/782 - loss: 1.8969 - acc_top1: 0.3199 - acc_top5: 0.8355 - 57ms/step\n",
      "step  80/782 - loss: 2.6343 - acc_top1: 0.3254 - acc_top5: 0.8367 - 55ms/step\n",
      "step  90/782 - loss: 1.7662 - acc_top1: 0.3306 - acc_top5: 0.8384 - 53ms/step\n",
      "step 100/782 - loss: 2.1043 - acc_top1: 0.3345 - acc_top5: 0.8397 - 52ms/step\n",
      "step 110/782 - loss: 1.9785 - acc_top1: 0.3379 - acc_top5: 0.8423 - 50ms/step\n",
      "step 120/782 - loss: 1.7114 - acc_top1: 0.3432 - acc_top5: 0.8426 - 50ms/step\n",
      "step 130/782 - loss: 3.2831 - acc_top1: 0.3431 - acc_top5: 0.8444 - 49ms/step\n",
      "step 140/782 - loss: 2.5907 - acc_top1: 0.3450 - acc_top5: 0.8463 - 48ms/step\n",
      "step 150/782 - loss: 2.9741 - acc_top1: 0.3452 - acc_top5: 0.8467 - 48ms/step\n",
      "step 160/782 - loss: 2.6131 - acc_top1: 0.3479 - acc_top5: 0.8480 - 47ms/step\n",
      "step 170/782 - loss: 2.0140 - acc_top1: 0.3498 - acc_top5: 0.8492 - 47ms/step\n",
      "step 180/782 - loss: 1.9720 - acc_top1: 0.3513 - acc_top5: 0.8502 - 47ms/step\n",
      "step 190/782 - loss: 1.7133 - acc_top1: 0.3530 - acc_top5: 0.8534 - 48ms/step\n",
      "step 200/782 - loss: 1.7436 - acc_top1: 0.3547 - acc_top5: 0.8541 - 48ms/step\n",
      "step 210/782 - loss: 2.7399 - acc_top1: 0.3561 - acc_top5: 0.8542 - 48ms/step\n",
      "step 220/782 - loss: 1.9819 - acc_top1: 0.3565 - acc_top5: 0.8545 - 47ms/step\n",
      "step 230/782 - loss: 1.5375 - acc_top1: 0.3592 - acc_top5: 0.8553 - 47ms/step\n",
      "step 240/782 - loss: 2.0916 - acc_top1: 0.3594 - acc_top5: 0.8564 - 47ms/step\n",
      "step 250/782 - loss: 2.4469 - acc_top1: 0.3601 - acc_top5: 0.8574 - 46ms/step\n",
      "step 260/782 - loss: 1.9395 - acc_top1: 0.3605 - acc_top5: 0.8582 - 46ms/step\n",
      "step 270/782 - loss: 1.3759 - acc_top1: 0.3609 - acc_top5: 0.8584 - 46ms/step\n",
      "step 280/782 - loss: 1.9491 - acc_top1: 0.3621 - acc_top5: 0.8590 - 45ms/step\n",
      "step 290/782 - loss: 2.0510 - acc_top1: 0.3624 - acc_top5: 0.8599 - 45ms/step\n",
      "step 300/782 - loss: 2.1399 - acc_top1: 0.3633 - acc_top5: 0.8613 - 45ms/step\n",
      "step 310/782 - loss: 2.3643 - acc_top1: 0.3645 - acc_top5: 0.8619 - 45ms/step\n",
      "step 320/782 - loss: 2.7970 - acc_top1: 0.3650 - acc_top5: 0.8625 - 44ms/step\n",
      "step 330/782 - loss: 1.9647 - acc_top1: 0.3651 - acc_top5: 0.8630 - 44ms/step\n",
      "step 340/782 - loss: 3.2558 - acc_top1: 0.3657 - acc_top5: 0.8639 - 44ms/step\n",
      "step 350/782 - loss: 1.5114 - acc_top1: 0.3667 - acc_top5: 0.8641 - 44ms/step\n",
      "step 360/782 - loss: 3.6055 - acc_top1: 0.3680 - acc_top5: 0.8645 - 44ms/step\n",
      "step 370/782 - loss: 3.0067 - acc_top1: 0.3688 - acc_top5: 0.8649 - 44ms/step\n",
      "step 380/782 - loss: 1.4425 - acc_top1: 0.3702 - acc_top5: 0.8661 - 43ms/step\n",
      "step 390/782 - loss: 2.3114 - acc_top1: 0.3701 - acc_top5: 0.8656 - 43ms/step\n",
      "step 400/782 - loss: 2.2175 - acc_top1: 0.3683 - acc_top5: 0.8648 - 43ms/step\n",
      "step 410/782 - loss: 3.9662 - acc_top1: 0.3675 - acc_top5: 0.8641 - 43ms/step\n",
      "step 420/782 - loss: 3.0962 - acc_top1: 0.3669 - acc_top5: 0.8629 - 43ms/step\n",
      "step 430/782 - loss: 2.3497 - acc_top1: 0.3658 - acc_top5: 0.8616 - 43ms/step\n",
      "step 440/782 - loss: 2.1337 - acc_top1: 0.3653 - acc_top5: 0.8615 - 43ms/step\n",
      "step 450/782 - loss: 2.2537 - acc_top1: 0.3649 - acc_top5: 0.8613 - 42ms/step\n",
      "step 460/782 - loss: 1.9796 - acc_top1: 0.3639 - acc_top5: 0.8609 - 42ms/step\n",
      "step 470/782 - loss: 2.4534 - acc_top1: 0.3635 - acc_top5: 0.8607 - 42ms/step\n",
      "step 480/782 - loss: 2.6735 - acc_top1: 0.3634 - acc_top5: 0.8607 - 42ms/step\n",
      "step 490/782 - loss: 1.9463 - acc_top1: 0.3631 - acc_top5: 0.8601 - 42ms/step\n",
      "step 500/782 - loss: 1.7927 - acc_top1: 0.3633 - acc_top5: 0.8605 - 42ms/step\n",
      "step 510/782 - loss: 2.4127 - acc_top1: 0.3631 - acc_top5: 0.8602 - 42ms/step\n",
      "step 520/782 - loss: 2.4246 - acc_top1: 0.3630 - acc_top5: 0.8600 - 42ms/step\n",
      "step 530/782 - loss: 1.8891 - acc_top1: 0.3631 - acc_top5: 0.8600 - 42ms/step\n",
      "step 540/782 - loss: 2.3786 - acc_top1: 0.3629 - acc_top5: 0.8604 - 42ms/step\n",
      "step 550/782 - loss: 1.7067 - acc_top1: 0.3626 - acc_top5: 0.8605 - 41ms/step\n",
      "step 560/782 - loss: 4.4715 - acc_top1: 0.3633 - acc_top5: 0.8607 - 41ms/step\n",
      "step 570/782 - loss: 2.4977 - acc_top1: 0.3634 - acc_top5: 0.8604 - 41ms/step\n",
      "step 580/782 - loss: 2.3781 - acc_top1: 0.3633 - acc_top5: 0.8604 - 41ms/step\n",
      "step 590/782 - loss: 1.8906 - acc_top1: 0.3637 - acc_top5: 0.8608 - 41ms/step\n",
      "step 600/782 - loss: 1.5648 - acc_top1: 0.3645 - acc_top5: 0.8610 - 41ms/step\n",
      "step 610/782 - loss: 3.5165 - acc_top1: 0.3645 - acc_top5: 0.8606 - 41ms/step\n",
      "step 620/782 - loss: 1.9946 - acc_top1: 0.3646 - acc_top5: 0.8606 - 41ms/step\n",
      "step 630/782 - loss: 3.3060 - acc_top1: 0.3651 - acc_top5: 0.8609 - 41ms/step\n",
      "step 640/782 - loss: 1.8457 - acc_top1: 0.3652 - acc_top5: 0.8607 - 41ms/step\n",
      "step 650/782 - loss: 2.1491 - acc_top1: 0.3652 - acc_top5: 0.8606 - 41ms/step\n",
      "step 660/782 - loss: 1.6807 - acc_top1: 0.3652 - acc_top5: 0.8606 - 41ms/step\n",
      "step 670/782 - loss: 2.3218 - acc_top1: 0.3653 - acc_top5: 0.8607 - 41ms/step\n",
      "step 680/782 - loss: 2.7707 - acc_top1: 0.3657 - acc_top5: 0.8609 - 41ms/step\n",
      "step 690/782 - loss: 1.5649 - acc_top1: 0.3659 - acc_top5: 0.8609 - 41ms/step\n",
      "step 700/782 - loss: 2.3149 - acc_top1: 0.3662 - acc_top5: 0.8611 - 41ms/step\n",
      "step 710/782 - loss: 2.4514 - acc_top1: 0.3661 - acc_top5: 0.8609 - 41ms/step\n",
      "step 720/782 - loss: 2.0557 - acc_top1: 0.3663 - acc_top5: 0.8609 - 41ms/step\n",
      "step 730/782 - loss: 2.1739 - acc_top1: 0.3665 - acc_top5: 0.8609 - 41ms/step\n",
      "step 740/782 - loss: 1.7441 - acc_top1: 0.3666 - acc_top5: 0.8608 - 41ms/step\n",
      "step 750/782 - loss: 2.8992 - acc_top1: 0.3666 - acc_top5: 0.8611 - 41ms/step\n",
      "step 760/782 - loss: 1.7430 - acc_top1: 0.3668 - acc_top5: 0.8615 - 41ms/step\n",
      "step 770/782 - loss: 2.2283 - acc_top1: 0.3675 - acc_top5: 0.8618 - 41ms/step\n",
      "step 780/782 - loss: 1.7319 - acc_top1: 0.3682 - acc_top5: 0.8620 - 41ms/step\n",
      "step 782/782 - loss: 1.8238 - acc_top1: 0.3682 - acc_top5: 0.8621 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/1\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.6127 - acc_top1: 0.4078 - acc_top5: 0.8547 - 164ms/step\n",
      "step  20/157 - loss: 1.8749 - acc_top1: 0.3883 - acc_top5: 0.8531 - 91ms/step\n",
      "step  30/157 - loss: 1.7538 - acc_top1: 0.3823 - acc_top5: 0.8516 - 67ms/step\n",
      "step  40/157 - loss: 5.6251 - acc_top1: 0.3703 - acc_top5: 0.8539 - 55ms/step\n",
      "step  50/157 - loss: 1.4798 - acc_top1: 0.3713 - acc_top5: 0.8572 - 47ms/step\n",
      "step  60/157 - loss: 1.7791 - acc_top1: 0.3714 - acc_top5: 0.8583 - 43ms/step\n",
      "step  70/157 - loss: 7.3909 - acc_top1: 0.3690 - acc_top5: 0.8585 - 39ms/step\n",
      "step  80/157 - loss: 1.5064 - acc_top1: 0.3699 - acc_top5: 0.8604 - 36ms/step\n",
      "step  90/157 - loss: 2.6468 - acc_top1: 0.3705 - acc_top5: 0.8620 - 34ms/step\n",
      "step 100/157 - loss: 1.7138 - acc_top1: 0.3697 - acc_top5: 0.8614 - 33ms/step\n",
      "step 110/157 - loss: 3.7801 - acc_top1: 0.3690 - acc_top5: 0.8597 - 31ms/step\n",
      "step 120/157 - loss: 1.6085 - acc_top1: 0.3698 - acc_top5: 0.8603 - 30ms/step\n",
      "step 130/157 - loss: 16.1911 - acc_top1: 0.3689 - acc_top5: 0.8591 - 29ms/step\n",
      "step 140/157 - loss: 12.7229 - acc_top1: 0.3711 - acc_top5: 0.8595 - 28ms/step\n",
      "step 150/157 - loss: 3.3593 - acc_top1: 0.3707 - acc_top5: 0.8600 - 28ms/step\n",
      "step 157/157 - loss: 1.3871 - acc_top1: 0.3696 - acc_top5: 0.8607 - 28ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 3/50\n",
      "step  10/782 - loss: 1.6863 - acc_top1: 0.3750 - acc_top5: 0.8625 - 159ms/step\n",
      "step  20/782 - loss: 1.9662 - acc_top1: 0.3828 - acc_top5: 0.8727 - 101ms/step\n",
      "step  30/782 - loss: 1.4500 - acc_top1: 0.3854 - acc_top5: 0.8833 - 82ms/step\n",
      "step  40/782 - loss: 3.8921 - acc_top1: 0.3922 - acc_top5: 0.8816 - 72ms/step\n",
      "step  50/782 - loss: 2.1994 - acc_top1: 0.3881 - acc_top5: 0.8812 - 65ms/step\n",
      "step  60/782 - loss: 4.0902 - acc_top1: 0.3911 - acc_top5: 0.8823 - 61ms/step\n",
      "step  70/782 - loss: 2.3620 - acc_top1: 0.3940 - acc_top5: 0.8866 - 58ms/step\n",
      "step  80/782 - loss: 1.6887 - acc_top1: 0.3939 - acc_top5: 0.8838 - 56ms/step\n",
      "step  90/782 - loss: 1.7572 - acc_top1: 0.3948 - acc_top5: 0.8835 - 54ms/step\n",
      "step 100/782 - loss: 1.3874 - acc_top1: 0.3977 - acc_top5: 0.8850 - 52ms/step\n",
      "step 110/782 - loss: 2.2334 - acc_top1: 0.4001 - acc_top5: 0.8857 - 51ms/step\n",
      "step 120/782 - loss: 1.4981 - acc_top1: 0.4016 - acc_top5: 0.8866 - 50ms/step\n",
      "step 130/782 - loss: 2.2140 - acc_top1: 0.4013 - acc_top5: 0.8873 - 50ms/step\n",
      "step 140/782 - loss: 2.2720 - acc_top1: 0.4027 - acc_top5: 0.8884 - 49ms/step\n",
      "step 150/782 - loss: 2.0299 - acc_top1: 0.4056 - acc_top5: 0.8883 - 48ms/step\n",
      "step 160/782 - loss: 1.7882 - acc_top1: 0.4065 - acc_top5: 0.8886 - 48ms/step\n",
      "step 170/782 - loss: 1.6590 - acc_top1: 0.4095 - acc_top5: 0.8890 - 47ms/step\n",
      "step 180/782 - loss: 1.3513 - acc_top1: 0.4097 - acc_top5: 0.8895 - 47ms/step\n",
      "step 190/782 - loss: 1.9680 - acc_top1: 0.4119 - acc_top5: 0.8900 - 46ms/step\n",
      "step 200/782 - loss: 1.5451 - acc_top1: 0.4140 - acc_top5: 0.8910 - 46ms/step\n",
      "step 210/782 - loss: 2.0769 - acc_top1: 0.4133 - acc_top5: 0.8906 - 45ms/step\n",
      "step 220/782 - loss: 2.1820 - acc_top1: 0.4136 - acc_top5: 0.8907 - 45ms/step\n",
      "step 230/782 - loss: 1.8171 - acc_top1: 0.4134 - acc_top5: 0.8907 - 45ms/step\n",
      "step 240/782 - loss: 3.6558 - acc_top1: 0.4139 - acc_top5: 0.8904 - 44ms/step\n",
      "step 250/782 - loss: 1.5247 - acc_top1: 0.4140 - acc_top5: 0.8902 - 44ms/step\n",
      "step 260/782 - loss: 2.4595 - acc_top1: 0.4152 - acc_top5: 0.8907 - 44ms/step\n",
      "step 270/782 - loss: 3.1456 - acc_top1: 0.4161 - acc_top5: 0.8914 - 44ms/step\n",
      "step 280/782 - loss: 2.1960 - acc_top1: 0.4166 - acc_top5: 0.8924 - 44ms/step\n",
      "step 290/782 - loss: 1.6427 - acc_top1: 0.4161 - acc_top5: 0.8921 - 43ms/step\n",
      "step 300/782 - loss: 1.9622 - acc_top1: 0.4165 - acc_top5: 0.8923 - 43ms/step\n",
      "step 310/782 - loss: 1.6327 - acc_top1: 0.4167 - acc_top5: 0.8916 - 43ms/step\n",
      "step 320/782 - loss: 2.2983 - acc_top1: 0.4177 - acc_top5: 0.8915 - 43ms/step\n",
      "step 330/782 - loss: 2.2839 - acc_top1: 0.4170 - acc_top5: 0.8908 - 43ms/step\n",
      "step 340/782 - loss: 1.4957 - acc_top1: 0.4174 - acc_top5: 0.8905 - 43ms/step\n",
      "step 350/782 - loss: 1.6746 - acc_top1: 0.4175 - acc_top5: 0.8903 - 42ms/step\n",
      "step 360/782 - loss: 1.6934 - acc_top1: 0.4171 - acc_top5: 0.8899 - 42ms/step\n",
      "step 370/782 - loss: 2.3329 - acc_top1: 0.4165 - acc_top5: 0.8901 - 43ms/step\n",
      "step 380/782 - loss: 3.0001 - acc_top1: 0.4161 - acc_top5: 0.8904 - 44ms/step\n",
      "step 390/782 - loss: 1.6236 - acc_top1: 0.4162 - acc_top5: 0.8902 - 44ms/step\n",
      "step 400/782 - loss: 1.8010 - acc_top1: 0.4159 - acc_top5: 0.8902 - 44ms/step\n",
      "step 410/782 - loss: 1.9874 - acc_top1: 0.4159 - acc_top5: 0.8903 - 45ms/step\n",
      "step 420/782 - loss: 2.6684 - acc_top1: 0.4161 - acc_top5: 0.8904 - 45ms/step\n",
      "step 430/782 - loss: 1.7575 - acc_top1: 0.4159 - acc_top5: 0.8906 - 46ms/step\n",
      "step 440/782 - loss: 3.4279 - acc_top1: 0.4156 - acc_top5: 0.8906 - 46ms/step\n",
      "step 450/782 - loss: 1.9378 - acc_top1: 0.4159 - acc_top5: 0.8906 - 46ms/step\n",
      "step 460/782 - loss: 1.6957 - acc_top1: 0.4156 - acc_top5: 0.8901 - 47ms/step\n",
      "step 470/782 - loss: 2.2602 - acc_top1: 0.4157 - acc_top5: 0.8905 - 47ms/step\n",
      "step 480/782 - loss: 2.3090 - acc_top1: 0.4156 - acc_top5: 0.8907 - 46ms/step\n",
      "step 490/782 - loss: 2.1884 - acc_top1: 0.4159 - acc_top5: 0.8910 - 46ms/step\n",
      "step 500/782 - loss: 2.1119 - acc_top1: 0.4169 - acc_top5: 0.8914 - 46ms/step\n",
      "step 510/782 - loss: 1.9123 - acc_top1: 0.4168 - acc_top5: 0.8911 - 46ms/step\n",
      "step 520/782 - loss: 3.1826 - acc_top1: 0.4176 - acc_top5: 0.8913 - 47ms/step\n",
      "step 530/782 - loss: 2.1080 - acc_top1: 0.4178 - acc_top5: 0.8916 - 47ms/step\n",
      "step 540/782 - loss: 1.5109 - acc_top1: 0.4174 - acc_top5: 0.8914 - 46ms/step\n",
      "step 550/782 - loss: 1.6833 - acc_top1: 0.4176 - acc_top5: 0.8912 - 46ms/step\n",
      "step 560/782 - loss: 2.6896 - acc_top1: 0.4173 - acc_top5: 0.8914 - 47ms/step\n",
      "step 570/782 - loss: 1.7505 - acc_top1: 0.4175 - acc_top5: 0.8915 - 47ms/step\n",
      "step 580/782 - loss: 2.5563 - acc_top1: 0.4178 - acc_top5: 0.8917 - 47ms/step\n",
      "step 590/782 - loss: 1.5734 - acc_top1: 0.4181 - acc_top5: 0.8918 - 47ms/step\n",
      "step 600/782 - loss: 1.5831 - acc_top1: 0.4184 - acc_top5: 0.8920 - 47ms/step\n",
      "step 610/782 - loss: 2.0277 - acc_top1: 0.4184 - acc_top5: 0.8923 - 46ms/step\n",
      "step 620/782 - loss: 3.3243 - acc_top1: 0.4189 - acc_top5: 0.8926 - 46ms/step\n",
      "step 630/782 - loss: 1.6585 - acc_top1: 0.4186 - acc_top5: 0.8926 - 46ms/step\n",
      "step 640/782 - loss: 1.2838 - acc_top1: 0.4182 - acc_top5: 0.8925 - 46ms/step\n",
      "step 650/782 - loss: 2.3694 - acc_top1: 0.4184 - acc_top5: 0.8927 - 46ms/step\n",
      "step 660/782 - loss: 1.4427 - acc_top1: 0.4187 - acc_top5: 0.8928 - 46ms/step\n",
      "step 670/782 - loss: 1.5812 - acc_top1: 0.4193 - acc_top5: 0.8929 - 46ms/step\n",
      "step 680/782 - loss: 1.8492 - acc_top1: 0.4193 - acc_top5: 0.8926 - 46ms/step\n",
      "step 690/782 - loss: 2.2266 - acc_top1: 0.4196 - acc_top5: 0.8926 - 46ms/step\n",
      "step 700/782 - loss: 1.9907 - acc_top1: 0.4200 - acc_top5: 0.8927 - 46ms/step\n",
      "step 710/782 - loss: 2.5638 - acc_top1: 0.4203 - acc_top5: 0.8926 - 46ms/step\n",
      "step 720/782 - loss: 2.3534 - acc_top1: 0.4205 - acc_top5: 0.8927 - 45ms/step\n",
      "step 730/782 - loss: 1.8432 - acc_top1: 0.4203 - acc_top5: 0.8928 - 45ms/step\n",
      "step 740/782 - loss: 2.3052 - acc_top1: 0.4205 - acc_top5: 0.8926 - 45ms/step\n",
      "step 750/782 - loss: 1.8384 - acc_top1: 0.4209 - acc_top5: 0.8928 - 45ms/step\n",
      "step 760/782 - loss: 2.1825 - acc_top1: 0.4209 - acc_top5: 0.8929 - 45ms/step\n",
      "step 770/782 - loss: 1.4571 - acc_top1: 0.4211 - acc_top5: 0.8931 - 45ms/step\n",
      "step 780/782 - loss: 1.5744 - acc_top1: 0.4212 - acc_top5: 0.8932 - 45ms/step\n",
      "step 782/782 - loss: 4.1225 - acc_top1: 0.4211 - acc_top5: 0.8932 - 45ms/step\n",
      "save checkpoint at /home/aistudio/output/2\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.5196 - acc_top1: 0.4266 - acc_top5: 0.9031 - 136ms/step\n",
      "step  20/157 - loss: 1.7724 - acc_top1: 0.4250 - acc_top5: 0.8953 - 78ms/step\n",
      "step  30/157 - loss: 3.9858 - acc_top1: 0.4167 - acc_top5: 0.8948 - 58ms/step\n",
      "step  40/157 - loss: 1.9178 - acc_top1: 0.4090 - acc_top5: 0.8918 - 48ms/step\n",
      "step  50/157 - loss: 1.6665 - acc_top1: 0.4113 - acc_top5: 0.8906 - 42ms/step\n",
      "step  60/157 - loss: 1.5056 - acc_top1: 0.4156 - acc_top5: 0.8883 - 38ms/step\n",
      "step  70/157 - loss: 4.1236 - acc_top1: 0.4152 - acc_top5: 0.8904 - 35ms/step\n",
      "step  80/157 - loss: 1.6260 - acc_top1: 0.4174 - acc_top5: 0.8930 - 33ms/step\n",
      "step  90/157 - loss: 1.4850 - acc_top1: 0.4188 - acc_top5: 0.8927 - 32ms/step\n",
      "step 100/157 - loss: 1.4482 - acc_top1: 0.4156 - acc_top5: 0.8930 - 30ms/step\n",
      "step 110/157 - loss: 1.7651 - acc_top1: 0.4146 - acc_top5: 0.8903 - 29ms/step\n",
      "step 120/157 - loss: 1.5418 - acc_top1: 0.4145 - acc_top5: 0.8909 - 28ms/step\n",
      "step 130/157 - loss: 4.3963 - acc_top1: 0.4144 - acc_top5: 0.8913 - 28ms/step\n",
      "step 140/157 - loss: 4.2611 - acc_top1: 0.4151 - acc_top5: 0.8911 - 27ms/step\n",
      "step 150/157 - loss: 1.5149 - acc_top1: 0.4173 - acc_top5: 0.8908 - 27ms/step\n",
      "step 157/157 - loss: 1.2867 - acc_top1: 0.4168 - acc_top5: 0.8917 - 27ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 4/50\n",
      "step  10/782 - loss: 1.7065 - acc_top1: 0.3875 - acc_top5: 0.8641 - 190ms/step\n",
      "step  20/782 - loss: 2.2727 - acc_top1: 0.3805 - acc_top5: 0.8672 - 115ms/step\n",
      "step  30/782 - loss: 3.2923 - acc_top1: 0.3734 - acc_top5: 0.8719 - 90ms/step\n",
      "step  40/782 - loss: 2.0200 - acc_top1: 0.3785 - acc_top5: 0.8742 - 77ms/step\n",
      "step  50/782 - loss: 1.5345 - acc_top1: 0.3806 - acc_top5: 0.8741 - 69ms/step\n",
      "step  60/782 - loss: 2.3732 - acc_top1: 0.3839 - acc_top5: 0.8734 - 64ms/step\n",
      "step  70/782 - loss: 2.3633 - acc_top1: 0.3868 - acc_top5: 0.8763 - 60ms/step\n",
      "step  80/782 - loss: 1.7717 - acc_top1: 0.3922 - acc_top5: 0.8801 - 57ms/step\n",
      "step  90/782 - loss: 4.2134 - acc_top1: 0.3970 - acc_top5: 0.8830 - 55ms/step\n",
      "step 100/782 - loss: 1.7660 - acc_top1: 0.4008 - acc_top5: 0.8827 - 54ms/step\n",
      "step 110/782 - loss: 2.3379 - acc_top1: 0.4026 - acc_top5: 0.8828 - 52ms/step\n",
      "step 120/782 - loss: 1.7384 - acc_top1: 0.4059 - acc_top5: 0.8842 - 51ms/step\n",
      "step 130/782 - loss: 2.9391 - acc_top1: 0.4058 - acc_top5: 0.8857 - 50ms/step\n",
      "step 140/782 - loss: 1.7160 - acc_top1: 0.4062 - acc_top5: 0.8866 - 50ms/step\n",
      "step 150/782 - loss: 2.6680 - acc_top1: 0.4059 - acc_top5: 0.8851 - 49ms/step\n",
      "step 160/782 - loss: 1.7562 - acc_top1: 0.4084 - acc_top5: 0.8858 - 48ms/step\n",
      "step 170/782 - loss: 2.2190 - acc_top1: 0.4104 - acc_top5: 0.8852 - 47ms/step\n",
      "step 180/782 - loss: 2.0454 - acc_top1: 0.4098 - acc_top5: 0.8855 - 47ms/step\n",
      "step 190/782 - loss: 1.6188 - acc_top1: 0.4095 - acc_top5: 0.8870 - 47ms/step\n",
      "step 200/782 - loss: 1.6261 - acc_top1: 0.4105 - acc_top5: 0.8876 - 46ms/step\n",
      "step 210/782 - loss: 2.0849 - acc_top1: 0.4120 - acc_top5: 0.8881 - 46ms/step\n",
      "step 220/782 - loss: 1.9238 - acc_top1: 0.4126 - acc_top5: 0.8887 - 45ms/step\n",
      "step 230/782 - loss: 1.9266 - acc_top1: 0.4136 - acc_top5: 0.8894 - 45ms/step\n",
      "step 240/782 - loss: 1.9784 - acc_top1: 0.4141 - acc_top5: 0.8906 - 45ms/step\n",
      "step 250/782 - loss: 1.5491 - acc_top1: 0.4151 - acc_top5: 0.8914 - 45ms/step\n",
      "step 260/782 - loss: 2.6405 - acc_top1: 0.4157 - acc_top5: 0.8919 - 44ms/step\n",
      "step 270/782 - loss: 1.8234 - acc_top1: 0.4166 - acc_top5: 0.8918 - 44ms/step\n",
      "step 280/782 - loss: 2.2960 - acc_top1: 0.4170 - acc_top5: 0.8926 - 44ms/step\n",
      "step 290/782 - loss: 1.4043 - acc_top1: 0.4185 - acc_top5: 0.8933 - 44ms/step\n",
      "step 300/782 - loss: 2.4140 - acc_top1: 0.4198 - acc_top5: 0.8936 - 44ms/step\n",
      "step 310/782 - loss: 1.6366 - acc_top1: 0.4203 - acc_top5: 0.8944 - 43ms/step\n",
      "step 320/782 - loss: 1.4110 - acc_top1: 0.4203 - acc_top5: 0.8946 - 43ms/step\n",
      "step 330/782 - loss: 1.6237 - acc_top1: 0.4204 - acc_top5: 0.8951 - 43ms/step\n",
      "step 340/782 - loss: 2.0629 - acc_top1: 0.4217 - acc_top5: 0.8960 - 43ms/step\n",
      "step 350/782 - loss: 1.8358 - acc_top1: 0.4225 - acc_top5: 0.8963 - 43ms/step\n",
      "step 360/782 - loss: 1.4355 - acc_top1: 0.4227 - acc_top5: 0.8967 - 43ms/step\n",
      "step 370/782 - loss: 1.3814 - acc_top1: 0.4238 - acc_top5: 0.8971 - 43ms/step\n",
      "step 380/782 - loss: 1.5729 - acc_top1: 0.4250 - acc_top5: 0.8974 - 43ms/step\n",
      "step 390/782 - loss: 1.6730 - acc_top1: 0.4266 - acc_top5: 0.8976 - 43ms/step\n",
      "step 400/782 - loss: 1.5799 - acc_top1: 0.4275 - acc_top5: 0.8976 - 43ms/step\n",
      "step 410/782 - loss: 3.3652 - acc_top1: 0.4281 - acc_top5: 0.8976 - 43ms/step\n",
      "step 420/782 - loss: 1.4762 - acc_top1: 0.4290 - acc_top5: 0.8975 - 43ms/step\n",
      "step 430/782 - loss: 1.7084 - acc_top1: 0.4295 - acc_top5: 0.8977 - 43ms/step\n",
      "step 440/782 - loss: 1.9792 - acc_top1: 0.4301 - acc_top5: 0.8978 - 43ms/step\n",
      "step 450/782 - loss: 1.5003 - acc_top1: 0.4305 - acc_top5: 0.8981 - 43ms/step\n",
      "step 460/782 - loss: 1.6294 - acc_top1: 0.4308 - acc_top5: 0.8983 - 43ms/step\n",
      "step 470/782 - loss: 1.3023 - acc_top1: 0.4316 - acc_top5: 0.8987 - 43ms/step\n",
      "step 480/782 - loss: 1.5775 - acc_top1: 0.4316 - acc_top5: 0.8987 - 42ms/step\n",
      "step 490/782 - loss: 1.4781 - acc_top1: 0.4327 - acc_top5: 0.8994 - 42ms/step\n",
      "step 500/782 - loss: 2.3100 - acc_top1: 0.4336 - acc_top5: 0.8998 - 42ms/step\n",
      "step 510/782 - loss: 2.2941 - acc_top1: 0.4344 - acc_top5: 0.8998 - 42ms/step\n",
      "step 520/782 - loss: 1.3150 - acc_top1: 0.4350 - acc_top5: 0.9003 - 42ms/step\n",
      "step 530/782 - loss: 2.1015 - acc_top1: 0.4357 - acc_top5: 0.9003 - 42ms/step\n",
      "step 540/782 - loss: 1.5691 - acc_top1: 0.4358 - acc_top5: 0.9001 - 42ms/step\n",
      "step 550/782 - loss: 1.4130 - acc_top1: 0.4362 - acc_top5: 0.9004 - 42ms/step\n",
      "step 560/782 - loss: 1.8948 - acc_top1: 0.4367 - acc_top5: 0.9009 - 42ms/step\n",
      "step 570/782 - loss: 1.6471 - acc_top1: 0.4373 - acc_top5: 0.9010 - 42ms/step\n",
      "step 580/782 - loss: 1.7648 - acc_top1: 0.4380 - acc_top5: 0.9012 - 42ms/step\n",
      "step 590/782 - loss: 1.6638 - acc_top1: 0.4386 - acc_top5: 0.9014 - 42ms/step\n",
      "step 600/782 - loss: 1.5153 - acc_top1: 0.4394 - acc_top5: 0.9017 - 42ms/step\n",
      "step 610/782 - loss: 1.9361 - acc_top1: 0.4406 - acc_top5: 0.9020 - 42ms/step\n",
      "step 620/782 - loss: 3.2506 - acc_top1: 0.4414 - acc_top5: 0.9020 - 42ms/step\n",
      "step 630/782 - loss: 1.5691 - acc_top1: 0.4417 - acc_top5: 0.9024 - 42ms/step\n",
      "step 640/782 - loss: 1.6924 - acc_top1: 0.4423 - acc_top5: 0.9024 - 42ms/step\n",
      "step 650/782 - loss: 1.4755 - acc_top1: 0.4428 - acc_top5: 0.9025 - 42ms/step\n",
      "step 660/782 - loss: 1.5350 - acc_top1: 0.4431 - acc_top5: 0.9028 - 42ms/step\n",
      "step 670/782 - loss: 2.4084 - acc_top1: 0.4440 - acc_top5: 0.9030 - 42ms/step\n",
      "step 680/782 - loss: 2.3561 - acc_top1: 0.4446 - acc_top5: 0.9030 - 42ms/step\n",
      "step 690/782 - loss: 1.7818 - acc_top1: 0.4455 - acc_top5: 0.9033 - 42ms/step\n",
      "step 700/782 - loss: 2.0136 - acc_top1: 0.4457 - acc_top5: 0.9033 - 41ms/step\n",
      "step 710/782 - loss: 1.3842 - acc_top1: 0.4461 - acc_top5: 0.9035 - 41ms/step\n",
      "step 720/782 - loss: 1.4519 - acc_top1: 0.4467 - acc_top5: 0.9036 - 41ms/step\n",
      "step 730/782 - loss: 1.7431 - acc_top1: 0.4469 - acc_top5: 0.9039 - 42ms/step\n",
      "step 740/782 - loss: 1.5323 - acc_top1: 0.4472 - acc_top5: 0.9039 - 42ms/step\n",
      "step 750/782 - loss: 1.7838 - acc_top1: 0.4471 - acc_top5: 0.9039 - 43ms/step\n",
      "step 760/782 - loss: 1.2579 - acc_top1: 0.4472 - acc_top5: 0.9044 - 43ms/step\n",
      "step 770/782 - loss: 1.4568 - acc_top1: 0.4478 - acc_top5: 0.9044 - 44ms/step\n",
      "step 780/782 - loss: 1.6406 - acc_top1: 0.4480 - acc_top5: 0.9045 - 44ms/step\n",
      "step 782/782 - loss: 4.6313 - acc_top1: 0.4479 - acc_top5: 0.9044 - 45ms/step\n",
      "save checkpoint at /home/aistudio/output/3\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.3030 - acc_top1: 0.4609 - acc_top5: 0.9156 - 128ms/step\n",
      "step  20/157 - loss: 1.7328 - acc_top1: 0.4492 - acc_top5: 0.8992 - 73ms/step\n",
      "step  30/157 - loss: 7.2826 - acc_top1: 0.4500 - acc_top5: 0.9031 - 55ms/step\n",
      "step  40/157 - loss: 1.9747 - acc_top1: 0.4488 - acc_top5: 0.8988 - 46ms/step\n",
      "step  50/157 - loss: 1.6572 - acc_top1: 0.4478 - acc_top5: 0.9012 - 40ms/step\n",
      "step  60/157 - loss: 1.3751 - acc_top1: 0.4529 - acc_top5: 0.9010 - 36ms/step\n",
      "step  70/157 - loss: 2.8995 - acc_top1: 0.4554 - acc_top5: 0.9016 - 34ms/step\n",
      "step  80/157 - loss: 1.6473 - acc_top1: 0.4570 - acc_top5: 0.9031 - 32ms/step\n",
      "step  90/157 - loss: 1.4857 - acc_top1: 0.4566 - acc_top5: 0.9024 - 30ms/step\n",
      "step 100/157 - loss: 1.4040 - acc_top1: 0.4556 - acc_top5: 0.9036 - 29ms/step\n",
      "step 110/157 - loss: 1.6884 - acc_top1: 0.4528 - acc_top5: 0.9018 - 28ms/step\n",
      "step 120/157 - loss: 1.4070 - acc_top1: 0.4552 - acc_top5: 0.9031 - 28ms/step\n",
      "step 130/157 - loss: 3.6561 - acc_top1: 0.4534 - acc_top5: 0.9016 - 28ms/step\n",
      "step 140/157 - loss: 23.9845 - acc_top1: 0.4537 - acc_top5: 0.9002 - 28ms/step\n",
      "step 150/157 - loss: 1.6696 - acc_top1: 0.4534 - acc_top5: 0.8997 - 27ms/step\n",
      "step 157/157 - loss: 1.0484 - acc_top1: 0.4536 - acc_top5: 0.9005 - 28ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 5/50\n",
      "step  10/782 - loss: 1.6343 - acc_top1: 0.3953 - acc_top5: 0.8609 - 151ms/step\n",
      "step  20/782 - loss: 1.3848 - acc_top1: 0.4125 - acc_top5: 0.8805 - 96ms/step\n",
      "step  30/782 - loss: 1.5781 - acc_top1: 0.4151 - acc_top5: 0.8880 - 81ms/step\n",
      "step  40/782 - loss: 1.6262 - acc_top1: 0.4336 - acc_top5: 0.8945 - 71ms/step\n",
      "step  50/782 - loss: 2.1402 - acc_top1: 0.4384 - acc_top5: 0.8981 - 65ms/step\n",
      "step  60/782 - loss: 2.1931 - acc_top1: 0.4448 - acc_top5: 0.8971 - 61ms/step\n",
      "step  70/782 - loss: 1.8114 - acc_top1: 0.4442 - acc_top5: 0.8991 - 58ms/step\n",
      "step  80/782 - loss: 1.7296 - acc_top1: 0.4482 - acc_top5: 0.9023 - 56ms/step\n",
      "step  90/782 - loss: 1.8574 - acc_top1: 0.4507 - acc_top5: 0.9049 - 54ms/step\n",
      "step 100/782 - loss: 2.6958 - acc_top1: 0.4542 - acc_top5: 0.9058 - 52ms/step\n",
      "step 110/782 - loss: 1.7184 - acc_top1: 0.4524 - acc_top5: 0.9067 - 51ms/step\n",
      "step 120/782 - loss: 1.5527 - acc_top1: 0.4536 - acc_top5: 0.9078 - 50ms/step\n",
      "step 130/782 - loss: 1.7435 - acc_top1: 0.4569 - acc_top5: 0.9076 - 49ms/step\n",
      "step 140/782 - loss: 1.6211 - acc_top1: 0.4592 - acc_top5: 0.9079 - 48ms/step\n",
      "step 150/782 - loss: 2.4127 - acc_top1: 0.4609 - acc_top5: 0.9068 - 48ms/step\n",
      "step 160/782 - loss: 1.9551 - acc_top1: 0.4638 - acc_top5: 0.9075 - 47ms/step\n",
      "step 170/782 - loss: 1.6329 - acc_top1: 0.4650 - acc_top5: 0.9075 - 47ms/step\n",
      "step 180/782 - loss: 1.1603 - acc_top1: 0.4661 - acc_top5: 0.9076 - 46ms/step\n",
      "step 190/782 - loss: 2.1932 - acc_top1: 0.4669 - acc_top5: 0.9075 - 46ms/step\n",
      "step 200/782 - loss: 1.4265 - acc_top1: 0.4679 - acc_top5: 0.9077 - 46ms/step\n",
      "step 210/782 - loss: 1.3598 - acc_top1: 0.4679 - acc_top5: 0.9089 - 45ms/step\n",
      "step 220/782 - loss: 1.4430 - acc_top1: 0.4682 - acc_top5: 0.9096 - 45ms/step\n",
      "step 230/782 - loss: 1.5513 - acc_top1: 0.4690 - acc_top5: 0.9092 - 45ms/step\n",
      "step 240/782 - loss: 1.5732 - acc_top1: 0.4691 - acc_top5: 0.9089 - 44ms/step\n",
      "step 250/782 - loss: 1.9486 - acc_top1: 0.4699 - acc_top5: 0.9098 - 44ms/step\n",
      "step 260/782 - loss: 2.6064 - acc_top1: 0.4706 - acc_top5: 0.9103 - 44ms/step\n",
      "step 270/782 - loss: 2.4845 - acc_top1: 0.4720 - acc_top5: 0.9108 - 44ms/step\n",
      "step 280/782 - loss: 2.1275 - acc_top1: 0.4718 - acc_top5: 0.9112 - 43ms/step\n",
      "step 290/782 - loss: 1.5725 - acc_top1: 0.4736 - acc_top5: 0.9115 - 43ms/step\n",
      "step 300/782 - loss: 1.6753 - acc_top1: 0.4742 - acc_top5: 0.9119 - 43ms/step\n",
      "step 310/782 - loss: 1.6232 - acc_top1: 0.4742 - acc_top5: 0.9120 - 43ms/step\n",
      "step 320/782 - loss: 1.9386 - acc_top1: 0.4741 - acc_top5: 0.9127 - 43ms/step\n",
      "step 330/782 - loss: 1.3719 - acc_top1: 0.4743 - acc_top5: 0.9131 - 43ms/step\n",
      "step 340/782 - loss: 1.9660 - acc_top1: 0.4743 - acc_top5: 0.9133 - 43ms/step\n",
      "step 350/782 - loss: 1.7863 - acc_top1: 0.4737 - acc_top5: 0.9130 - 42ms/step\n",
      "step 360/782 - loss: 2.2834 - acc_top1: 0.4746 - acc_top5: 0.9135 - 42ms/step\n",
      "step 370/782 - loss: 2.0676 - acc_top1: 0.4742 - acc_top5: 0.9139 - 42ms/step\n",
      "step 380/782 - loss: 2.2424 - acc_top1: 0.4746 - acc_top5: 0.9137 - 42ms/step\n",
      "step 390/782 - loss: 1.4706 - acc_top1: 0.4746 - acc_top5: 0.9138 - 42ms/step\n",
      "step 400/782 - loss: 1.6080 - acc_top1: 0.4757 - acc_top5: 0.9140 - 42ms/step\n",
      "step 410/782 - loss: 2.1939 - acc_top1: 0.4760 - acc_top5: 0.9140 - 42ms/step\n",
      "step 420/782 - loss: 1.3815 - acc_top1: 0.4756 - acc_top5: 0.9142 - 42ms/step\n",
      "step 430/782 - loss: 1.6147 - acc_top1: 0.4764 - acc_top5: 0.9143 - 42ms/step\n",
      "step 440/782 - loss: 1.6883 - acc_top1: 0.4767 - acc_top5: 0.9145 - 42ms/step\n",
      "step 450/782 - loss: 1.5431 - acc_top1: 0.4770 - acc_top5: 0.9143 - 42ms/step\n",
      "step 460/782 - loss: 1.3903 - acc_top1: 0.4769 - acc_top5: 0.9147 - 42ms/step\n",
      "step 470/782 - loss: 1.5438 - acc_top1: 0.4766 - acc_top5: 0.9149 - 42ms/step\n",
      "step 480/782 - loss: 1.4335 - acc_top1: 0.4766 - acc_top5: 0.9152 - 42ms/step\n",
      "step 490/782 - loss: 1.8855 - acc_top1: 0.4767 - acc_top5: 0.9153 - 42ms/step\n",
      "step 500/782 - loss: 1.3609 - acc_top1: 0.4772 - acc_top5: 0.9156 - 41ms/step\n",
      "step 510/782 - loss: 1.3349 - acc_top1: 0.4763 - acc_top5: 0.9156 - 41ms/step\n",
      "step 520/782 - loss: 1.4438 - acc_top1: 0.4768 - acc_top5: 0.9155 - 41ms/step\n",
      "step 530/782 - loss: 1.7054 - acc_top1: 0.4775 - acc_top5: 0.9162 - 41ms/step\n",
      "step 540/782 - loss: 1.9239 - acc_top1: 0.4779 - acc_top5: 0.9164 - 41ms/step\n",
      "step 550/782 - loss: 1.7821 - acc_top1: 0.4783 - acc_top5: 0.9163 - 41ms/step\n",
      "step 560/782 - loss: 2.1376 - acc_top1: 0.4791 - acc_top5: 0.9163 - 41ms/step\n",
      "step 570/782 - loss: 2.4542 - acc_top1: 0.4797 - acc_top5: 0.9164 - 41ms/step\n",
      "step 580/782 - loss: 2.3643 - acc_top1: 0.4797 - acc_top5: 0.9165 - 41ms/step\n",
      "step 590/782 - loss: 1.3123 - acc_top1: 0.4803 - acc_top5: 0.9169 - 41ms/step\n",
      "step 600/782 - loss: 1.1838 - acc_top1: 0.4812 - acc_top5: 0.9172 - 41ms/step\n",
      "step 610/782 - loss: 1.4150 - acc_top1: 0.4807 - acc_top5: 0.9175 - 41ms/step\n",
      "step 620/782 - loss: 1.3044 - acc_top1: 0.4812 - acc_top5: 0.9177 - 41ms/step\n",
      "step 630/782 - loss: 1.7196 - acc_top1: 0.4815 - acc_top5: 0.9178 - 41ms/step\n",
      "step 640/782 - loss: 1.4121 - acc_top1: 0.4813 - acc_top5: 0.9177 - 41ms/step\n",
      "step 650/782 - loss: 1.9533 - acc_top1: 0.4813 - acc_top5: 0.9177 - 41ms/step\n",
      "step 660/782 - loss: 2.7739 - acc_top1: 0.4803 - acc_top5: 0.9170 - 41ms/step\n",
      "step 670/782 - loss: 2.6497 - acc_top1: 0.4790 - acc_top5: 0.9166 - 41ms/step\n",
      "step 680/782 - loss: 3.1522 - acc_top1: 0.4785 - acc_top5: 0.9164 - 41ms/step\n",
      "step 690/782 - loss: 1.4775 - acc_top1: 0.4780 - acc_top5: 0.9162 - 41ms/step\n",
      "step 700/782 - loss: 1.4797 - acc_top1: 0.4777 - acc_top5: 0.9160 - 41ms/step\n",
      "step 710/782 - loss: 2.1766 - acc_top1: 0.4769 - acc_top5: 0.9159 - 41ms/step\n",
      "step 720/782 - loss: 1.8846 - acc_top1: 0.4767 - acc_top5: 0.9157 - 41ms/step\n",
      "step 730/782 - loss: 1.3392 - acc_top1: 0.4764 - acc_top5: 0.9157 - 41ms/step\n",
      "step 740/782 - loss: 1.6661 - acc_top1: 0.4763 - acc_top5: 0.9154 - 41ms/step\n",
      "step 750/782 - loss: 1.6177 - acc_top1: 0.4760 - acc_top5: 0.9152 - 41ms/step\n",
      "step 760/782 - loss: 1.6490 - acc_top1: 0.4758 - acc_top5: 0.9150 - 41ms/step\n",
      "step 770/782 - loss: 2.1274 - acc_top1: 0.4754 - acc_top5: 0.9149 - 41ms/step\n",
      "step 780/782 - loss: 2.1751 - acc_top1: 0.4751 - acc_top5: 0.9146 - 40ms/step\n",
      "step 782/782 - loss: 1.1793 - acc_top1: 0.4751 - acc_top5: 0.9146 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/4\n",
      "Eval begin...\n",
      "step  10/157 - loss: 22.4728 - acc_top1: 0.3953 - acc_top5: 0.8766 - 122ms/step\n",
      "step  20/157 - loss: 24.9223 - acc_top1: 0.4203 - acc_top5: 0.8812 - 70ms/step\n",
      "step  30/157 - loss: 1.7837 - acc_top1: 0.4245 - acc_top5: 0.8755 - 53ms/step\n",
      "step  40/157 - loss: 3.9664 - acc_top1: 0.4188 - acc_top5: 0.8789 - 44ms/step\n",
      "step  50/157 - loss: 4.4694 - acc_top1: 0.4194 - acc_top5: 0.8825 - 39ms/step\n",
      "step  60/157 - loss: 1.9815 - acc_top1: 0.4247 - acc_top5: 0.8844 - 36ms/step\n",
      "step  70/157 - loss: 9.0188 - acc_top1: 0.4254 - acc_top5: 0.8846 - 33ms/step\n",
      "step  80/157 - loss: 7.1688 - acc_top1: 0.4273 - acc_top5: 0.8873 - 31ms/step\n",
      "step  90/157 - loss: 9.8442 - acc_top1: 0.4241 - acc_top5: 0.8847 - 30ms/step\n",
      "step 100/157 - loss: 2.5560 - acc_top1: 0.4238 - acc_top5: 0.8866 - 29ms/step\n",
      "step 110/157 - loss: 4.7043 - acc_top1: 0.4230 - acc_top5: 0.8866 - 28ms/step\n",
      "step 120/157 - loss: 4.1273 - acc_top1: 0.4223 - acc_top5: 0.8870 - 27ms/step\n",
      "step 130/157 - loss: 5.8543 - acc_top1: 0.4226 - acc_top5: 0.8868 - 26ms/step\n",
      "step 140/157 - loss: 2.1189 - acc_top1: 0.4214 - acc_top5: 0.8865 - 26ms/step\n",
      "step 150/157 - loss: 5.9510 - acc_top1: 0.4217 - acc_top5: 0.8865 - 25ms/step\n",
      "step 157/157 - loss: 1.6338 - acc_top1: 0.4208 - acc_top5: 0.8874 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 6/50\n",
      "step  10/782 - loss: 1.7197 - acc_top1: 0.4453 - acc_top5: 0.9047 - 154ms/step\n",
      "step  20/782 - loss: 2.1294 - acc_top1: 0.4313 - acc_top5: 0.9016 - 97ms/step\n",
      "step  30/782 - loss: 1.7277 - acc_top1: 0.4417 - acc_top5: 0.9016 - 78ms/step\n",
      "step  40/782 - loss: 1.4150 - acc_top1: 0.4469 - acc_top5: 0.9008 - 69ms/step\n",
      "step  50/782 - loss: 1.5464 - acc_top1: 0.4425 - acc_top5: 0.9044 - 63ms/step\n",
      "step  60/782 - loss: 1.5442 - acc_top1: 0.4404 - acc_top5: 0.9065 - 59ms/step\n",
      "step  70/782 - loss: 1.7620 - acc_top1: 0.4362 - acc_top5: 0.9067 - 56ms/step\n",
      "step  80/782 - loss: 1.5694 - acc_top1: 0.4328 - acc_top5: 0.9057 - 54ms/step\n",
      "step  90/782 - loss: 1.8767 - acc_top1: 0.4345 - acc_top5: 0.9062 - 52ms/step\n",
      "step 100/782 - loss: 1.3603 - acc_top1: 0.4375 - acc_top5: 0.9087 - 51ms/step\n",
      "step 110/782 - loss: 1.5730 - acc_top1: 0.4391 - acc_top5: 0.9080 - 50ms/step\n",
      "step 120/782 - loss: 1.6545 - acc_top1: 0.4415 - acc_top5: 0.9083 - 49ms/step\n",
      "step 130/782 - loss: 1.4537 - acc_top1: 0.4415 - acc_top5: 0.9067 - 49ms/step\n",
      "step 140/782 - loss: 1.9121 - acc_top1: 0.4420 - acc_top5: 0.9079 - 48ms/step\n",
      "step 150/782 - loss: 1.5347 - acc_top1: 0.4395 - acc_top5: 0.9085 - 47ms/step\n",
      "step 160/782 - loss: 1.5656 - acc_top1: 0.4417 - acc_top5: 0.9087 - 47ms/step\n",
      "step 170/782 - loss: 1.3843 - acc_top1: 0.4403 - acc_top5: 0.9095 - 46ms/step\n",
      "step 180/782 - loss: 1.6288 - acc_top1: 0.4418 - acc_top5: 0.9101 - 46ms/step\n",
      "step 190/782 - loss: 1.7744 - acc_top1: 0.4435 - acc_top5: 0.9107 - 46ms/step\n",
      "step 200/782 - loss: 1.8764 - acc_top1: 0.4429 - acc_top5: 0.9102 - 45ms/step\n",
      "step 210/782 - loss: 1.6054 - acc_top1: 0.4451 - acc_top5: 0.9096 - 45ms/step\n",
      "step 220/782 - loss: 1.2416 - acc_top1: 0.4457 - acc_top5: 0.9104 - 45ms/step\n",
      "step 230/782 - loss: 1.5286 - acc_top1: 0.4461 - acc_top5: 0.9103 - 45ms/step\n",
      "step 240/782 - loss: 1.6602 - acc_top1: 0.4474 - acc_top5: 0.9111 - 45ms/step\n",
      "step 250/782 - loss: 1.3427 - acc_top1: 0.4491 - acc_top5: 0.9127 - 44ms/step\n",
      "step 260/782 - loss: 1.5271 - acc_top1: 0.4499 - acc_top5: 0.9129 - 44ms/step\n",
      "step 270/782 - loss: 2.0131 - acc_top1: 0.4511 - acc_top5: 0.9130 - 44ms/step\n",
      "step 280/782 - loss: 2.0029 - acc_top1: 0.4521 - acc_top5: 0.9134 - 44ms/step\n",
      "step 290/782 - loss: 1.9282 - acc_top1: 0.4529 - acc_top5: 0.9137 - 44ms/step\n",
      "step 300/782 - loss: 1.9983 - acc_top1: 0.4526 - acc_top5: 0.9137 - 44ms/step\n",
      "step 310/782 - loss: 2.2990 - acc_top1: 0.4524 - acc_top5: 0.9136 - 43ms/step\n",
      "step 320/782 - loss: 1.3853 - acc_top1: 0.4532 - acc_top5: 0.9138 - 43ms/step\n",
      "step 330/782 - loss: 1.8764 - acc_top1: 0.4538 - acc_top5: 0.9136 - 43ms/step\n",
      "step 340/782 - loss: 1.7782 - acc_top1: 0.4541 - acc_top5: 0.9136 - 43ms/step\n",
      "step 350/782 - loss: 1.4167 - acc_top1: 0.4561 - acc_top5: 0.9143 - 43ms/step\n",
      "step 360/782 - loss: 2.2742 - acc_top1: 0.4564 - acc_top5: 0.9149 - 43ms/step\n",
      "step 370/782 - loss: 2.0459 - acc_top1: 0.4577 - acc_top5: 0.9151 - 43ms/step\n",
      "step 380/782 - loss: 1.8149 - acc_top1: 0.4575 - acc_top5: 0.9154 - 43ms/step\n",
      "step 390/782 - loss: 1.5711 - acc_top1: 0.4572 - acc_top5: 0.9157 - 43ms/step\n",
      "step 400/782 - loss: 1.4048 - acc_top1: 0.4579 - acc_top5: 0.9159 - 43ms/step\n",
      "step 410/782 - loss: 1.3638 - acc_top1: 0.4586 - acc_top5: 0.9162 - 43ms/step\n",
      "step 420/782 - loss: 1.4182 - acc_top1: 0.4595 - acc_top5: 0.9161 - 43ms/step\n",
      "step 430/782 - loss: 1.3830 - acc_top1: 0.4601 - acc_top5: 0.9164 - 43ms/step\n",
      "step 440/782 - loss: 2.5770 - acc_top1: 0.4608 - acc_top5: 0.9167 - 43ms/step\n",
      "step 450/782 - loss: 2.7583 - acc_top1: 0.4614 - acc_top5: 0.9171 - 42ms/step\n",
      "step 460/782 - loss: 1.4226 - acc_top1: 0.4622 - acc_top5: 0.9173 - 42ms/step\n",
      "step 470/782 - loss: 1.4924 - acc_top1: 0.4623 - acc_top5: 0.9171 - 42ms/step\n",
      "step 480/782 - loss: 1.2479 - acc_top1: 0.4628 - acc_top5: 0.9175 - 42ms/step\n",
      "step 490/782 - loss: 1.5122 - acc_top1: 0.4629 - acc_top5: 0.9176 - 42ms/step\n",
      "step 500/782 - loss: 1.5347 - acc_top1: 0.4631 - acc_top5: 0.9180 - 42ms/step\n",
      "step 510/782 - loss: 1.4755 - acc_top1: 0.4634 - acc_top5: 0.9182 - 42ms/step\n",
      "step 520/782 - loss: 1.6320 - acc_top1: 0.4638 - acc_top5: 0.9186 - 42ms/step\n",
      "step 530/782 - loss: 1.9575 - acc_top1: 0.4636 - acc_top5: 0.9186 - 42ms/step\n",
      "step 540/782 - loss: 1.4223 - acc_top1: 0.4639 - acc_top5: 0.9189 - 42ms/step\n",
      "step 550/782 - loss: 1.3942 - acc_top1: 0.4643 - acc_top5: 0.9188 - 42ms/step\n",
      "step 560/782 - loss: 2.4313 - acc_top1: 0.4639 - acc_top5: 0.9189 - 42ms/step\n",
      "step 570/782 - loss: 1.6342 - acc_top1: 0.4646 - acc_top5: 0.9187 - 42ms/step\n",
      "step 580/782 - loss: 1.8296 - acc_top1: 0.4649 - acc_top5: 0.9183 - 42ms/step\n",
      "step 590/782 - loss: 1.4342 - acc_top1: 0.4654 - acc_top5: 0.9185 - 42ms/step\n",
      "step 600/782 - loss: 1.6762 - acc_top1: 0.4657 - acc_top5: 0.9183 - 42ms/step\n",
      "step 610/782 - loss: 2.2980 - acc_top1: 0.4656 - acc_top5: 0.9184 - 42ms/step\n",
      "step 620/782 - loss: 2.2190 - acc_top1: 0.4653 - acc_top5: 0.9182 - 42ms/step\n",
      "step 630/782 - loss: 1.4344 - acc_top1: 0.4656 - acc_top5: 0.9181 - 42ms/step\n",
      "step 640/782 - loss: 1.9029 - acc_top1: 0.4659 - acc_top5: 0.9181 - 42ms/step\n",
      "step 650/782 - loss: 2.0572 - acc_top1: 0.4662 - acc_top5: 0.9178 - 42ms/step\n",
      "step 660/782 - loss: 1.1554 - acc_top1: 0.4666 - acc_top5: 0.9176 - 42ms/step\n",
      "step 670/782 - loss: 1.5251 - acc_top1: 0.4670 - acc_top5: 0.9175 - 42ms/step\n",
      "step 680/782 - loss: 1.6957 - acc_top1: 0.4676 - acc_top5: 0.9175 - 42ms/step\n",
      "step 690/782 - loss: 1.6853 - acc_top1: 0.4678 - acc_top5: 0.9176 - 42ms/step\n",
      "step 700/782 - loss: 1.6529 - acc_top1: 0.4681 - acc_top5: 0.9173 - 42ms/step\n",
      "step 710/782 - loss: 1.3755 - acc_top1: 0.4680 - acc_top5: 0.9171 - 42ms/step\n",
      "step 720/782 - loss: 1.7451 - acc_top1: 0.4680 - acc_top5: 0.9169 - 42ms/step\n",
      "step 730/782 - loss: 2.0158 - acc_top1: 0.4682 - acc_top5: 0.9169 - 42ms/step\n",
      "step 740/782 - loss: 1.6661 - acc_top1: 0.4685 - acc_top5: 0.9169 - 41ms/step\n",
      "step 750/782 - loss: 1.9330 - acc_top1: 0.4686 - acc_top5: 0.9171 - 41ms/step\n",
      "step 760/782 - loss: 1.4928 - acc_top1: 0.4689 - acc_top5: 0.9171 - 41ms/step\n",
      "step 770/782 - loss: 1.5479 - acc_top1: 0.4692 - acc_top5: 0.9170 - 41ms/step\n",
      "step 780/782 - loss: 1.4770 - acc_top1: 0.4695 - acc_top5: 0.9170 - 41ms/step\n",
      "step 782/782 - loss: 2.1577 - acc_top1: 0.4695 - acc_top5: 0.9171 - 42ms/step\n",
      "save checkpoint at /home/aistudio/output/5\n",
      "Eval begin...\n",
      "step  10/157 - loss: 11.2460 - acc_top1: 0.4938 - acc_top5: 0.9187 - 132ms/step\n",
      "step  20/157 - loss: 1.8923 - acc_top1: 0.4977 - acc_top5: 0.9133 - 75ms/step\n",
      "step  30/157 - loss: 1.4766 - acc_top1: 0.4922 - acc_top5: 0.9115 - 56ms/step\n",
      "step  40/157 - loss: 8.5802 - acc_top1: 0.4871 - acc_top5: 0.9141 - 47ms/step\n",
      "step  50/157 - loss: 1.4100 - acc_top1: 0.4866 - acc_top5: 0.9169 - 41ms/step\n",
      "step  60/157 - loss: 1.1621 - acc_top1: 0.4888 - acc_top5: 0.9182 - 37ms/step\n",
      "step  70/157 - loss: 2.6755 - acc_top1: 0.4924 - acc_top5: 0.9172 - 34ms/step\n",
      "step  80/157 - loss: 18.3098 - acc_top1: 0.4918 - acc_top5: 0.9178 - 32ms/step\n",
      "step  90/157 - loss: 47.9622 - acc_top1: 0.4877 - acc_top5: 0.9153 - 31ms/step\n",
      "step 100/157 - loss: 31.8429 - acc_top1: 0.4864 - acc_top5: 0.9166 - 29ms/step\n",
      "step 110/157 - loss: 1.4466 - acc_top1: 0.4838 - acc_top5: 0.9149 - 28ms/step\n",
      "step 120/157 - loss: 1.3081 - acc_top1: 0.4837 - acc_top5: 0.9160 - 27ms/step\n",
      "step 130/157 - loss: 71.7336 - acc_top1: 0.4863 - acc_top5: 0.9157 - 27ms/step\n",
      "step 140/157 - loss: 8.4487 - acc_top1: 0.4857 - acc_top5: 0.9157 - 26ms/step\n",
      "step 150/157 - loss: 24.7576 - acc_top1: 0.4870 - acc_top5: 0.9161 - 26ms/step\n",
      "step 157/157 - loss: 1.0817 - acc_top1: 0.4866 - acc_top5: 0.9171 - 27ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 7/50\n",
      "step  10/782 - loss: 2.1309 - acc_top1: 0.4813 - acc_top5: 0.9187 - 150ms/step\n",
      "step  20/782 - loss: 2.0282 - acc_top1: 0.4867 - acc_top5: 0.9133 - 94ms/step\n",
      "step  30/782 - loss: 1.0641 - acc_top1: 0.5068 - acc_top5: 0.9224 - 76ms/step\n",
      "step  40/782 - loss: 2.1274 - acc_top1: 0.5008 - acc_top5: 0.9258 - 67ms/step\n",
      "step  50/782 - loss: 1.7402 - acc_top1: 0.4972 - acc_top5: 0.9241 - 61ms/step\n",
      "step  60/782 - loss: 2.2917 - acc_top1: 0.4971 - acc_top5: 0.9237 - 57ms/step\n",
      "step  70/782 - loss: 1.4552 - acc_top1: 0.4980 - acc_top5: 0.9234 - 54ms/step\n",
      "step  80/782 - loss: 1.8091 - acc_top1: 0.5014 - acc_top5: 0.9248 - 52ms/step\n",
      "step  90/782 - loss: 1.5592 - acc_top1: 0.5033 - acc_top5: 0.9271 - 51ms/step\n",
      "step 100/782 - loss: 1.4371 - acc_top1: 0.5045 - acc_top5: 0.9272 - 50ms/step\n",
      "step 110/782 - loss: 1.4590 - acc_top1: 0.5024 - acc_top5: 0.9259 - 49ms/step\n",
      "step 120/782 - loss: 1.4061 - acc_top1: 0.5023 - acc_top5: 0.9259 - 48ms/step\n",
      "step 130/782 - loss: 1.4621 - acc_top1: 0.5029 - acc_top5: 0.9269 - 47ms/step\n",
      "step 140/782 - loss: 1.1503 - acc_top1: 0.5041 - acc_top5: 0.9275 - 47ms/step\n",
      "step 150/782 - loss: 1.2255 - acc_top1: 0.5039 - acc_top5: 0.9281 - 46ms/step\n",
      "step 160/782 - loss: 1.4488 - acc_top1: 0.5055 - acc_top5: 0.9287 - 46ms/step\n",
      "step 170/782 - loss: 0.9837 - acc_top1: 0.5076 - acc_top5: 0.9295 - 46ms/step\n",
      "step 180/782 - loss: 1.3768 - acc_top1: 0.5065 - acc_top5: 0.9293 - 47ms/step\n",
      "step 190/782 - loss: 1.5062 - acc_top1: 0.5070 - acc_top5: 0.9286 - 47ms/step\n",
      "step 200/782 - loss: 1.7061 - acc_top1: 0.5086 - acc_top5: 0.9287 - 47ms/step\n",
      "step 210/782 - loss: 1.2823 - acc_top1: 0.5079 - acc_top5: 0.9289 - 47ms/step\n",
      "step 220/782 - loss: 1.5227 - acc_top1: 0.5080 - acc_top5: 0.9288 - 46ms/step\n",
      "step 230/782 - loss: 2.8662 - acc_top1: 0.5069 - acc_top5: 0.9281 - 46ms/step\n",
      "step 240/782 - loss: 2.2949 - acc_top1: 0.5068 - acc_top5: 0.9287 - 46ms/step\n",
      "step 250/782 - loss: 1.6530 - acc_top1: 0.5059 - acc_top5: 0.9285 - 46ms/step\n",
      "step 260/782 - loss: 1.4106 - acc_top1: 0.5064 - acc_top5: 0.9279 - 45ms/step\n",
      "step 270/782 - loss: 1.7314 - acc_top1: 0.5062 - acc_top5: 0.9283 - 45ms/step\n",
      "step 280/782 - loss: 1.3178 - acc_top1: 0.5073 - acc_top5: 0.9283 - 45ms/step\n",
      "step 290/782 - loss: 1.1426 - acc_top1: 0.5075 - acc_top5: 0.9290 - 45ms/step\n",
      "step 300/782 - loss: 1.6247 - acc_top1: 0.5076 - acc_top5: 0.9285 - 45ms/step\n",
      "step 310/782 - loss: 1.3956 - acc_top1: 0.5073 - acc_top5: 0.9279 - 45ms/step\n",
      "step 320/782 - loss: 1.5650 - acc_top1: 0.5083 - acc_top5: 0.9281 - 44ms/step\n",
      "step 330/782 - loss: 1.6297 - acc_top1: 0.5099 - acc_top5: 0.9282 - 44ms/step\n",
      "step 340/782 - loss: 1.8676 - acc_top1: 0.5096 - acc_top5: 0.9279 - 44ms/step\n",
      "step 350/782 - loss: 1.1566 - acc_top1: 0.5095 - acc_top5: 0.9283 - 44ms/step\n",
      "step 360/782 - loss: 1.4369 - acc_top1: 0.5100 - acc_top5: 0.9284 - 44ms/step\n",
      "step 370/782 - loss: 2.1333 - acc_top1: 0.5106 - acc_top5: 0.9289 - 44ms/step\n",
      "step 380/782 - loss: 1.5638 - acc_top1: 0.5112 - acc_top5: 0.9294 - 44ms/step\n",
      "step 390/782 - loss: 1.4281 - acc_top1: 0.5119 - acc_top5: 0.9302 - 43ms/step\n",
      "step 400/782 - loss: 1.8119 - acc_top1: 0.5128 - acc_top5: 0.9300 - 43ms/step\n",
      "step 410/782 - loss: 1.2772 - acc_top1: 0.5125 - acc_top5: 0.9298 - 43ms/step\n",
      "step 420/782 - loss: 1.4583 - acc_top1: 0.5124 - acc_top5: 0.9298 - 43ms/step\n",
      "step 430/782 - loss: 1.1568 - acc_top1: 0.5122 - acc_top5: 0.9293 - 43ms/step\n",
      "step 440/782 - loss: 1.6377 - acc_top1: 0.5125 - acc_top5: 0.9294 - 43ms/step\n",
      "step 450/782 - loss: 1.6212 - acc_top1: 0.5127 - acc_top5: 0.9292 - 43ms/step\n",
      "step 460/782 - loss: 1.6101 - acc_top1: 0.5126 - acc_top5: 0.9292 - 43ms/step\n",
      "step 470/782 - loss: 1.6494 - acc_top1: 0.5137 - acc_top5: 0.9291 - 43ms/step\n",
      "step 480/782 - loss: 1.3690 - acc_top1: 0.5134 - acc_top5: 0.9290 - 43ms/step\n",
      "step 490/782 - loss: 1.7498 - acc_top1: 0.5136 - acc_top5: 0.9293 - 43ms/step\n",
      "step 500/782 - loss: 1.5784 - acc_top1: 0.5134 - acc_top5: 0.9293 - 43ms/step\n",
      "step 510/782 - loss: 1.6109 - acc_top1: 0.5134 - acc_top5: 0.9293 - 43ms/step\n",
      "step 520/782 - loss: 1.4406 - acc_top1: 0.5134 - acc_top5: 0.9291 - 43ms/step\n",
      "step 530/782 - loss: 1.5622 - acc_top1: 0.5139 - acc_top5: 0.9291 - 43ms/step\n",
      "step 540/782 - loss: 1.4435 - acc_top1: 0.5142 - acc_top5: 0.9291 - 43ms/step\n",
      "step 550/782 - loss: 1.4202 - acc_top1: 0.5141 - acc_top5: 0.9292 - 43ms/step\n",
      "step 560/782 - loss: 1.9130 - acc_top1: 0.5140 - acc_top5: 0.9289 - 43ms/step\n",
      "step 570/782 - loss: 1.3057 - acc_top1: 0.5137 - acc_top5: 0.9290 - 43ms/step\n",
      "step 580/782 - loss: 2.0341 - acc_top1: 0.5141 - acc_top5: 0.9294 - 43ms/step\n",
      "step 590/782 - loss: 1.1549 - acc_top1: 0.5141 - acc_top5: 0.9293 - 43ms/step\n",
      "step 600/782 - loss: 2.4240 - acc_top1: 0.5142 - acc_top5: 0.9294 - 42ms/step\n",
      "step 610/782 - loss: 1.3256 - acc_top1: 0.5140 - acc_top5: 0.9293 - 42ms/step\n",
      "step 620/782 - loss: 2.3174 - acc_top1: 0.5149 - acc_top5: 0.9294 - 42ms/step\n",
      "step 630/782 - loss: 1.4164 - acc_top1: 0.5147 - acc_top5: 0.9295 - 42ms/step\n",
      "step 640/782 - loss: 1.6927 - acc_top1: 0.5148 - acc_top5: 0.9295 - 42ms/step\n",
      "step 650/782 - loss: 1.8135 - acc_top1: 0.5147 - acc_top5: 0.9298 - 42ms/step\n",
      "step 660/782 - loss: 1.4263 - acc_top1: 0.5147 - acc_top5: 0.9299 - 42ms/step\n",
      "step 670/782 - loss: 1.5731 - acc_top1: 0.5148 - acc_top5: 0.9300 - 42ms/step\n",
      "step 680/782 - loss: 2.2043 - acc_top1: 0.5150 - acc_top5: 0.9299 - 42ms/step\n",
      "step 690/782 - loss: 1.6166 - acc_top1: 0.5152 - acc_top5: 0.9300 - 42ms/step\n",
      "step 700/782 - loss: 1.7698 - acc_top1: 0.5152 - acc_top5: 0.9299 - 42ms/step\n",
      "step 710/782 - loss: 1.3477 - acc_top1: 0.5148 - acc_top5: 0.9298 - 42ms/step\n",
      "step 720/782 - loss: 2.0653 - acc_top1: 0.5147 - acc_top5: 0.9298 - 42ms/step\n",
      "step 730/782 - loss: 1.9852 - acc_top1: 0.5143 - acc_top5: 0.9296 - 42ms/step\n",
      "step 740/782 - loss: 1.4231 - acc_top1: 0.5138 - acc_top5: 0.9295 - 42ms/step\n",
      "step 750/782 - loss: 1.4650 - acc_top1: 0.5136 - acc_top5: 0.9293 - 42ms/step\n",
      "step 760/782 - loss: 1.4782 - acc_top1: 0.5139 - acc_top5: 0.9290 - 42ms/step\n",
      "step 770/782 - loss: 1.9408 - acc_top1: 0.5140 - acc_top5: 0.9291 - 42ms/step\n",
      "step 780/782 - loss: 1.6033 - acc_top1: 0.5138 - acc_top5: 0.9290 - 42ms/step\n",
      "step 782/782 - loss: 1.8183 - acc_top1: 0.5138 - acc_top5: 0.9291 - 42ms/step\n",
      "save checkpoint at /home/aistudio/output/6\n",
      "Eval begin...\n",
      "step  10/157 - loss: 244.8906 - acc_top1: 0.3781 - acc_top5: 0.8141 - 135ms/step\n",
      "step  20/157 - loss: 247.0872 - acc_top1: 0.3781 - acc_top5: 0.8352 - 78ms/step\n",
      "step  30/157 - loss: 117.0045 - acc_top1: 0.3807 - acc_top5: 0.8328 - 58ms/step\n",
      "step  40/157 - loss: 137.8670 - acc_top1: 0.3855 - acc_top5: 0.8422 - 49ms/step\n",
      "step  50/157 - loss: 83.7872 - acc_top1: 0.3906 - acc_top5: 0.8459 - 43ms/step\n",
      "step  60/157 - loss: 265.5562 - acc_top1: 0.3857 - acc_top5: 0.8438 - 39ms/step\n",
      "step  70/157 - loss: 254.8255 - acc_top1: 0.3846 - acc_top5: 0.8442 - 36ms/step\n",
      "step  80/157 - loss: 168.4786 - acc_top1: 0.3865 - acc_top5: 0.8473 - 34ms/step\n",
      "step  90/157 - loss: 563.9721 - acc_top1: 0.3825 - acc_top5: 0.8486 - 32ms/step\n",
      "step 100/157 - loss: 40.5036 - acc_top1: 0.3831 - acc_top5: 0.8486 - 31ms/step\n",
      "step 110/157 - loss: 269.7094 - acc_top1: 0.3803 - acc_top5: 0.8456 - 30ms/step\n",
      "step 120/157 - loss: 190.7809 - acc_top1: 0.3815 - acc_top5: 0.8474 - 29ms/step\n",
      "step 130/157 - loss: 154.8593 - acc_top1: 0.3804 - acc_top5: 0.8487 - 28ms/step\n",
      "step 140/157 - loss: 239.1382 - acc_top1: 0.3811 - acc_top5: 0.8496 - 27ms/step\n",
      "step 150/157 - loss: 233.4641 - acc_top1: 0.3821 - acc_top5: 0.8484 - 26ms/step\n",
      "step 157/157 - loss: 1.9996 - acc_top1: 0.3824 - acc_top5: 0.8481 - 27ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 8/50\n",
      "step  10/782 - loss: 1.4596 - acc_top1: 0.4797 - acc_top5: 0.9500 - 156ms/step\n",
      "step  20/782 - loss: 2.1171 - acc_top1: 0.4461 - acc_top5: 0.9195 - 97ms/step\n",
      "step  30/782 - loss: 1.4548 - acc_top1: 0.4599 - acc_top5: 0.9208 - 78ms/step\n",
      "step  40/782 - loss: 1.8997 - acc_top1: 0.4691 - acc_top5: 0.9277 - 68ms/step\n",
      "step  50/782 - loss: 1.2807 - acc_top1: 0.4844 - acc_top5: 0.9263 - 63ms/step\n",
      "step  60/782 - loss: 1.1284 - acc_top1: 0.4862 - acc_top5: 0.9247 - 59ms/step\n",
      "step  70/782 - loss: 1.8544 - acc_top1: 0.4871 - acc_top5: 0.9241 - 56ms/step\n",
      "step  80/782 - loss: 1.7892 - acc_top1: 0.4893 - acc_top5: 0.9244 - 54ms/step\n",
      "step  90/782 - loss: 1.1807 - acc_top1: 0.4939 - acc_top5: 0.9260 - 52ms/step\n",
      "step 100/782 - loss: 1.5949 - acc_top1: 0.4939 - acc_top5: 0.9248 - 51ms/step\n",
      "step 110/782 - loss: 1.3144 - acc_top1: 0.4957 - acc_top5: 0.9247 - 50ms/step\n",
      "step 120/782 - loss: 1.3730 - acc_top1: 0.4979 - acc_top5: 0.9242 - 49ms/step\n",
      "step 130/782 - loss: 1.5411 - acc_top1: 0.5012 - acc_top5: 0.9226 - 49ms/step\n",
      "step 140/782 - loss: 1.5014 - acc_top1: 0.5058 - acc_top5: 0.9232 - 48ms/step\n",
      "step 150/782 - loss: 1.6597 - acc_top1: 0.5040 - acc_top5: 0.9228 - 48ms/step\n",
      "step 160/782 - loss: 1.4009 - acc_top1: 0.5041 - acc_top5: 0.9237 - 47ms/step\n",
      "step 170/782 - loss: 1.5092 - acc_top1: 0.5048 - acc_top5: 0.9250 - 47ms/step\n",
      "step 180/782 - loss: 1.4015 - acc_top1: 0.5049 - acc_top5: 0.9254 - 46ms/step\n",
      "step 190/782 - loss: 1.4709 - acc_top1: 0.5062 - acc_top5: 0.9259 - 46ms/step\n",
      "step 200/782 - loss: 2.0703 - acc_top1: 0.5070 - acc_top5: 0.9256 - 45ms/step\n",
      "step 210/782 - loss: 1.0927 - acc_top1: 0.5085 - acc_top5: 0.9262 - 45ms/step\n",
      "step 220/782 - loss: 1.4808 - acc_top1: 0.5097 - acc_top5: 0.9263 - 45ms/step\n",
      "step 230/782 - loss: 1.3712 - acc_top1: 0.5084 - acc_top5: 0.9260 - 45ms/step\n",
      "step 240/782 - loss: 1.3438 - acc_top1: 0.5101 - acc_top5: 0.9262 - 44ms/step\n",
      "step 250/782 - loss: 1.4777 - acc_top1: 0.5114 - acc_top5: 0.9266 - 44ms/step\n",
      "step 260/782 - loss: 1.7569 - acc_top1: 0.5129 - acc_top5: 0.9266 - 44ms/step\n",
      "step 270/782 - loss: 2.1326 - acc_top1: 0.5144 - acc_top5: 0.9269 - 44ms/step\n",
      "step 280/782 - loss: 1.2647 - acc_top1: 0.5145 - acc_top5: 0.9273 - 44ms/step\n",
      "step 290/782 - loss: 2.0832 - acc_top1: 0.5152 - acc_top5: 0.9278 - 44ms/step\n",
      "step 300/782 - loss: 1.6017 - acc_top1: 0.5157 - acc_top5: 0.9276 - 44ms/step\n",
      "step 310/782 - loss: 1.5433 - acc_top1: 0.5164 - acc_top5: 0.9280 - 44ms/step\n",
      "step 320/782 - loss: 1.6388 - acc_top1: 0.5173 - acc_top5: 0.9282 - 44ms/step\n",
      "step 330/782 - loss: 1.7125 - acc_top1: 0.5176 - acc_top5: 0.9280 - 43ms/step\n",
      "step 340/782 - loss: 1.5170 - acc_top1: 0.5182 - acc_top5: 0.9279 - 43ms/step\n",
      "step 350/782 - loss: 1.6054 - acc_top1: 0.5188 - acc_top5: 0.9283 - 43ms/step\n",
      "step 360/782 - loss: 1.2609 - acc_top1: 0.5194 - acc_top5: 0.9289 - 43ms/step\n",
      "step 370/782 - loss: 1.6257 - acc_top1: 0.5202 - acc_top5: 0.9285 - 43ms/step\n",
      "step 380/782 - loss: 1.4070 - acc_top1: 0.5204 - acc_top5: 0.9288 - 43ms/step\n",
      "step 390/782 - loss: 1.2241 - acc_top1: 0.5213 - acc_top5: 0.9291 - 43ms/step\n",
      "step 400/782 - loss: 2.1175 - acc_top1: 0.5209 - acc_top5: 0.9294 - 43ms/step\n",
      "step 410/782 - loss: 1.4736 - acc_top1: 0.5218 - acc_top5: 0.9296 - 43ms/step\n",
      "step 420/782 - loss: 1.3371 - acc_top1: 0.5218 - acc_top5: 0.9294 - 43ms/step\n",
      "step 430/782 - loss: 1.2972 - acc_top1: 0.5227 - acc_top5: 0.9299 - 43ms/step\n",
      "step 440/782 - loss: 1.1584 - acc_top1: 0.5239 - acc_top5: 0.9298 - 42ms/step\n",
      "step 450/782 - loss: 1.4711 - acc_top1: 0.5240 - acc_top5: 0.9298 - 42ms/step\n",
      "step 460/782 - loss: 1.3581 - acc_top1: 0.5244 - acc_top5: 0.9295 - 42ms/step\n",
      "step 470/782 - loss: 1.4889 - acc_top1: 0.5242 - acc_top5: 0.9294 - 42ms/step\n",
      "step 480/782 - loss: 1.2263 - acc_top1: 0.5254 - acc_top5: 0.9295 - 43ms/step\n",
      "step 490/782 - loss: 1.3883 - acc_top1: 0.5253 - acc_top5: 0.9293 - 43ms/step\n",
      "step 500/782 - loss: 1.0697 - acc_top1: 0.5260 - acc_top5: 0.9297 - 43ms/step\n",
      "step 510/782 - loss: 1.0742 - acc_top1: 0.5267 - acc_top5: 0.9300 - 42ms/step\n",
      "step 520/782 - loss: 1.4588 - acc_top1: 0.5264 - acc_top5: 0.9299 - 42ms/step\n",
      "step 530/782 - loss: 1.7773 - acc_top1: 0.5270 - acc_top5: 0.9300 - 42ms/step\n",
      "step 540/782 - loss: 1.6163 - acc_top1: 0.5275 - acc_top5: 0.9304 - 42ms/step\n",
      "step 550/782 - loss: 1.5049 - acc_top1: 0.5277 - acc_top5: 0.9306 - 42ms/step\n",
      "step 560/782 - loss: 1.4330 - acc_top1: 0.5280 - acc_top5: 0.9306 - 42ms/step\n",
      "step 570/782 - loss: 1.7545 - acc_top1: 0.5277 - acc_top5: 0.9304 - 42ms/step\n",
      "step 580/782 - loss: 1.5615 - acc_top1: 0.5276 - acc_top5: 0.9303 - 42ms/step\n",
      "step 590/782 - loss: 1.4978 - acc_top1: 0.5280 - acc_top5: 0.9306 - 42ms/step\n",
      "step 600/782 - loss: 1.5926 - acc_top1: 0.5286 - acc_top5: 0.9305 - 42ms/step\n",
      "step 610/782 - loss: 1.6263 - acc_top1: 0.5284 - acc_top5: 0.9302 - 42ms/step\n",
      "step 620/782 - loss: 1.4339 - acc_top1: 0.5291 - acc_top5: 0.9306 - 42ms/step\n",
      "step 630/782 - loss: 1.2861 - acc_top1: 0.5292 - acc_top5: 0.9306 - 42ms/step\n",
      "step 640/782 - loss: 1.5040 - acc_top1: 0.5292 - acc_top5: 0.9306 - 42ms/step\n",
      "step 650/782 - loss: 1.5748 - acc_top1: 0.5293 - acc_top5: 0.9306 - 42ms/step\n",
      "step 660/782 - loss: 1.7800 - acc_top1: 0.5295 - acc_top5: 0.9306 - 42ms/step\n",
      "step 670/782 - loss: 1.7326 - acc_top1: 0.5295 - acc_top5: 0.9307 - 42ms/step\n",
      "step 680/782 - loss: 1.2286 - acc_top1: 0.5299 - acc_top5: 0.9308 - 42ms/step\n",
      "step 690/782 - loss: 1.6953 - acc_top1: 0.5301 - acc_top5: 0.9310 - 42ms/step\n",
      "step 700/782 - loss: 1.6383 - acc_top1: 0.5300 - acc_top5: 0.9310 - 42ms/step\n",
      "step 710/782 - loss: 1.9644 - acc_top1: 0.5302 - acc_top5: 0.9312 - 42ms/step\n",
      "step 720/782 - loss: 2.0324 - acc_top1: 0.5301 - acc_top5: 0.9311 - 42ms/step\n",
      "step 730/782 - loss: 1.4443 - acc_top1: 0.5304 - acc_top5: 0.9312 - 41ms/step\n",
      "step 740/782 - loss: 1.7467 - acc_top1: 0.5306 - acc_top5: 0.9313 - 41ms/step\n",
      "step 750/782 - loss: 1.5173 - acc_top1: 0.5311 - acc_top5: 0.9313 - 41ms/step\n",
      "step 760/782 - loss: 2.0579 - acc_top1: 0.5312 - acc_top5: 0.9316 - 41ms/step\n",
      "step 770/782 - loss: 1.3216 - acc_top1: 0.5313 - acc_top5: 0.9317 - 41ms/step\n",
      "step 780/782 - loss: 1.7534 - acc_top1: 0.5314 - acc_top5: 0.9316 - 41ms/step\n",
      "step 782/782 - loss: 2.9132 - acc_top1: 0.5313 - acc_top5: 0.9315 - 42ms/step\n",
      "save checkpoint at /home/aistudio/output/7\n",
      "Eval begin...\n",
      "step  10/157 - loss: 2.4087 - acc_top1: 0.5391 - acc_top5: 0.9547 - 132ms/step\n",
      "step  20/157 - loss: 7.2413 - acc_top1: 0.5477 - acc_top5: 0.9500 - 75ms/step\n",
      "step  30/157 - loss: 1.4251 - acc_top1: 0.5464 - acc_top5: 0.9411 - 56ms/step\n",
      "step  40/157 - loss: 1.7841 - acc_top1: 0.5469 - acc_top5: 0.9391 - 47ms/step\n",
      "step  50/157 - loss: 1.3387 - acc_top1: 0.5453 - acc_top5: 0.9363 - 41ms/step\n",
      "step  60/157 - loss: 3.0469 - acc_top1: 0.5492 - acc_top5: 0.9326 - 37ms/step\n",
      "step  70/157 - loss: 2.5180 - acc_top1: 0.5542 - acc_top5: 0.9324 - 35ms/step\n",
      "step  80/157 - loss: 29.3621 - acc_top1: 0.5557 - acc_top5: 0.9342 - 33ms/step\n",
      "step  90/157 - loss: 3.4722 - acc_top1: 0.5510 - acc_top5: 0.9337 - 31ms/step\n",
      "step 100/157 - loss: 12.9938 - acc_top1: 0.5494 - acc_top5: 0.9333 - 30ms/step\n",
      "step 110/157 - loss: 7.5719 - acc_top1: 0.5477 - acc_top5: 0.9332 - 29ms/step\n",
      "step 120/157 - loss: 2.8622 - acc_top1: 0.5480 - acc_top5: 0.9345 - 28ms/step\n",
      "step 130/157 - loss: 11.2019 - acc_top1: 0.5481 - acc_top5: 0.9337 - 27ms/step\n",
      "step 140/157 - loss: 2.7725 - acc_top1: 0.5491 - acc_top5: 0.9342 - 27ms/step\n",
      "step 150/157 - loss: 1.1255 - acc_top1: 0.5492 - acc_top5: 0.9345 - 26ms/step\n",
      "step 157/157 - loss: 0.9433 - acc_top1: 0.5484 - acc_top5: 0.9347 - 28ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 9/50\n",
      "step  10/782 - loss: 1.6143 - acc_top1: 0.5281 - acc_top5: 0.9172 - 154ms/step\n",
      "step  20/782 - loss: 1.2484 - acc_top1: 0.5484 - acc_top5: 0.9242 - 97ms/step\n",
      "step  30/782 - loss: 1.4722 - acc_top1: 0.5526 - acc_top5: 0.9271 - 78ms/step\n",
      "step  40/782 - loss: 0.9181 - acc_top1: 0.5605 - acc_top5: 0.9320 - 69ms/step\n",
      "step  50/782 - loss: 1.2559 - acc_top1: 0.5594 - acc_top5: 0.9309 - 63ms/step\n",
      "step  60/782 - loss: 1.5124 - acc_top1: 0.5536 - acc_top5: 0.9326 - 59ms/step\n",
      "step  70/782 - loss: 1.2989 - acc_top1: 0.5607 - acc_top5: 0.9348 - 56ms/step\n",
      "step  80/782 - loss: 1.1279 - acc_top1: 0.5613 - acc_top5: 0.9375 - 54ms/step\n",
      "step  90/782 - loss: 1.4480 - acc_top1: 0.5634 - acc_top5: 0.9373 - 52ms/step\n",
      "step 100/782 - loss: 1.3546 - acc_top1: 0.5616 - acc_top5: 0.9375 - 51ms/step\n",
      "step 110/782 - loss: 1.3579 - acc_top1: 0.5629 - acc_top5: 0.9376 - 50ms/step\n",
      "step 120/782 - loss: 1.5904 - acc_top1: 0.5637 - acc_top5: 0.9372 - 49ms/step\n",
      "step 130/782 - loss: 1.4806 - acc_top1: 0.5638 - acc_top5: 0.9385 - 48ms/step\n",
      "step 140/782 - loss: 1.2769 - acc_top1: 0.5629 - acc_top5: 0.9381 - 47ms/step\n",
      "step 150/782 - loss: 1.4171 - acc_top1: 0.5621 - acc_top5: 0.9374 - 47ms/step\n",
      "step 160/782 - loss: 1.7158 - acc_top1: 0.5601 - acc_top5: 0.9372 - 46ms/step\n",
      "step 170/782 - loss: 1.4663 - acc_top1: 0.5592 - acc_top5: 0.9373 - 46ms/step\n",
      "step 180/782 - loss: 1.6034 - acc_top1: 0.5601 - acc_top5: 0.9377 - 46ms/step\n",
      "step 190/782 - loss: 0.9341 - acc_top1: 0.5610 - acc_top5: 0.9382 - 45ms/step\n",
      "step 200/782 - loss: 1.2574 - acc_top1: 0.5617 - acc_top5: 0.9384 - 45ms/step\n",
      "step 210/782 - loss: 1.6411 - acc_top1: 0.5604 - acc_top5: 0.9390 - 45ms/step\n",
      "step 220/782 - loss: 1.8258 - acc_top1: 0.5612 - acc_top5: 0.9391 - 44ms/step\n",
      "step 230/782 - loss: 1.4800 - acc_top1: 0.5631 - acc_top5: 0.9392 - 44ms/step\n",
      "step 240/782 - loss: 1.3943 - acc_top1: 0.5632 - acc_top5: 0.9388 - 44ms/step\n",
      "step 250/782 - loss: 1.4238 - acc_top1: 0.5637 - acc_top5: 0.9393 - 44ms/step\n",
      "step 260/782 - loss: 1.4086 - acc_top1: 0.5650 - acc_top5: 0.9392 - 44ms/step\n",
      "step 270/782 - loss: 1.4871 - acc_top1: 0.5655 - acc_top5: 0.9394 - 44ms/step\n",
      "step 280/782 - loss: 1.2401 - acc_top1: 0.5644 - acc_top5: 0.9396 - 43ms/step\n",
      "step 290/782 - loss: 1.8796 - acc_top1: 0.5630 - acc_top5: 0.9399 - 43ms/step\n",
      "step 300/782 - loss: 1.0179 - acc_top1: 0.5626 - acc_top5: 0.9402 - 43ms/step\n",
      "step 310/782 - loss: 1.7212 - acc_top1: 0.5637 - acc_top5: 0.9400 - 43ms/step\n",
      "step 320/782 - loss: 1.0974 - acc_top1: 0.5638 - acc_top5: 0.9400 - 43ms/step\n",
      "step 330/782 - loss: 1.2230 - acc_top1: 0.5623 - acc_top5: 0.9402 - 43ms/step\n",
      "step 340/782 - loss: 1.6375 - acc_top1: 0.5614 - acc_top5: 0.9409 - 43ms/step\n",
      "step 350/782 - loss: 1.3932 - acc_top1: 0.5616 - acc_top5: 0.9415 - 43ms/step\n",
      "step 360/782 - loss: 1.3678 - acc_top1: 0.5619 - acc_top5: 0.9411 - 43ms/step\n",
      "step 370/782 - loss: 1.4699 - acc_top1: 0.5618 - acc_top5: 0.9410 - 43ms/step\n",
      "step 380/782 - loss: 1.1214 - acc_top1: 0.5613 - acc_top5: 0.9409 - 43ms/step\n",
      "step 390/782 - loss: 1.1848 - acc_top1: 0.5612 - acc_top5: 0.9407 - 42ms/step\n",
      "step 400/782 - loss: 1.1181 - acc_top1: 0.5616 - acc_top5: 0.9411 - 42ms/step\n",
      "step 410/782 - loss: 1.1470 - acc_top1: 0.5609 - acc_top5: 0.9406 - 42ms/step\n",
      "step 420/782 - loss: 1.1428 - acc_top1: 0.5611 - acc_top5: 0.9409 - 42ms/step\n",
      "step 430/782 - loss: 1.4464 - acc_top1: 0.5617 - acc_top5: 0.9411 - 42ms/step\n",
      "step 440/782 - loss: 1.7910 - acc_top1: 0.5619 - acc_top5: 0.9412 - 42ms/step\n",
      "step 450/782 - loss: 1.2316 - acc_top1: 0.5619 - acc_top5: 0.9408 - 42ms/step\n",
      "step 460/782 - loss: 1.1981 - acc_top1: 0.5619 - acc_top5: 0.9407 - 42ms/step\n",
      "step 470/782 - loss: 1.4008 - acc_top1: 0.5616 - acc_top5: 0.9406 - 42ms/step\n",
      "step 480/782 - loss: 1.2725 - acc_top1: 0.5617 - acc_top5: 0.9407 - 42ms/step\n",
      "step 490/782 - loss: 1.4771 - acc_top1: 0.5614 - acc_top5: 0.9404 - 42ms/step\n",
      "step 500/782 - loss: 1.3317 - acc_top1: 0.5619 - acc_top5: 0.9404 - 42ms/step\n",
      "step 510/782 - loss: 1.4122 - acc_top1: 0.5625 - acc_top5: 0.9404 - 42ms/step\n",
      "step 520/782 - loss: 1.2210 - acc_top1: 0.5620 - acc_top5: 0.9408 - 42ms/step\n",
      "step 530/782 - loss: 1.2694 - acc_top1: 0.5614 - acc_top5: 0.9409 - 42ms/step\n",
      "step 540/782 - loss: 1.3880 - acc_top1: 0.5614 - acc_top5: 0.9407 - 42ms/step\n",
      "step 550/782 - loss: 1.4554 - acc_top1: 0.5614 - acc_top5: 0.9407 - 42ms/step\n",
      "step 560/782 - loss: 1.3032 - acc_top1: 0.5616 - acc_top5: 0.9410 - 42ms/step\n",
      "step 570/782 - loss: 1.1633 - acc_top1: 0.5618 - acc_top5: 0.9411 - 42ms/step\n",
      "step 580/782 - loss: 1.3402 - acc_top1: 0.5619 - acc_top5: 0.9411 - 42ms/step\n",
      "step 590/782 - loss: 1.4021 - acc_top1: 0.5621 - acc_top5: 0.9412 - 42ms/step\n",
      "step 600/782 - loss: 1.1717 - acc_top1: 0.5622 - acc_top5: 0.9412 - 42ms/step\n",
      "step 610/782 - loss: 1.4071 - acc_top1: 0.5623 - acc_top5: 0.9413 - 42ms/step\n",
      "step 620/782 - loss: 1.2310 - acc_top1: 0.5622 - acc_top5: 0.9414 - 42ms/step\n",
      "step 630/782 - loss: 1.0540 - acc_top1: 0.5624 - acc_top5: 0.9415 - 42ms/step\n",
      "step 640/782 - loss: 1.1190 - acc_top1: 0.5622 - acc_top5: 0.9417 - 42ms/step\n",
      "step 650/782 - loss: 1.1911 - acc_top1: 0.5622 - acc_top5: 0.9419 - 42ms/step\n",
      "step 660/782 - loss: 1.1542 - acc_top1: 0.5622 - acc_top5: 0.9419 - 42ms/step\n",
      "step 670/782 - loss: 1.1179 - acc_top1: 0.5627 - acc_top5: 0.9419 - 42ms/step\n",
      "step 680/782 - loss: 1.3901 - acc_top1: 0.5627 - acc_top5: 0.9419 - 42ms/step\n",
      "step 690/782 - loss: 1.0499 - acc_top1: 0.5627 - acc_top5: 0.9420 - 42ms/step\n",
      "step 700/782 - loss: 1.3234 - acc_top1: 0.5627 - acc_top5: 0.9421 - 42ms/step\n",
      "step 710/782 - loss: 1.1742 - acc_top1: 0.5622 - acc_top5: 0.9419 - 41ms/step\n",
      "step 720/782 - loss: 1.5942 - acc_top1: 0.5619 - acc_top5: 0.9421 - 41ms/step\n",
      "step 730/782 - loss: 1.3822 - acc_top1: 0.5619 - acc_top5: 0.9420 - 41ms/step\n",
      "step 740/782 - loss: 1.5173 - acc_top1: 0.5618 - acc_top5: 0.9417 - 41ms/step\n",
      "step 750/782 - loss: 1.4276 - acc_top1: 0.5614 - acc_top5: 0.9416 - 41ms/step\n",
      "step 760/782 - loss: 1.4180 - acc_top1: 0.5614 - acc_top5: 0.9416 - 41ms/step\n",
      "step 770/782 - loss: 1.1662 - acc_top1: 0.5613 - acc_top5: 0.9414 - 41ms/step\n",
      "step 780/782 - loss: 1.4877 - acc_top1: 0.5609 - acc_top5: 0.9414 - 41ms/step\n",
      "step 782/782 - loss: 1.2786 - acc_top1: 0.5608 - acc_top5: 0.9414 - 42ms/step\n",
      "save checkpoint at /home/aistudio/output/8\n",
      "Eval begin...\n",
      "step  10/157 - loss: 2.4861 - acc_top1: 0.5016 - acc_top5: 0.9375 - 123ms/step\n",
      "step  20/157 - loss: 6.4344 - acc_top1: 0.5117 - acc_top5: 0.9359 - 71ms/step\n",
      "step  30/157 - loss: 16.3662 - acc_top1: 0.5073 - acc_top5: 0.9344 - 53ms/step\n",
      "step  40/157 - loss: 13.9803 - acc_top1: 0.5070 - acc_top5: 0.9344 - 44ms/step\n",
      "step  50/157 - loss: 1.3257 - acc_top1: 0.5112 - acc_top5: 0.9350 - 39ms/step\n",
      "step  60/157 - loss: 27.3051 - acc_top1: 0.5182 - acc_top5: 0.9320 - 35ms/step\n",
      "step  70/157 - loss: 32.3655 - acc_top1: 0.5201 - acc_top5: 0.9304 - 33ms/step\n",
      "step  80/157 - loss: 1.3149 - acc_top1: 0.5184 - acc_top5: 0.9314 - 31ms/step\n",
      "step  90/157 - loss: 11.1875 - acc_top1: 0.5184 - acc_top5: 0.9321 - 30ms/step\n",
      "step 100/157 - loss: 1.1248 - acc_top1: 0.5181 - acc_top5: 0.9331 - 28ms/step\n",
      "step 110/157 - loss: 15.6661 - acc_top1: 0.5138 - acc_top5: 0.9318 - 27ms/step\n",
      "step 120/157 - loss: 1.1297 - acc_top1: 0.5159 - acc_top5: 0.9328 - 27ms/step\n",
      "step 130/157 - loss: 2.0959 - acc_top1: 0.5172 - acc_top5: 0.9323 - 26ms/step\n",
      "step 140/157 - loss: 17.8230 - acc_top1: 0.5175 - acc_top5: 0.9316 - 25ms/step\n",
      "step 150/157 - loss: 1.2033 - acc_top1: 0.5180 - acc_top5: 0.9307 - 25ms/step\n",
      "step 157/157 - loss: 1.0238 - acc_top1: 0.5176 - acc_top5: 0.9310 - 25ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 10/50\n",
      "step  10/782 - loss: 1.2780 - acc_top1: 0.5266 - acc_top5: 0.9391 - 171ms/step\n",
      "step  20/782 - loss: 1.3870 - acc_top1: 0.5242 - acc_top5: 0.9352 - 105ms/step\n",
      "step  30/782 - loss: 1.0588 - acc_top1: 0.5365 - acc_top5: 0.9354 - 84ms/step\n",
      "step  40/782 - loss: 1.3267 - acc_top1: 0.5441 - acc_top5: 0.9437 - 73ms/step\n",
      "step  50/782 - loss: 1.4331 - acc_top1: 0.5425 - acc_top5: 0.9387 - 67ms/step\n",
      "step  60/782 - loss: 1.2423 - acc_top1: 0.5437 - acc_top5: 0.9404 - 62ms/step\n",
      "step  70/782 - loss: 1.5493 - acc_top1: 0.5415 - acc_top5: 0.9395 - 59ms/step\n",
      "step  80/782 - loss: 1.4982 - acc_top1: 0.5406 - acc_top5: 0.9387 - 57ms/step\n",
      "step  90/782 - loss: 2.0498 - acc_top1: 0.5417 - acc_top5: 0.9361 - 55ms/step\n",
      "step 100/782 - loss: 1.4174 - acc_top1: 0.5387 - acc_top5: 0.9359 - 53ms/step\n",
      "step 110/782 - loss: 1.6935 - acc_top1: 0.5408 - acc_top5: 0.9372 - 52ms/step\n",
      "step 120/782 - loss: 1.5362 - acc_top1: 0.5413 - acc_top5: 0.9367 - 51ms/step\n",
      "step 130/782 - loss: 1.1801 - acc_top1: 0.5418 - acc_top5: 0.9387 - 50ms/step\n",
      "step 140/782 - loss: 1.2323 - acc_top1: 0.5424 - acc_top5: 0.9394 - 49ms/step\n",
      "step 150/782 - loss: 1.2495 - acc_top1: 0.5406 - acc_top5: 0.9384 - 48ms/step\n",
      "step 160/782 - loss: 1.2304 - acc_top1: 0.5408 - acc_top5: 0.9389 - 48ms/step\n",
      "step 170/782 - loss: 1.3440 - acc_top1: 0.5406 - acc_top5: 0.9397 - 47ms/step\n",
      "step 180/782 - loss: 1.3148 - acc_top1: 0.5412 - acc_top5: 0.9395 - 47ms/step\n",
      "step 190/782 - loss: 1.4397 - acc_top1: 0.5411 - acc_top5: 0.9400 - 46ms/step\n",
      "step 200/782 - loss: 1.2271 - acc_top1: 0.5433 - acc_top5: 0.9404 - 46ms/step\n",
      "step 210/782 - loss: 1.9097 - acc_top1: 0.5434 - acc_top5: 0.9406 - 46ms/step\n",
      "step 220/782 - loss: 1.1908 - acc_top1: 0.5445 - acc_top5: 0.9408 - 45ms/step\n",
      "step 230/782 - loss: 1.4159 - acc_top1: 0.5449 - acc_top5: 0.9412 - 45ms/step\n",
      "step 240/782 - loss: 1.5197 - acc_top1: 0.5451 - acc_top5: 0.9410 - 45ms/step\n",
      "step 250/782 - loss: 1.4063 - acc_top1: 0.5459 - acc_top5: 0.9413 - 45ms/step\n",
      "step 260/782 - loss: 1.2472 - acc_top1: 0.5463 - acc_top5: 0.9412 - 44ms/step\n",
      "step 270/782 - loss: 1.1744 - acc_top1: 0.5462 - acc_top5: 0.9421 - 45ms/step\n",
      "step 280/782 - loss: 1.5540 - acc_top1: 0.5485 - acc_top5: 0.9427 - 44ms/step\n",
      "step 290/782 - loss: 1.3149 - acc_top1: 0.5496 - acc_top5: 0.9430 - 44ms/step\n",
      "step 300/782 - loss: 1.0972 - acc_top1: 0.5496 - acc_top5: 0.9433 - 44ms/step\n",
      "step 310/782 - loss: 1.4804 - acc_top1: 0.5502 - acc_top5: 0.9432 - 44ms/step\n",
      "step 320/782 - loss: 1.3441 - acc_top1: 0.5513 - acc_top5: 0.9433 - 44ms/step\n",
      "step 330/782 - loss: 1.1677 - acc_top1: 0.5518 - acc_top5: 0.9440 - 44ms/step\n",
      "step 340/782 - loss: 1.2004 - acc_top1: 0.5523 - acc_top5: 0.9439 - 43ms/step\n",
      "step 350/782 - loss: 1.1248 - acc_top1: 0.5533 - acc_top5: 0.9445 - 43ms/step\n",
      "step 360/782 - loss: 1.1733 - acc_top1: 0.5535 - acc_top5: 0.9444 - 43ms/step\n",
      "step 370/782 - loss: 1.5401 - acc_top1: 0.5536 - acc_top5: 0.9448 - 43ms/step\n",
      "step 380/782 - loss: 1.7445 - acc_top1: 0.5533 - acc_top5: 0.9447 - 43ms/step\n",
      "step 390/782 - loss: 1.3188 - acc_top1: 0.5539 - acc_top5: 0.9448 - 43ms/step\n",
      "step 400/782 - loss: 1.2765 - acc_top1: 0.5553 - acc_top5: 0.9450 - 43ms/step\n",
      "step 410/782 - loss: 1.1856 - acc_top1: 0.5558 - acc_top5: 0.9452 - 43ms/step\n",
      "step 420/782 - loss: 1.0978 - acc_top1: 0.5558 - acc_top5: 0.9453 - 43ms/step\n",
      "step 430/782 - loss: 1.3121 - acc_top1: 0.5561 - acc_top5: 0.9452 - 43ms/step\n",
      "step 440/782 - loss: 1.3459 - acc_top1: 0.5562 - acc_top5: 0.9454 - 43ms/step\n",
      "step 450/782 - loss: 1.4983 - acc_top1: 0.5559 - acc_top5: 0.9455 - 43ms/step\n",
      "step 460/782 - loss: 1.2488 - acc_top1: 0.5564 - acc_top5: 0.9456 - 43ms/step\n",
      "step 470/782 - loss: 1.4691 - acc_top1: 0.5569 - acc_top5: 0.9461 - 43ms/step\n",
      "step 480/782 - loss: 1.2296 - acc_top1: 0.5572 - acc_top5: 0.9461 - 42ms/step\n",
      "step 490/782 - loss: 1.3274 - acc_top1: 0.5574 - acc_top5: 0.9464 - 42ms/step\n",
      "step 500/782 - loss: 1.5209 - acc_top1: 0.5576 - acc_top5: 0.9462 - 42ms/step\n",
      "step 510/782 - loss: 1.2301 - acc_top1: 0.5575 - acc_top5: 0.9462 - 42ms/step\n",
      "step 520/782 - loss: 1.2600 - acc_top1: 0.5576 - acc_top5: 0.9466 - 42ms/step\n",
      "step 530/782 - loss: 1.1878 - acc_top1: 0.5579 - acc_top5: 0.9467 - 42ms/step\n",
      "step 540/782 - loss: 1.3252 - acc_top1: 0.5580 - acc_top5: 0.9468 - 42ms/step\n",
      "step 550/782 - loss: 1.1362 - acc_top1: 0.5582 - acc_top5: 0.9470 - 42ms/step\n",
      "step 560/782 - loss: 1.2011 - acc_top1: 0.5586 - acc_top5: 0.9471 - 42ms/step\n",
      "step 570/782 - loss: 1.2696 - acc_top1: 0.5585 - acc_top5: 0.9467 - 42ms/step\n",
      "step 580/782 - loss: 1.2233 - acc_top1: 0.5590 - acc_top5: 0.9468 - 42ms/step\n",
      "step 590/782 - loss: 1.2618 - acc_top1: 0.5585 - acc_top5: 0.9465 - 42ms/step\n",
      "step 600/782 - loss: 1.7843 - acc_top1: 0.5584 - acc_top5: 0.9463 - 42ms/step\n",
      "step 610/782 - loss: 1.1581 - acc_top1: 0.5583 - acc_top5: 0.9464 - 42ms/step\n",
      "step 620/782 - loss: 1.4030 - acc_top1: 0.5592 - acc_top5: 0.9466 - 42ms/step\n",
      "step 630/782 - loss: 1.4045 - acc_top1: 0.5598 - acc_top5: 0.9466 - 42ms/step\n",
      "step 640/782 - loss: 1.1082 - acc_top1: 0.5599 - acc_top5: 0.9467 - 42ms/step\n",
      "step 650/782 - loss: 1.3608 - acc_top1: 0.5599 - acc_top5: 0.9466 - 42ms/step\n",
      "step 660/782 - loss: 1.2242 - acc_top1: 0.5600 - acc_top5: 0.9466 - 42ms/step\n",
      "step 670/782 - loss: 1.0947 - acc_top1: 0.5605 - acc_top5: 0.9467 - 42ms/step\n",
      "step 680/782 - loss: 1.5098 - acc_top1: 0.5606 - acc_top5: 0.9469 - 41ms/step\n",
      "step 690/782 - loss: 1.1569 - acc_top1: 0.5603 - acc_top5: 0.9469 - 41ms/step\n",
      "step 700/782 - loss: 1.2996 - acc_top1: 0.5606 - acc_top5: 0.9471 - 41ms/step\n",
      "step 710/782 - loss: 1.6255 - acc_top1: 0.5602 - acc_top5: 0.9471 - 41ms/step\n",
      "step 720/782 - loss: 1.7221 - acc_top1: 0.5603 - acc_top5: 0.9469 - 41ms/step\n",
      "step 730/782 - loss: 1.3025 - acc_top1: 0.5605 - acc_top5: 0.9467 - 41ms/step\n",
      "step 740/782 - loss: 1.3659 - acc_top1: 0.5604 - acc_top5: 0.9465 - 41ms/step\n",
      "step 750/782 - loss: 1.1633 - acc_top1: 0.5609 - acc_top5: 0.9464 - 41ms/step\n",
      "step 760/782 - loss: 1.2135 - acc_top1: 0.5609 - acc_top5: 0.9464 - 41ms/step\n",
      "step 770/782 - loss: 1.3073 - acc_top1: 0.5608 - acc_top5: 0.9464 - 41ms/step\n",
      "step 780/782 - loss: 1.3344 - acc_top1: 0.5611 - acc_top5: 0.9467 - 41ms/step\n",
      "step 782/782 - loss: 2.1918 - acc_top1: 0.5610 - acc_top5: 0.9467 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/9\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.0328 - acc_top1: 0.5641 - acc_top5: 0.9609 - 126ms/step\n",
      "step  20/157 - loss: 1.5169 - acc_top1: 0.5625 - acc_top5: 0.9578 - 73ms/step\n",
      "step  30/157 - loss: 1.2710 - acc_top1: 0.5687 - acc_top5: 0.9578 - 55ms/step\n",
      "step  40/157 - loss: 3.0019 - acc_top1: 0.5687 - acc_top5: 0.9531 - 46ms/step\n",
      "step  50/157 - loss: 1.2440 - acc_top1: 0.5663 - acc_top5: 0.9525 - 41ms/step\n",
      "step  60/157 - loss: 1.9804 - acc_top1: 0.5643 - acc_top5: 0.9508 - 37ms/step\n",
      "step  70/157 - loss: 1.7075 - acc_top1: 0.5667 - acc_top5: 0.9502 - 34ms/step\n",
      "step  80/157 - loss: 1.2565 - acc_top1: 0.5697 - acc_top5: 0.9521 - 32ms/step\n",
      "step  90/157 - loss: 1.0157 - acc_top1: 0.5674 - acc_top5: 0.9524 - 31ms/step\n",
      "step 100/157 - loss: 1.1704 - acc_top1: 0.5644 - acc_top5: 0.9522 - 29ms/step\n",
      "step 110/157 - loss: 1.3311 - acc_top1: 0.5642 - acc_top5: 0.9513 - 28ms/step\n",
      "step 120/157 - loss: 1.0134 - acc_top1: 0.5647 - acc_top5: 0.9525 - 27ms/step\n",
      "step 130/157 - loss: 5.1689 - acc_top1: 0.5653 - acc_top5: 0.9523 - 27ms/step\n",
      "step 140/157 - loss: 7.3813 - acc_top1: 0.5667 - acc_top5: 0.9522 - 26ms/step\n",
      "step 150/157 - loss: 1.0645 - acc_top1: 0.5654 - acc_top5: 0.9513 - 25ms/step\n",
      "step 157/157 - loss: 0.6634 - acc_top1: 0.5645 - acc_top5: 0.9516 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 11/50\n",
      "step  10/782 - loss: 1.3554 - acc_top1: 0.5859 - acc_top5: 0.9656 - 145ms/step\n",
      "step  20/782 - loss: 1.5046 - acc_top1: 0.5953 - acc_top5: 0.9570 - 92ms/step\n",
      "step  30/782 - loss: 1.1472 - acc_top1: 0.6052 - acc_top5: 0.9542 - 74ms/step\n",
      "step  40/782 - loss: 1.1452 - acc_top1: 0.5980 - acc_top5: 0.9531 - 65ms/step\n",
      "step  50/782 - loss: 1.2123 - acc_top1: 0.5969 - acc_top5: 0.9528 - 60ms/step\n",
      "step  60/782 - loss: 1.1644 - acc_top1: 0.5961 - acc_top5: 0.9526 - 56ms/step\n",
      "step  70/782 - loss: 1.3223 - acc_top1: 0.5946 - acc_top5: 0.9520 - 53ms/step\n",
      "step  80/782 - loss: 1.2686 - acc_top1: 0.6008 - acc_top5: 0.9525 - 51ms/step\n",
      "step  90/782 - loss: 0.9321 - acc_top1: 0.6059 - acc_top5: 0.9540 - 50ms/step\n",
      "step 100/782 - loss: 1.6066 - acc_top1: 0.6059 - acc_top5: 0.9537 - 49ms/step\n",
      "step 110/782 - loss: 1.1665 - acc_top1: 0.6054 - acc_top5: 0.9540 - 48ms/step\n",
      "step 120/782 - loss: 1.3212 - acc_top1: 0.6021 - acc_top5: 0.9553 - 47ms/step\n",
      "step 130/782 - loss: 1.2349 - acc_top1: 0.6016 - acc_top5: 0.9553 - 46ms/step\n",
      "step 140/782 - loss: 1.0641 - acc_top1: 0.6018 - acc_top5: 0.9556 - 46ms/step\n",
      "step 150/782 - loss: 1.2087 - acc_top1: 0.6023 - acc_top5: 0.9558 - 45ms/step\n",
      "step 160/782 - loss: 1.6519 - acc_top1: 0.6013 - acc_top5: 0.9551 - 45ms/step\n",
      "step 170/782 - loss: 1.0563 - acc_top1: 0.6011 - acc_top5: 0.9555 - 44ms/step\n",
      "step 180/782 - loss: 1.1918 - acc_top1: 0.6000 - acc_top5: 0.9556 - 44ms/step\n",
      "step 190/782 - loss: 1.0325 - acc_top1: 0.6000 - acc_top5: 0.9552 - 44ms/step\n",
      "step 200/782 - loss: 1.4776 - acc_top1: 0.5991 - acc_top5: 0.9557 - 43ms/step\n",
      "step 210/782 - loss: 1.2101 - acc_top1: 0.5995 - acc_top5: 0.9559 - 43ms/step\n",
      "step 220/782 - loss: 1.4735 - acc_top1: 0.6001 - acc_top5: 0.9560 - 43ms/step\n",
      "step 230/782 - loss: 0.9849 - acc_top1: 0.5990 - acc_top5: 0.9562 - 43ms/step\n",
      "step 240/782 - loss: 1.0070 - acc_top1: 0.5995 - acc_top5: 0.9564 - 43ms/step\n",
      "step 250/782 - loss: 1.3617 - acc_top1: 0.5989 - acc_top5: 0.9563 - 43ms/step\n",
      "step 260/782 - loss: 0.9639 - acc_top1: 0.5993 - acc_top5: 0.9570 - 43ms/step\n",
      "step 270/782 - loss: 1.2019 - acc_top1: 0.5998 - acc_top5: 0.9571 - 43ms/step\n",
      "step 280/782 - loss: 1.5477 - acc_top1: 0.5993 - acc_top5: 0.9570 - 42ms/step\n",
      "step 290/782 - loss: 0.9952 - acc_top1: 0.5985 - acc_top5: 0.9570 - 42ms/step\n",
      "step 300/782 - loss: 1.4050 - acc_top1: 0.5976 - acc_top5: 0.9568 - 42ms/step\n",
      "step 310/782 - loss: 0.9771 - acc_top1: 0.5974 - acc_top5: 0.9569 - 42ms/step\n",
      "step 320/782 - loss: 1.2142 - acc_top1: 0.5972 - acc_top5: 0.9566 - 42ms/step\n",
      "step 330/782 - loss: 1.1491 - acc_top1: 0.5972 - acc_top5: 0.9565 - 42ms/step\n",
      "step 340/782 - loss: 1.0545 - acc_top1: 0.5970 - acc_top5: 0.9566 - 42ms/step\n",
      "step 350/782 - loss: 1.2905 - acc_top1: 0.5968 - acc_top5: 0.9565 - 42ms/step\n",
      "step 360/782 - loss: 1.2788 - acc_top1: 0.5961 - acc_top5: 0.9558 - 42ms/step\n",
      "step 370/782 - loss: 0.9741 - acc_top1: 0.5966 - acc_top5: 0.9558 - 42ms/step\n",
      "step 380/782 - loss: 1.2962 - acc_top1: 0.5963 - acc_top5: 0.9556 - 41ms/step\n",
      "step 390/782 - loss: 1.2211 - acc_top1: 0.5965 - acc_top5: 0.9556 - 41ms/step\n",
      "step 400/782 - loss: 1.2797 - acc_top1: 0.5968 - acc_top5: 0.9559 - 41ms/step\n",
      "step 410/782 - loss: 1.0380 - acc_top1: 0.5968 - acc_top5: 0.9561 - 41ms/step\n",
      "step 420/782 - loss: 1.2412 - acc_top1: 0.5968 - acc_top5: 0.9561 - 41ms/step\n",
      "step 430/782 - loss: 1.2921 - acc_top1: 0.5971 - acc_top5: 0.9561 - 41ms/step\n",
      "step 440/782 - loss: 1.7121 - acc_top1: 0.5971 - acc_top5: 0.9559 - 41ms/step\n",
      "step 450/782 - loss: 1.1979 - acc_top1: 0.5974 - acc_top5: 0.9560 - 41ms/step\n",
      "step 460/782 - loss: 1.3229 - acc_top1: 0.5970 - acc_top5: 0.9559 - 41ms/step\n",
      "step 470/782 - loss: 1.1139 - acc_top1: 0.5974 - acc_top5: 0.9560 - 41ms/step\n",
      "step 480/782 - loss: 1.3964 - acc_top1: 0.5973 - acc_top5: 0.9561 - 41ms/step\n",
      "step 490/782 - loss: 1.4289 - acc_top1: 0.5967 - acc_top5: 0.9561 - 41ms/step\n",
      "step 500/782 - loss: 1.3472 - acc_top1: 0.5969 - acc_top5: 0.9560 - 41ms/step\n",
      "step 510/782 - loss: 1.2151 - acc_top1: 0.5975 - acc_top5: 0.9561 - 41ms/step\n",
      "step 520/782 - loss: 1.1461 - acc_top1: 0.5970 - acc_top5: 0.9562 - 41ms/step\n",
      "step 530/782 - loss: 1.2920 - acc_top1: 0.5970 - acc_top5: 0.9561 - 41ms/step\n",
      "step 540/782 - loss: 1.3423 - acc_top1: 0.5967 - acc_top5: 0.9558 - 41ms/step\n",
      "step 550/782 - loss: 1.3314 - acc_top1: 0.5969 - acc_top5: 0.9556 - 41ms/step\n",
      "step 560/782 - loss: 1.0046 - acc_top1: 0.5963 - acc_top5: 0.9559 - 41ms/step\n",
      "step 570/782 - loss: 1.6169 - acc_top1: 0.5964 - acc_top5: 0.9560 - 41ms/step\n",
      "step 580/782 - loss: 1.1848 - acc_top1: 0.5967 - acc_top5: 0.9561 - 41ms/step\n",
      "step 590/782 - loss: 0.9374 - acc_top1: 0.5967 - acc_top5: 0.9559 - 41ms/step\n",
      "step 600/782 - loss: 1.0064 - acc_top1: 0.5968 - acc_top5: 0.9559 - 41ms/step\n",
      "step 610/782 - loss: 1.4166 - acc_top1: 0.5965 - acc_top5: 0.9558 - 41ms/step\n",
      "step 620/782 - loss: 1.1455 - acc_top1: 0.5970 - acc_top5: 0.9559 - 41ms/step\n",
      "step 630/782 - loss: 1.2574 - acc_top1: 0.5972 - acc_top5: 0.9560 - 41ms/step\n",
      "step 640/782 - loss: 1.0596 - acc_top1: 0.5979 - acc_top5: 0.9562 - 41ms/step\n",
      "step 650/782 - loss: 1.1544 - acc_top1: 0.5976 - acc_top5: 0.9562 - 41ms/step\n",
      "step 660/782 - loss: 1.4449 - acc_top1: 0.5976 - acc_top5: 0.9564 - 41ms/step\n",
      "step 670/782 - loss: 1.3472 - acc_top1: 0.5976 - acc_top5: 0.9566 - 41ms/step\n",
      "step 680/782 - loss: 1.6942 - acc_top1: 0.5985 - acc_top5: 0.9567 - 41ms/step\n",
      "step 690/782 - loss: 1.3524 - acc_top1: 0.5983 - acc_top5: 0.9566 - 41ms/step\n",
      "step 700/782 - loss: 1.2324 - acc_top1: 0.5984 - acc_top5: 0.9568 - 41ms/step\n",
      "step 710/782 - loss: 0.9779 - acc_top1: 0.5984 - acc_top5: 0.9569 - 41ms/step\n",
      "step 720/782 - loss: 1.0953 - acc_top1: 0.5983 - acc_top5: 0.9569 - 41ms/step\n",
      "step 730/782 - loss: 1.0081 - acc_top1: 0.5987 - acc_top5: 0.9569 - 41ms/step\n",
      "step 740/782 - loss: 1.0328 - acc_top1: 0.5989 - acc_top5: 0.9570 - 41ms/step\n",
      "step 750/782 - loss: 0.9361 - acc_top1: 0.5988 - acc_top5: 0.9570 - 41ms/step\n",
      "step 760/782 - loss: 1.0092 - acc_top1: 0.5992 - acc_top5: 0.9572 - 41ms/step\n",
      "step 770/782 - loss: 1.1743 - acc_top1: 0.5996 - acc_top5: 0.9574 - 41ms/step\n",
      "step 780/782 - loss: 1.1425 - acc_top1: 0.6002 - acc_top5: 0.9575 - 40ms/step\n",
      "step 782/782 - loss: 1.8983 - acc_top1: 0.6002 - acc_top5: 0.9574 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/10\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.1564 - acc_top1: 0.5797 - acc_top5: 0.9641 - 120ms/step\n",
      "step  20/157 - loss: 1.4692 - acc_top1: 0.5836 - acc_top5: 0.9609 - 69ms/step\n",
      "step  30/157 - loss: 1.2198 - acc_top1: 0.5833 - acc_top5: 0.9609 - 52ms/step\n",
      "step  40/157 - loss: 2.5503 - acc_top1: 0.5828 - acc_top5: 0.9582 - 44ms/step\n",
      "step  50/157 - loss: 1.1924 - acc_top1: 0.5831 - acc_top5: 0.9566 - 39ms/step\n",
      "step  60/157 - loss: 1.5015 - acc_top1: 0.5823 - acc_top5: 0.9531 - 36ms/step\n",
      "step  70/157 - loss: 1.1335 - acc_top1: 0.5810 - acc_top5: 0.9525 - 33ms/step\n",
      "step  80/157 - loss: 1.2650 - acc_top1: 0.5805 - acc_top5: 0.9537 - 31ms/step\n",
      "step  90/157 - loss: 0.9462 - acc_top1: 0.5797 - acc_top5: 0.9530 - 30ms/step\n",
      "step 100/157 - loss: 1.1607 - acc_top1: 0.5791 - acc_top5: 0.9541 - 28ms/step\n",
      "step 110/157 - loss: 1.3106 - acc_top1: 0.5793 - acc_top5: 0.9548 - 27ms/step\n",
      "step 120/157 - loss: 1.0099 - acc_top1: 0.5790 - acc_top5: 0.9549 - 26ms/step\n",
      "step 130/157 - loss: 5.9876 - acc_top1: 0.5784 - acc_top5: 0.9538 - 26ms/step\n",
      "step 140/157 - loss: 6.3279 - acc_top1: 0.5769 - acc_top5: 0.9529 - 25ms/step\n",
      "step 150/157 - loss: 0.9634 - acc_top1: 0.5768 - acc_top5: 0.9523 - 25ms/step\n",
      "step 157/157 - loss: 0.8490 - acc_top1: 0.5755 - acc_top5: 0.9523 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 12/50\n",
      "step  10/782 - loss: 1.1611 - acc_top1: 0.6047 - acc_top5: 0.9656 - 149ms/step\n",
      "step  20/782 - loss: 1.1032 - acc_top1: 0.6320 - acc_top5: 0.9609 - 96ms/step\n",
      "step  30/782 - loss: 0.7794 - acc_top1: 0.6339 - acc_top5: 0.9630 - 79ms/step\n",
      "step  40/782 - loss: 1.0549 - acc_top1: 0.6320 - acc_top5: 0.9645 - 69ms/step\n",
      "step  50/782 - loss: 1.0524 - acc_top1: 0.6369 - acc_top5: 0.9650 - 63ms/step\n",
      "step  60/782 - loss: 1.0664 - acc_top1: 0.6312 - acc_top5: 0.9643 - 59ms/step\n",
      "step  70/782 - loss: 0.9635 - acc_top1: 0.6326 - acc_top5: 0.9638 - 57ms/step\n",
      "step  80/782 - loss: 1.1504 - acc_top1: 0.6307 - acc_top5: 0.9641 - 55ms/step\n",
      "step  90/782 - loss: 1.4651 - acc_top1: 0.6299 - acc_top5: 0.9637 - 53ms/step\n",
      "step 100/782 - loss: 1.1547 - acc_top1: 0.6297 - acc_top5: 0.9634 - 52ms/step\n",
      "step 110/782 - loss: 0.9097 - acc_top1: 0.6301 - acc_top5: 0.9635 - 51ms/step\n",
      "step 120/782 - loss: 1.1923 - acc_top1: 0.6305 - acc_top5: 0.9626 - 50ms/step\n",
      "step 130/782 - loss: 1.1815 - acc_top1: 0.6302 - acc_top5: 0.9626 - 49ms/step\n",
      "step 140/782 - loss: 1.2376 - acc_top1: 0.6299 - acc_top5: 0.9619 - 48ms/step\n",
      "step 150/782 - loss: 1.1787 - acc_top1: 0.6310 - acc_top5: 0.9619 - 48ms/step\n",
      "step 160/782 - loss: 1.2147 - acc_top1: 0.6306 - acc_top5: 0.9615 - 47ms/step\n",
      "step 170/782 - loss: 1.1522 - acc_top1: 0.6285 - acc_top5: 0.9606 - 47ms/step\n",
      "step 180/782 - loss: 1.1566 - acc_top1: 0.6287 - acc_top5: 0.9618 - 46ms/step\n",
      "step 190/782 - loss: 0.8988 - acc_top1: 0.6281 - acc_top5: 0.9624 - 46ms/step\n",
      "step 200/782 - loss: 1.1791 - acc_top1: 0.6271 - acc_top5: 0.9623 - 46ms/step\n",
      "step 210/782 - loss: 1.0395 - acc_top1: 0.6269 - acc_top5: 0.9622 - 46ms/step\n",
      "step 220/782 - loss: 0.9760 - acc_top1: 0.6270 - acc_top5: 0.9619 - 45ms/step\n",
      "step 230/782 - loss: 1.5789 - acc_top1: 0.6276 - acc_top5: 0.9619 - 45ms/step\n",
      "step 240/782 - loss: 1.0707 - acc_top1: 0.6268 - acc_top5: 0.9619 - 45ms/step\n",
      "step 250/782 - loss: 1.0322 - acc_top1: 0.6267 - acc_top5: 0.9614 - 44ms/step\n",
      "step 260/782 - loss: 1.2012 - acc_top1: 0.6267 - acc_top5: 0.9611 - 44ms/step\n",
      "step 270/782 - loss: 1.1573 - acc_top1: 0.6261 - acc_top5: 0.9617 - 44ms/step\n",
      "step 280/782 - loss: 1.2577 - acc_top1: 0.6275 - acc_top5: 0.9621 - 44ms/step\n",
      "step 290/782 - loss: 1.0126 - acc_top1: 0.6274 - acc_top5: 0.9623 - 43ms/step\n",
      "step 300/782 - loss: 1.1245 - acc_top1: 0.6276 - acc_top5: 0.9622 - 43ms/step\n",
      "step 310/782 - loss: 1.2967 - acc_top1: 0.6269 - acc_top5: 0.9619 - 43ms/step\n",
      "step 320/782 - loss: 1.0779 - acc_top1: 0.6260 - acc_top5: 0.9622 - 43ms/step\n",
      "step 330/782 - loss: 1.2336 - acc_top1: 0.6243 - acc_top5: 0.9623 - 43ms/step\n",
      "step 340/782 - loss: 1.5701 - acc_top1: 0.6239 - acc_top5: 0.9618 - 43ms/step\n",
      "step 350/782 - loss: 1.0630 - acc_top1: 0.6245 - acc_top5: 0.9619 - 43ms/step\n",
      "step 360/782 - loss: 1.2413 - acc_top1: 0.6247 - acc_top5: 0.9618 - 43ms/step\n",
      "step 370/782 - loss: 1.2165 - acc_top1: 0.6241 - acc_top5: 0.9618 - 44ms/step\n",
      "step 380/782 - loss: 1.2008 - acc_top1: 0.6235 - acc_top5: 0.9617 - 44ms/step\n",
      "step 390/782 - loss: 1.0317 - acc_top1: 0.6225 - acc_top5: 0.9615 - 44ms/step\n",
      "step 400/782 - loss: 1.7108 - acc_top1: 0.6218 - acc_top5: 0.9608 - 44ms/step\n",
      "step 410/782 - loss: 1.6976 - acc_top1: 0.6208 - acc_top5: 0.9607 - 44ms/step\n",
      "step 420/782 - loss: 1.3651 - acc_top1: 0.6204 - acc_top5: 0.9605 - 44ms/step\n",
      "step 430/782 - loss: 1.4093 - acc_top1: 0.6199 - acc_top5: 0.9601 - 44ms/step\n",
      "step 440/782 - loss: 0.9544 - acc_top1: 0.6201 - acc_top5: 0.9598 - 44ms/step\n",
      "step 450/782 - loss: 1.0696 - acc_top1: 0.6197 - acc_top5: 0.9597 - 44ms/step\n",
      "step 460/782 - loss: 1.1735 - acc_top1: 0.6196 - acc_top5: 0.9597 - 44ms/step\n",
      "step 470/782 - loss: 1.3343 - acc_top1: 0.6191 - acc_top5: 0.9595 - 44ms/step\n",
      "step 480/782 - loss: 1.3020 - acc_top1: 0.6198 - acc_top5: 0.9597 - 44ms/step\n",
      "step 490/782 - loss: 1.3776 - acc_top1: 0.6188 - acc_top5: 0.9597 - 44ms/step\n",
      "step 500/782 - loss: 1.0453 - acc_top1: 0.6189 - acc_top5: 0.9598 - 44ms/step\n",
      "step 510/782 - loss: 1.4859 - acc_top1: 0.6186 - acc_top5: 0.9597 - 44ms/step\n",
      "step 520/782 - loss: 1.5363 - acc_top1: 0.6195 - acc_top5: 0.9598 - 43ms/step\n",
      "step 530/782 - loss: 1.0134 - acc_top1: 0.6195 - acc_top5: 0.9599 - 43ms/step\n",
      "step 540/782 - loss: 1.0617 - acc_top1: 0.6190 - acc_top5: 0.9599 - 43ms/step\n",
      "step 550/782 - loss: 1.1109 - acc_top1: 0.6188 - acc_top5: 0.9597 - 43ms/step\n",
      "step 560/782 - loss: 1.1216 - acc_top1: 0.6189 - acc_top5: 0.9597 - 43ms/step\n",
      "step 570/782 - loss: 1.0417 - acc_top1: 0.6186 - acc_top5: 0.9597 - 43ms/step\n",
      "step 580/782 - loss: 1.0973 - acc_top1: 0.6185 - acc_top5: 0.9595 - 43ms/step\n",
      "step 590/782 - loss: 1.2628 - acc_top1: 0.6189 - acc_top5: 0.9595 - 43ms/step\n",
      "step 600/782 - loss: 1.0433 - acc_top1: 0.6194 - acc_top5: 0.9594 - 43ms/step\n",
      "step 610/782 - loss: 1.0646 - acc_top1: 0.6195 - acc_top5: 0.9594 - 43ms/step\n",
      "step 620/782 - loss: 1.2417 - acc_top1: 0.6200 - acc_top5: 0.9596 - 43ms/step\n",
      "step 630/782 - loss: 1.6168 - acc_top1: 0.6200 - acc_top5: 0.9596 - 43ms/step\n",
      "step 640/782 - loss: 1.3735 - acc_top1: 0.6203 - acc_top5: 0.9596 - 43ms/step\n",
      "step 650/782 - loss: 0.9192 - acc_top1: 0.6203 - acc_top5: 0.9596 - 43ms/step\n",
      "step 660/782 - loss: 1.0154 - acc_top1: 0.6201 - acc_top5: 0.9596 - 43ms/step\n",
      "step 670/782 - loss: 0.9309 - acc_top1: 0.6199 - acc_top5: 0.9596 - 43ms/step\n",
      "step 680/782 - loss: 1.0738 - acc_top1: 0.6204 - acc_top5: 0.9597 - 43ms/step\n",
      "step 690/782 - loss: 0.9423 - acc_top1: 0.6203 - acc_top5: 0.9596 - 43ms/step\n",
      "step 700/782 - loss: 0.8576 - acc_top1: 0.6204 - acc_top5: 0.9596 - 43ms/step\n",
      "step 710/782 - loss: 0.8988 - acc_top1: 0.6206 - acc_top5: 0.9596 - 43ms/step\n",
      "step 720/782 - loss: 1.2464 - acc_top1: 0.6206 - acc_top5: 0.9595 - 43ms/step\n",
      "step 730/782 - loss: 1.0599 - acc_top1: 0.6208 - acc_top5: 0.9594 - 43ms/step\n",
      "step 740/782 - loss: 1.0336 - acc_top1: 0.6211 - acc_top5: 0.9594 - 43ms/step\n",
      "step 750/782 - loss: 1.0120 - acc_top1: 0.6209 - acc_top5: 0.9593 - 43ms/step\n",
      "step 760/782 - loss: 1.2192 - acc_top1: 0.6209 - acc_top5: 0.9593 - 43ms/step\n",
      "step 770/782 - loss: 1.1584 - acc_top1: 0.6210 - acc_top5: 0.9591 - 42ms/step\n",
      "step 780/782 - loss: 0.9212 - acc_top1: 0.6215 - acc_top5: 0.9592 - 42ms/step\n",
      "step 782/782 - loss: 2.0734 - acc_top1: 0.6214 - acc_top5: 0.9593 - 43ms/step\n",
      "save checkpoint at /home/aistudio/output/11\n",
      "Eval begin...\n",
      "step  10/157 - loss: 0.9752 - acc_top1: 0.5813 - acc_top5: 0.9656 - 126ms/step\n",
      "step  20/157 - loss: 1.4892 - acc_top1: 0.5789 - acc_top5: 0.9609 - 72ms/step\n",
      "step  30/157 - loss: 1.2331 - acc_top1: 0.5802 - acc_top5: 0.9568 - 54ms/step\n",
      "step  40/157 - loss: 1.7198 - acc_top1: 0.5828 - acc_top5: 0.9535 - 45ms/step\n",
      "step  50/157 - loss: 1.2091 - acc_top1: 0.5819 - acc_top5: 0.9506 - 39ms/step\n",
      "step  60/157 - loss: 1.0781 - acc_top1: 0.5810 - acc_top5: 0.9497 - 36ms/step\n",
      "step  70/157 - loss: 3.2682 - acc_top1: 0.5826 - acc_top5: 0.9475 - 33ms/step\n",
      "step  80/157 - loss: 2.3858 - acc_top1: 0.5824 - acc_top5: 0.9490 - 31ms/step\n",
      "step  90/157 - loss: 1.0221 - acc_top1: 0.5793 - acc_top5: 0.9503 - 30ms/step\n",
      "step 100/157 - loss: 1.1021 - acc_top1: 0.5775 - acc_top5: 0.9506 - 29ms/step\n",
      "step 110/157 - loss: 1.5057 - acc_top1: 0.5740 - acc_top5: 0.9496 - 28ms/step\n",
      "step 120/157 - loss: 1.9533 - acc_top1: 0.5743 - acc_top5: 0.9503 - 27ms/step\n",
      "step 130/157 - loss: 19.0157 - acc_top1: 0.5745 - acc_top5: 0.9502 - 27ms/step\n",
      "step 140/157 - loss: 6.4069 - acc_top1: 0.5740 - acc_top5: 0.9509 - 27ms/step\n",
      "step 150/157 - loss: 1.0599 - acc_top1: 0.5742 - acc_top5: 0.9513 - 27ms/step\n",
      "step 157/157 - loss: 0.8641 - acc_top1: 0.5727 - acc_top5: 0.9512 - 28ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 13/50\n",
      "step  10/782 - loss: 0.8478 - acc_top1: 0.6609 - acc_top5: 0.9609 - 151ms/step\n",
      "step  20/782 - loss: 0.9759 - acc_top1: 0.6562 - acc_top5: 0.9672 - 95ms/step\n",
      "step  30/782 - loss: 1.1999 - acc_top1: 0.6557 - acc_top5: 0.9688 - 77ms/step\n",
      "step  40/782 - loss: 1.1417 - acc_top1: 0.6504 - acc_top5: 0.9645 - 68ms/step\n",
      "step  50/782 - loss: 1.0446 - acc_top1: 0.6512 - acc_top5: 0.9669 - 62ms/step\n",
      "step  60/782 - loss: 1.2022 - acc_top1: 0.6484 - acc_top5: 0.9667 - 58ms/step\n",
      "step  70/782 - loss: 1.2134 - acc_top1: 0.6531 - acc_top5: 0.9645 - 55ms/step\n",
      "step  80/782 - loss: 1.0400 - acc_top1: 0.6496 - acc_top5: 0.9656 - 53ms/step\n",
      "step  90/782 - loss: 0.9049 - acc_top1: 0.6467 - acc_top5: 0.9651 - 52ms/step\n",
      "step 100/782 - loss: 1.3064 - acc_top1: 0.6472 - acc_top5: 0.9664 - 50ms/step\n",
      "step 110/782 - loss: 1.1803 - acc_top1: 0.6457 - acc_top5: 0.9662 - 49ms/step\n",
      "step 120/782 - loss: 1.1364 - acc_top1: 0.6427 - acc_top5: 0.9663 - 48ms/step\n",
      "step 130/782 - loss: 1.0111 - acc_top1: 0.6389 - acc_top5: 0.9653 - 47ms/step\n",
      "step 140/782 - loss: 1.3072 - acc_top1: 0.6338 - acc_top5: 0.9651 - 47ms/step\n",
      "step 150/782 - loss: 0.9938 - acc_top1: 0.6324 - acc_top5: 0.9644 - 46ms/step\n",
      "step 160/782 - loss: 1.0421 - acc_top1: 0.6318 - acc_top5: 0.9644 - 46ms/step\n",
      "step 170/782 - loss: 1.4108 - acc_top1: 0.6295 - acc_top5: 0.9638 - 45ms/step\n",
      "step 180/782 - loss: 1.1067 - acc_top1: 0.6306 - acc_top5: 0.9637 - 45ms/step\n",
      "step 190/782 - loss: 1.0996 - acc_top1: 0.6303 - acc_top5: 0.9637 - 45ms/step\n",
      "step 200/782 - loss: 0.9243 - acc_top1: 0.6297 - acc_top5: 0.9634 - 44ms/step\n",
      "step 210/782 - loss: 1.1842 - acc_top1: 0.6301 - acc_top5: 0.9633 - 44ms/step\n",
      "step 220/782 - loss: 1.0331 - acc_top1: 0.6287 - acc_top5: 0.9634 - 44ms/step\n",
      "step 230/782 - loss: 1.0935 - acc_top1: 0.6258 - acc_top5: 0.9630 - 44ms/step\n",
      "step 240/782 - loss: 1.2125 - acc_top1: 0.6261 - acc_top5: 0.9626 - 43ms/step\n",
      "step 250/782 - loss: 1.0278 - acc_top1: 0.6267 - acc_top5: 0.9626 - 43ms/step\n",
      "step 260/782 - loss: 1.1649 - acc_top1: 0.6266 - acc_top5: 0.9622 - 43ms/step\n",
      "step 270/782 - loss: 0.9174 - acc_top1: 0.6271 - acc_top5: 0.9624 - 43ms/step\n",
      "step 280/782 - loss: 1.3065 - acc_top1: 0.6270 - acc_top5: 0.9622 - 43ms/step\n",
      "step 290/782 - loss: 0.9935 - acc_top1: 0.6271 - acc_top5: 0.9620 - 43ms/step\n",
      "step 300/782 - loss: 0.9519 - acc_top1: 0.6264 - acc_top5: 0.9622 - 42ms/step\n",
      "step 310/782 - loss: 0.8382 - acc_top1: 0.6269 - acc_top5: 0.9624 - 42ms/step\n",
      "step 320/782 - loss: 1.1875 - acc_top1: 0.6276 - acc_top5: 0.9623 - 42ms/step\n",
      "step 330/782 - loss: 1.0852 - acc_top1: 0.6278 - acc_top5: 0.9625 - 42ms/step\n",
      "step 340/782 - loss: 0.9450 - acc_top1: 0.6293 - acc_top5: 0.9628 - 42ms/step\n",
      "step 350/782 - loss: 1.0691 - acc_top1: 0.6293 - acc_top5: 0.9628 - 42ms/step\n",
      "step 360/782 - loss: 1.0939 - acc_top1: 0.6296 - acc_top5: 0.9632 - 42ms/step\n",
      "step 370/782 - loss: 1.1188 - acc_top1: 0.6298 - acc_top5: 0.9632 - 42ms/step\n",
      "step 380/782 - loss: 0.9536 - acc_top1: 0.6303 - acc_top5: 0.9633 - 42ms/step\n",
      "step 390/782 - loss: 0.8516 - acc_top1: 0.6299 - acc_top5: 0.9633 - 42ms/step\n",
      "step 400/782 - loss: 0.9861 - acc_top1: 0.6291 - acc_top5: 0.9632 - 41ms/step\n",
      "step 410/782 - loss: 1.2411 - acc_top1: 0.6290 - acc_top5: 0.9631 - 41ms/step\n",
      "step 420/782 - loss: 1.0869 - acc_top1: 0.6286 - acc_top5: 0.9630 - 41ms/step\n",
      "step 430/782 - loss: 1.4069 - acc_top1: 0.6279 - acc_top5: 0.9630 - 41ms/step\n",
      "step 440/782 - loss: 0.9389 - acc_top1: 0.6284 - acc_top5: 0.9630 - 41ms/step\n",
      "step 450/782 - loss: 1.1344 - acc_top1: 0.6286 - acc_top5: 0.9631 - 41ms/step\n",
      "step 460/782 - loss: 1.1371 - acc_top1: 0.6287 - acc_top5: 0.9630 - 41ms/step\n",
      "step 470/782 - loss: 1.1262 - acc_top1: 0.6292 - acc_top5: 0.9629 - 41ms/step\n",
      "step 480/782 - loss: 0.9341 - acc_top1: 0.6295 - acc_top5: 0.9629 - 41ms/step\n",
      "step 490/782 - loss: 1.2738 - acc_top1: 0.6298 - acc_top5: 0.9630 - 41ms/step\n",
      "step 500/782 - loss: 1.0966 - acc_top1: 0.6310 - acc_top5: 0.9630 - 41ms/step\n",
      "step 510/782 - loss: 1.1019 - acc_top1: 0.6310 - acc_top5: 0.9630 - 41ms/step\n",
      "step 520/782 - loss: 1.1181 - acc_top1: 0.6318 - acc_top5: 0.9633 - 41ms/step\n",
      "step 530/782 - loss: 0.8765 - acc_top1: 0.6318 - acc_top5: 0.9634 - 41ms/step\n",
      "step 540/782 - loss: 0.9909 - acc_top1: 0.6317 - acc_top5: 0.9636 - 41ms/step\n",
      "step 550/782 - loss: 1.2154 - acc_top1: 0.6312 - acc_top5: 0.9637 - 41ms/step\n",
      "step 560/782 - loss: 0.8403 - acc_top1: 0.6313 - acc_top5: 0.9636 - 41ms/step\n",
      "step 570/782 - loss: 0.9278 - acc_top1: 0.6319 - acc_top5: 0.9638 - 41ms/step\n",
      "step 580/782 - loss: 1.2405 - acc_top1: 0.6323 - acc_top5: 0.9638 - 41ms/step\n",
      "step 590/782 - loss: 1.0742 - acc_top1: 0.6326 - acc_top5: 0.9636 - 41ms/step\n",
      "step 600/782 - loss: 0.9932 - acc_top1: 0.6328 - acc_top5: 0.9637 - 41ms/step\n",
      "step 610/782 - loss: 1.0295 - acc_top1: 0.6330 - acc_top5: 0.9638 - 41ms/step\n",
      "step 620/782 - loss: 0.9833 - acc_top1: 0.6332 - acc_top5: 0.9638 - 41ms/step\n",
      "step 630/782 - loss: 0.9872 - acc_top1: 0.6337 - acc_top5: 0.9637 - 41ms/step\n",
      "step 640/782 - loss: 0.8789 - acc_top1: 0.6341 - acc_top5: 0.9637 - 41ms/step\n",
      "step 650/782 - loss: 0.8291 - acc_top1: 0.6346 - acc_top5: 0.9638 - 41ms/step\n",
      "step 660/782 - loss: 0.9088 - acc_top1: 0.6345 - acc_top5: 0.9638 - 41ms/step\n",
      "step 670/782 - loss: 0.8174 - acc_top1: 0.6350 - acc_top5: 0.9638 - 41ms/step\n",
      "step 680/782 - loss: 0.9940 - acc_top1: 0.6354 - acc_top5: 0.9639 - 41ms/step\n",
      "step 690/782 - loss: 1.2409 - acc_top1: 0.6354 - acc_top5: 0.9638 - 41ms/step\n",
      "step 700/782 - loss: 1.1594 - acc_top1: 0.6354 - acc_top5: 0.9640 - 41ms/step\n",
      "step 710/782 - loss: 1.0494 - acc_top1: 0.6352 - acc_top5: 0.9640 - 41ms/step\n",
      "step 720/782 - loss: 1.0842 - acc_top1: 0.6351 - acc_top5: 0.9638 - 41ms/step\n",
      "step 730/782 - loss: 0.9863 - acc_top1: 0.6349 - acc_top5: 0.9638 - 41ms/step\n",
      "step 740/782 - loss: 0.8595 - acc_top1: 0.6348 - acc_top5: 0.9639 - 41ms/step\n",
      "step 750/782 - loss: 0.8019 - acc_top1: 0.6353 - acc_top5: 0.9640 - 41ms/step\n",
      "step 760/782 - loss: 0.6700 - acc_top1: 0.6359 - acc_top5: 0.9640 - 41ms/step\n",
      "step 770/782 - loss: 0.8993 - acc_top1: 0.6360 - acc_top5: 0.9642 - 41ms/step\n",
      "step 780/782 - loss: 1.1337 - acc_top1: 0.6357 - acc_top5: 0.9642 - 40ms/step\n",
      "step 782/782 - loss: 1.4243 - acc_top1: 0.6356 - acc_top5: 0.9642 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/12\n",
      "Eval begin...\n",
      "step  10/157 - loss: 2.3632 - acc_top1: 0.6188 - acc_top5: 0.9594 - 122ms/step\n",
      "step  20/157 - loss: 1.5724 - acc_top1: 0.6156 - acc_top5: 0.9570 - 70ms/step\n",
      "step  30/157 - loss: 1.2713 - acc_top1: 0.6250 - acc_top5: 0.9578 - 52ms/step\n",
      "step  40/157 - loss: 1.6100 - acc_top1: 0.6148 - acc_top5: 0.9578 - 44ms/step\n",
      "step  50/157 - loss: 1.1591 - acc_top1: 0.6122 - acc_top5: 0.9578 - 39ms/step\n",
      "step  60/157 - loss: 5.1151 - acc_top1: 0.6161 - acc_top5: 0.9555 - 35ms/step\n",
      "step  70/157 - loss: 0.9578 - acc_top1: 0.6167 - acc_top5: 0.9551 - 33ms/step\n",
      "step  80/157 - loss: 1.2174 - acc_top1: 0.6150 - acc_top5: 0.9561 - 31ms/step\n",
      "step  90/157 - loss: 9.4756 - acc_top1: 0.6135 - acc_top5: 0.9569 - 29ms/step\n",
      "step 100/157 - loss: 1.5789 - acc_top1: 0.6131 - acc_top5: 0.9577 - 28ms/step\n",
      "step 110/157 - loss: 8.6104 - acc_top1: 0.6074 - acc_top5: 0.9560 - 27ms/step\n",
      "step 120/157 - loss: 0.9370 - acc_top1: 0.6044 - acc_top5: 0.9565 - 26ms/step\n",
      "step 130/157 - loss: 21.3226 - acc_top1: 0.6054 - acc_top5: 0.9563 - 26ms/step\n",
      "step 140/157 - loss: 20.2290 - acc_top1: 0.6038 - acc_top5: 0.9558 - 25ms/step\n",
      "step 150/157 - loss: 0.9051 - acc_top1: 0.6040 - acc_top5: 0.9556 - 25ms/step\n",
      "step 157/157 - loss: 1.7615 - acc_top1: 0.6026 - acc_top5: 0.9558 - 25ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 14/50\n",
      "step  10/782 - loss: 1.0899 - acc_top1: 0.6766 - acc_top5: 0.9609 - 152ms/step\n",
      "step  20/782 - loss: 0.9784 - acc_top1: 0.6711 - acc_top5: 0.9672 - 96ms/step\n",
      "step  30/782 - loss: 1.0106 - acc_top1: 0.6641 - acc_top5: 0.9651 - 77ms/step\n",
      "step  40/782 - loss: 1.0005 - acc_top1: 0.6652 - acc_top5: 0.9680 - 68ms/step\n",
      "step  50/782 - loss: 1.0898 - acc_top1: 0.6691 - acc_top5: 0.9688 - 62ms/step\n",
      "step  60/782 - loss: 0.8307 - acc_top1: 0.6708 - acc_top5: 0.9672 - 58ms/step\n",
      "step  70/782 - loss: 0.7782 - acc_top1: 0.6775 - acc_top5: 0.9679 - 55ms/step\n",
      "step  80/782 - loss: 0.9405 - acc_top1: 0.6783 - acc_top5: 0.9686 - 53ms/step\n",
      "step  90/782 - loss: 0.8936 - acc_top1: 0.6797 - acc_top5: 0.9701 - 51ms/step\n",
      "step 100/782 - loss: 1.5072 - acc_top1: 0.6748 - acc_top5: 0.9694 - 50ms/step\n",
      "step 110/782 - loss: 0.8162 - acc_top1: 0.6733 - acc_top5: 0.9700 - 50ms/step\n",
      "step 120/782 - loss: 0.9940 - acc_top1: 0.6750 - acc_top5: 0.9702 - 51ms/step\n",
      "step 130/782 - loss: 0.9705 - acc_top1: 0.6730 - acc_top5: 0.9706 - 50ms/step\n",
      "step 140/782 - loss: 0.6318 - acc_top1: 0.6757 - acc_top5: 0.9708 - 49ms/step\n",
      "step 150/782 - loss: 0.8170 - acc_top1: 0.6734 - acc_top5: 0.9697 - 49ms/step\n",
      "step 160/782 - loss: 0.8349 - acc_top1: 0.6743 - acc_top5: 0.9705 - 48ms/step\n",
      "step 170/782 - loss: 1.0056 - acc_top1: 0.6733 - acc_top5: 0.9711 - 48ms/step\n",
      "step 180/782 - loss: 0.7111 - acc_top1: 0.6723 - acc_top5: 0.9709 - 47ms/step\n",
      "step 190/782 - loss: 1.0244 - acc_top1: 0.6724 - acc_top5: 0.9711 - 47ms/step\n",
      "step 200/782 - loss: 0.9147 - acc_top1: 0.6734 - acc_top5: 0.9712 - 46ms/step\n",
      "step 210/782 - loss: 0.8148 - acc_top1: 0.6740 - acc_top5: 0.9714 - 46ms/step\n",
      "step 220/782 - loss: 1.1977 - acc_top1: 0.6744 - acc_top5: 0.9713 - 45ms/step\n",
      "step 230/782 - loss: 1.3609 - acc_top1: 0.6747 - acc_top5: 0.9708 - 46ms/step\n",
      "step 240/782 - loss: 0.9763 - acc_top1: 0.6757 - acc_top5: 0.9708 - 45ms/step\n",
      "step 250/782 - loss: 0.8340 - acc_top1: 0.6748 - acc_top5: 0.9711 - 45ms/step\n",
      "step 260/782 - loss: 0.9886 - acc_top1: 0.6755 - acc_top5: 0.9715 - 45ms/step\n",
      "step 270/782 - loss: 0.9841 - acc_top1: 0.6751 - acc_top5: 0.9715 - 44ms/step\n",
      "step 280/782 - loss: 0.9061 - acc_top1: 0.6739 - acc_top5: 0.9715 - 44ms/step\n",
      "step 290/782 - loss: 0.9264 - acc_top1: 0.6744 - acc_top5: 0.9717 - 44ms/step\n",
      "step 300/782 - loss: 0.7790 - acc_top1: 0.6753 - acc_top5: 0.9716 - 44ms/step\n",
      "step 310/782 - loss: 0.7767 - acc_top1: 0.6759 - acc_top5: 0.9715 - 44ms/step\n",
      "step 320/782 - loss: 0.8107 - acc_top1: 0.6765 - acc_top5: 0.9714 - 44ms/step\n",
      "step 330/782 - loss: 1.1733 - acc_top1: 0.6771 - acc_top5: 0.9716 - 43ms/step\n",
      "step 340/782 - loss: 0.8854 - acc_top1: 0.6773 - acc_top5: 0.9719 - 43ms/step\n",
      "step 350/782 - loss: 1.1899 - acc_top1: 0.6768 - acc_top5: 0.9717 - 43ms/step\n",
      "step 360/782 - loss: 0.9734 - acc_top1: 0.6766 - acc_top5: 0.9717 - 43ms/step\n",
      "step 370/782 - loss: 0.9109 - acc_top1: 0.6756 - acc_top5: 0.9711 - 43ms/step\n",
      "step 380/782 - loss: 0.9877 - acc_top1: 0.6757 - acc_top5: 0.9711 - 43ms/step\n",
      "step 390/782 - loss: 0.8782 - acc_top1: 0.6756 - acc_top5: 0.9712 - 43ms/step\n",
      "step 400/782 - loss: 0.8884 - acc_top1: 0.6754 - acc_top5: 0.9711 - 43ms/step\n",
      "step 410/782 - loss: 1.2175 - acc_top1: 0.6754 - acc_top5: 0.9712 - 43ms/step\n",
      "step 420/782 - loss: 0.9914 - acc_top1: 0.6752 - acc_top5: 0.9713 - 43ms/step\n",
      "step 430/782 - loss: 1.4514 - acc_top1: 0.6746 - acc_top5: 0.9715 - 43ms/step\n",
      "step 440/782 - loss: 1.0923 - acc_top1: 0.6736 - acc_top5: 0.9717 - 42ms/step\n",
      "step 450/782 - loss: 1.0543 - acc_top1: 0.6733 - acc_top5: 0.9719 - 42ms/step\n",
      "step 460/782 - loss: 1.0416 - acc_top1: 0.6731 - acc_top5: 0.9718 - 42ms/step\n",
      "step 470/782 - loss: 1.0564 - acc_top1: 0.6722 - acc_top5: 0.9717 - 42ms/step\n",
      "step 480/782 - loss: 1.3754 - acc_top1: 0.6715 - acc_top5: 0.9715 - 42ms/step\n",
      "step 490/782 - loss: 1.2356 - acc_top1: 0.6709 - acc_top5: 0.9716 - 42ms/step\n",
      "step 500/782 - loss: 1.0748 - acc_top1: 0.6705 - acc_top5: 0.9712 - 42ms/step\n",
      "step 510/782 - loss: 1.2690 - acc_top1: 0.6699 - acc_top5: 0.9711 - 42ms/step\n",
      "step 520/782 - loss: 0.8669 - acc_top1: 0.6690 - acc_top5: 0.9706 - 42ms/step\n",
      "step 530/782 - loss: 1.1235 - acc_top1: 0.6687 - acc_top5: 0.9704 - 42ms/step\n",
      "step 540/782 - loss: 1.1026 - acc_top1: 0.6683 - acc_top5: 0.9701 - 42ms/step\n",
      "step 550/782 - loss: 0.9592 - acc_top1: 0.6679 - acc_top5: 0.9701 - 42ms/step\n",
      "step 560/782 - loss: 1.3123 - acc_top1: 0.6673 - acc_top5: 0.9698 - 42ms/step\n",
      "step 570/782 - loss: 0.9326 - acc_top1: 0.6666 - acc_top5: 0.9697 - 42ms/step\n",
      "step 580/782 - loss: 1.0746 - acc_top1: 0.6660 - acc_top5: 0.9696 - 42ms/step\n",
      "step 590/782 - loss: 1.0551 - acc_top1: 0.6654 - acc_top5: 0.9696 - 42ms/step\n",
      "step 600/782 - loss: 1.1404 - acc_top1: 0.6654 - acc_top5: 0.9692 - 42ms/step\n",
      "step 610/782 - loss: 1.0934 - acc_top1: 0.6654 - acc_top5: 0.9692 - 42ms/step\n",
      "step 620/782 - loss: 0.9435 - acc_top1: 0.6652 - acc_top5: 0.9691 - 42ms/step\n",
      "step 630/782 - loss: 1.0412 - acc_top1: 0.6653 - acc_top5: 0.9690 - 42ms/step\n",
      "step 640/782 - loss: 0.9374 - acc_top1: 0.6651 - acc_top5: 0.9691 - 42ms/step\n",
      "step 650/782 - loss: 1.1366 - acc_top1: 0.6648 - acc_top5: 0.9692 - 42ms/step\n",
      "step 660/782 - loss: 1.3907 - acc_top1: 0.6644 - acc_top5: 0.9691 - 42ms/step\n",
      "step 670/782 - loss: 1.1990 - acc_top1: 0.6641 - acc_top5: 0.9691 - 41ms/step\n",
      "step 680/782 - loss: 0.8395 - acc_top1: 0.6639 - acc_top5: 0.9693 - 41ms/step\n",
      "step 690/782 - loss: 0.9350 - acc_top1: 0.6637 - acc_top5: 0.9693 - 41ms/step\n",
      "step 700/782 - loss: 1.0376 - acc_top1: 0.6638 - acc_top5: 0.9693 - 41ms/step\n",
      "step 710/782 - loss: 0.8444 - acc_top1: 0.6637 - acc_top5: 0.9694 - 41ms/step\n",
      "step 720/782 - loss: 0.8352 - acc_top1: 0.6639 - acc_top5: 0.9695 - 41ms/step\n",
      "step 730/782 - loss: 1.0294 - acc_top1: 0.6642 - acc_top5: 0.9696 - 41ms/step\n",
      "step 740/782 - loss: 0.9146 - acc_top1: 0.6636 - acc_top5: 0.9694 - 41ms/step\n",
      "step 750/782 - loss: 1.1693 - acc_top1: 0.6635 - acc_top5: 0.9694 - 41ms/step\n",
      "step 760/782 - loss: 1.1509 - acc_top1: 0.6634 - acc_top5: 0.9696 - 41ms/step\n",
      "step 770/782 - loss: 1.1464 - acc_top1: 0.6629 - acc_top5: 0.9696 - 41ms/step\n",
      "step 780/782 - loss: 1.1002 - acc_top1: 0.6626 - acc_top5: 0.9696 - 41ms/step\n",
      "step 782/782 - loss: 1.8784 - acc_top1: 0.6626 - acc_top5: 0.9696 - 41ms/step\n",
      "save checkpoint at /home/aistudio/output/13\n",
      "Eval begin...\n",
      "step  10/157 - loss: 0.9925 - acc_top1: 0.5984 - acc_top5: 0.9594 - 126ms/step\n",
      "step  20/157 - loss: 1.5037 - acc_top1: 0.6055 - acc_top5: 0.9570 - 73ms/step\n",
      "step  30/157 - loss: 1.1411 - acc_top1: 0.6141 - acc_top5: 0.9557 - 54ms/step\n",
      "step  40/157 - loss: 1.6466 - acc_top1: 0.6137 - acc_top5: 0.9570 - 45ms/step\n",
      "step  50/157 - loss: 1.1364 - acc_top1: 0.6119 - acc_top5: 0.9559 - 40ms/step\n",
      "step  60/157 - loss: 1.1194 - acc_top1: 0.6117 - acc_top5: 0.9557 - 36ms/step\n",
      "step  70/157 - loss: 3.2380 - acc_top1: 0.6100 - acc_top5: 0.9549 - 33ms/step\n",
      "step  80/157 - loss: 1.1259 - acc_top1: 0.6107 - acc_top5: 0.9561 - 31ms/step\n",
      "step  90/157 - loss: 0.9188 - acc_top1: 0.6097 - acc_top5: 0.9569 - 30ms/step\n",
      "step 100/157 - loss: 1.1229 - acc_top1: 0.6075 - acc_top5: 0.9573 - 29ms/step\n",
      "step 110/157 - loss: 5.1450 - acc_top1: 0.6045 - acc_top5: 0.9570 - 28ms/step\n",
      "step 120/157 - loss: 1.0198 - acc_top1: 0.6051 - acc_top5: 0.9578 - 27ms/step\n",
      "step 130/157 - loss: 5.4479 - acc_top1: 0.6056 - acc_top5: 0.9573 - 26ms/step\n",
      "step 140/157 - loss: 10.2446 - acc_top1: 0.6052 - acc_top5: 0.9574 - 26ms/step\n",
      "step 150/157 - loss: 0.8838 - acc_top1: 0.6060 - acc_top5: 0.9572 - 25ms/step\n",
      "step 157/157 - loss: 0.8826 - acc_top1: 0.6036 - acc_top5: 0.9568 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 15/50\n",
      "step  10/782 - loss: 0.8735 - acc_top1: 0.6531 - acc_top5: 0.9734 - 157ms/step\n",
      "step  20/782 - loss: 0.9908 - acc_top1: 0.6633 - acc_top5: 0.9711 - 98ms/step\n",
      "step  30/782 - loss: 0.9205 - acc_top1: 0.6635 - acc_top5: 0.9693 - 80ms/step\n",
      "step  40/782 - loss: 0.9800 - acc_top1: 0.6578 - acc_top5: 0.9684 - 70ms/step\n",
      "step  50/782 - loss: 0.9435 - acc_top1: 0.6644 - acc_top5: 0.9697 - 64ms/step\n",
      "step  60/782 - loss: 0.9855 - acc_top1: 0.6648 - acc_top5: 0.9693 - 60ms/step\n",
      "step  70/782 - loss: 0.7983 - acc_top1: 0.6696 - acc_top5: 0.9701 - 57ms/step\n",
      "step  80/782 - loss: 0.6049 - acc_top1: 0.6738 - acc_top5: 0.9713 - 55ms/step\n",
      "step  90/782 - loss: 0.9668 - acc_top1: 0.6755 - acc_top5: 0.9708 - 53ms/step\n",
      "step 100/782 - loss: 0.8488 - acc_top1: 0.6780 - acc_top5: 0.9719 - 52ms/step\n",
      "step 110/782 - loss: 0.8480 - acc_top1: 0.6759 - acc_top5: 0.9716 - 51ms/step\n",
      "step 120/782 - loss: 0.8307 - acc_top1: 0.6772 - acc_top5: 0.9725 - 50ms/step\n",
      "step 130/782 - loss: 0.7862 - acc_top1: 0.6781 - acc_top5: 0.9733 - 50ms/step\n",
      "step 140/782 - loss: 0.9747 - acc_top1: 0.6809 - acc_top5: 0.9734 - 49ms/step\n",
      "step 150/782 - loss: 0.7561 - acc_top1: 0.6813 - acc_top5: 0.9731 - 48ms/step\n",
      "step 160/782 - loss: 0.6552 - acc_top1: 0.6828 - acc_top5: 0.9738 - 47ms/step\n",
      "step 170/782 - loss: 0.6128 - acc_top1: 0.6842 - acc_top5: 0.9737 - 47ms/step\n",
      "step 180/782 - loss: 0.8425 - acc_top1: 0.6850 - acc_top5: 0.9741 - 47ms/step\n",
      "step 190/782 - loss: 0.9323 - acc_top1: 0.6861 - acc_top5: 0.9743 - 46ms/step\n",
      "step 200/782 - loss: 1.0796 - acc_top1: 0.6858 - acc_top5: 0.9741 - 46ms/step\n",
      "step 210/782 - loss: 0.7798 - acc_top1: 0.6858 - acc_top5: 0.9738 - 46ms/step\n",
      "step 220/782 - loss: 1.0069 - acc_top1: 0.6839 - acc_top5: 0.9740 - 45ms/step\n",
      "step 230/782 - loss: 1.1820 - acc_top1: 0.6845 - acc_top5: 0.9745 - 45ms/step\n",
      "step 240/782 - loss: 0.8003 - acc_top1: 0.6844 - acc_top5: 0.9745 - 45ms/step\n",
      "step 250/782 - loss: 0.7835 - acc_top1: 0.6843 - acc_top5: 0.9745 - 44ms/step\n",
      "step 260/782 - loss: 1.0911 - acc_top1: 0.6849 - acc_top5: 0.9744 - 44ms/step\n",
      "step 270/782 - loss: 1.2907 - acc_top1: 0.6853 - acc_top5: 0.9744 - 44ms/step\n",
      "step 280/782 - loss: 0.9698 - acc_top1: 0.6857 - acc_top5: 0.9744 - 44ms/step\n",
      "step 290/782 - loss: 0.9225 - acc_top1: 0.6858 - acc_top5: 0.9744 - 44ms/step\n",
      "step 300/782 - loss: 1.1380 - acc_top1: 0.6854 - acc_top5: 0.9744 - 43ms/step\n",
      "step 310/782 - loss: 0.8858 - acc_top1: 0.6861 - acc_top5: 0.9742 - 43ms/step\n",
      "step 320/782 - loss: 0.9519 - acc_top1: 0.6861 - acc_top5: 0.9745 - 43ms/step\n",
      "step 330/782 - loss: 1.1175 - acc_top1: 0.6862 - acc_top5: 0.9742 - 43ms/step\n",
      "step 340/782 - loss: 0.8497 - acc_top1: 0.6874 - acc_top5: 0.9743 - 43ms/step\n",
      "step 350/782 - loss: 0.9632 - acc_top1: 0.6870 - acc_top5: 0.9744 - 43ms/step\n",
      "step 360/782 - loss: 0.8788 - acc_top1: 0.6872 - acc_top5: 0.9742 - 43ms/step\n",
      "step 370/782 - loss: 1.0423 - acc_top1: 0.6872 - acc_top5: 0.9742 - 43ms/step\n",
      "step 380/782 - loss: 0.9921 - acc_top1: 0.6873 - acc_top5: 0.9741 - 43ms/step\n",
      "step 390/782 - loss: 0.9040 - acc_top1: 0.6866 - acc_top5: 0.9743 - 43ms/step\n",
      "step 400/782 - loss: 0.7984 - acc_top1: 0.6868 - acc_top5: 0.9741 - 43ms/step\n",
      "step 410/782 - loss: 0.7409 - acc_top1: 0.6870 - acc_top5: 0.9741 - 43ms/step\n",
      "step 420/782 - loss: 0.7648 - acc_top1: 0.6868 - acc_top5: 0.9738 - 43ms/step\n",
      "step 430/782 - loss: 0.7322 - acc_top1: 0.6864 - acc_top5: 0.9735 - 42ms/step\n",
      "step 440/782 - loss: 0.9378 - acc_top1: 0.6864 - acc_top5: 0.9737 - 42ms/step\n",
      "step 450/782 - loss: 1.1441 - acc_top1: 0.6857 - acc_top5: 0.9736 - 42ms/step\n",
      "step 460/782 - loss: 0.6785 - acc_top1: 0.6855 - acc_top5: 0.9734 - 42ms/step\n",
      "step 470/782 - loss: 0.9576 - acc_top1: 0.6862 - acc_top5: 0.9735 - 42ms/step\n",
      "step 480/782 - loss: 0.8155 - acc_top1: 0.6869 - acc_top5: 0.9738 - 42ms/step\n",
      "step 490/782 - loss: 0.6965 - acc_top1: 0.6869 - acc_top5: 0.9738 - 42ms/step\n",
      "step 500/782 - loss: 0.8700 - acc_top1: 0.6870 - acc_top5: 0.9738 - 43ms/step\n",
      "step 510/782 - loss: 0.9571 - acc_top1: 0.6877 - acc_top5: 0.9739 - 43ms/step\n",
      "step 520/782 - loss: 0.8204 - acc_top1: 0.6879 - acc_top5: 0.9738 - 43ms/step\n",
      "step 530/782 - loss: 0.8857 - acc_top1: 0.6882 - acc_top5: 0.9738 - 43ms/step\n",
      "step 540/782 - loss: 0.9525 - acc_top1: 0.6881 - acc_top5: 0.9739 - 43ms/step\n",
      "step 550/782 - loss: 1.3814 - acc_top1: 0.6881 - acc_top5: 0.9737 - 43ms/step\n",
      "step 560/782 - loss: 0.8854 - acc_top1: 0.6883 - acc_top5: 0.9737 - 43ms/step\n",
      "step 570/782 - loss: 0.8798 - acc_top1: 0.6884 - acc_top5: 0.9737 - 43ms/step\n",
      "step 580/782 - loss: 0.7131 - acc_top1: 0.6882 - acc_top5: 0.9738 - 43ms/step\n",
      "step 590/782 - loss: 1.0210 - acc_top1: 0.6880 - acc_top5: 0.9741 - 43ms/step\n",
      "step 600/782 - loss: 0.8457 - acc_top1: 0.6880 - acc_top5: 0.9741 - 42ms/step\n",
      "step 610/782 - loss: 0.7631 - acc_top1: 0.6884 - acc_top5: 0.9741 - 42ms/step\n",
      "step 620/782 - loss: 0.8255 - acc_top1: 0.6890 - acc_top5: 0.9740 - 42ms/step\n",
      "step 630/782 - loss: 1.0459 - acc_top1: 0.6888 - acc_top5: 0.9739 - 42ms/step\n",
      "step 640/782 - loss: 0.8980 - acc_top1: 0.6890 - acc_top5: 0.9740 - 42ms/step\n",
      "step 650/782 - loss: 0.9476 - acc_top1: 0.6892 - acc_top5: 0.9742 - 42ms/step\n",
      "step 660/782 - loss: 1.0482 - acc_top1: 0.6895 - acc_top5: 0.9742 - 42ms/step\n",
      "step 670/782 - loss: 0.8917 - acc_top1: 0.6899 - acc_top5: 0.9743 - 42ms/step\n",
      "step 680/782 - loss: 0.9354 - acc_top1: 0.6899 - acc_top5: 0.9744 - 42ms/step\n",
      "step 690/782 - loss: 0.9783 - acc_top1: 0.6904 - acc_top5: 0.9745 - 42ms/step\n",
      "step 700/782 - loss: 0.6572 - acc_top1: 0.6904 - acc_top5: 0.9745 - 42ms/step\n",
      "step 710/782 - loss: 0.8458 - acc_top1: 0.6907 - acc_top5: 0.9746 - 42ms/step\n",
      "step 720/782 - loss: 1.2156 - acc_top1: 0.6905 - acc_top5: 0.9746 - 42ms/step\n",
      "step 730/782 - loss: 1.0192 - acc_top1: 0.6900 - acc_top5: 0.9747 - 42ms/step\n",
      "step 740/782 - loss: 0.9645 - acc_top1: 0.6899 - acc_top5: 0.9748 - 42ms/step\n",
      "step 750/782 - loss: 0.8516 - acc_top1: 0.6900 - acc_top5: 0.9749 - 42ms/step\n",
      "step 760/782 - loss: 0.9647 - acc_top1: 0.6898 - acc_top5: 0.9748 - 42ms/step\n",
      "step 770/782 - loss: 0.9237 - acc_top1: 0.6897 - acc_top5: 0.9748 - 42ms/step\n",
      "step 780/782 - loss: 0.5758 - acc_top1: 0.6900 - acc_top5: 0.9750 - 42ms/step\n",
      "step 782/782 - loss: 1.7726 - acc_top1: 0.6898 - acc_top5: 0.9749 - 42ms/step\n",
      "save checkpoint at /home/aistudio/output/14\n",
      "Eval begin...\n",
      "step  10/157 - loss: 1.4529 - acc_top1: 0.6000 - acc_top5: 0.9641 - 124ms/step\n",
      "step  20/157 - loss: 1.4597 - acc_top1: 0.6109 - acc_top5: 0.9609 - 71ms/step\n",
      "step  30/157 - loss: 1.1284 - acc_top1: 0.6068 - acc_top5: 0.9568 - 53ms/step\n",
      "step  40/157 - loss: 2.6128 - acc_top1: 0.6023 - acc_top5: 0.9539 - 44ms/step\n",
      "step  50/157 - loss: 1.2681 - acc_top1: 0.6072 - acc_top5: 0.9547 - 39ms/step\n",
      "step  60/157 - loss: 2.9917 - acc_top1: 0.6070 - acc_top5: 0.9516 - 36ms/step\n",
      "step  70/157 - loss: 1.0797 - acc_top1: 0.6051 - acc_top5: 0.9489 - 33ms/step\n",
      "step  80/157 - loss: 1.1749 - acc_top1: 0.6037 - acc_top5: 0.9500 - 32ms/step\n",
      "step  90/157 - loss: 2.3619 - acc_top1: 0.6005 - acc_top5: 0.9512 - 30ms/step\n",
      "step 100/157 - loss: 1.0627 - acc_top1: 0.6008 - acc_top5: 0.9517 - 29ms/step\n",
      "step 110/157 - loss: 4.1284 - acc_top1: 0.5993 - acc_top5: 0.9501 - 28ms/step\n",
      "step 120/157 - loss: 1.0099 - acc_top1: 0.5990 - acc_top5: 0.9500 - 27ms/step\n",
      "step 130/157 - loss: 4.2597 - acc_top1: 0.6013 - acc_top5: 0.9488 - 27ms/step\n",
      "step 140/157 - loss: 11.1688 - acc_top1: 0.6008 - acc_top5: 0.9494 - 26ms/step\n",
      "step 150/157 - loss: 1.0346 - acc_top1: 0.6007 - acc_top5: 0.9484 - 25ms/step\n",
      "step 157/157 - loss: 2.9027 - acc_top1: 0.5976 - acc_top5: 0.9477 - 26ms/step\n",
      "Eval samples: 10000\n",
      "Epoch 16/50\n",
      "step  10/782 - loss: 0.8264 - acc_top1: 0.7453 - acc_top5: 0.9766 - 167ms/step\n",
      "step  20/782 - loss: 0.6475 - acc_top1: 0.7375 - acc_top5: 0.9789 - 113ms/step\n",
      "step  30/782 - loss: 0.8607 - acc_top1: 0.7396 - acc_top5: 0.9792 - 89ms/step\n",
      "step  40/782 - loss: 0.7602 - acc_top1: 0.7391 - acc_top5: 0.9793 - 76ms/step\n",
      "step  50/782 - loss: 0.6492 - acc_top1: 0.7359 - acc_top5: 0.9803 - 69ms/step\n",
      "step  60/782 - loss: 1.1043 - acc_top1: 0.7333 - acc_top5: 0.9805 - 64ms/step\n",
      "step  70/782 - loss: 0.9116 - acc_top1: 0.7312 - acc_top5: 0.9801 - 61ms/step\n",
      "step  80/782 - loss: 0.7111 - acc_top1: 0.7250 - acc_top5: 0.9814 - 58ms/step\n",
      "step  90/782 - loss: 0.8301 - acc_top1: 0.7264 - acc_top5: 0.9823 - 56ms/step\n",
      "step 100/782 - loss: 1.0029 - acc_top1: 0.7281 - acc_top5: 0.9830 - 54ms/step\n",
      "step 110/782 - loss: 0.6427 - acc_top1: 0.7283 - acc_top5: 0.9832 - 53ms/step\n",
      "step 120/782 - loss: 0.6421 - acc_top1: 0.7283 - acc_top5: 0.9839 - 52ms/step\n",
      "step 130/782 - loss: 0.6232 - acc_top1: 0.7302 - acc_top5: 0.9845 - 51ms/step\n"
     ]
    },
    {
     "ename": "KeyboardInterrupt",
     "evalue": "",
     "output_type": "error",
     "traceback": [
      "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
      "\u001b[0;31mKeyboardInterrupt\u001b[0m                         Traceback (most recent call last)",
      "\u001b[0;32m/tmp/ipykernel_152/1147452275.py\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n\u001b[1;32m     18\u001b[0m           \u001b[0mbatch_size\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;36m64\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m     19\u001b[0m           \u001b[0msave_dir\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;34m\"./output\"\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m---> 20\u001b[0;31m           num_workers=8)\n\u001b[0m",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model.py\u001b[0m in \u001b[0;36mfit\u001b[0;34m(self, train_data, eval_data, batch_size, epochs, eval_freq, log_freq, save_dir, save_freq, verbose, drop_last, shuffle, num_workers, callbacks, accumulate_grad_batches, num_iters)\u001b[0m\n\u001b[1;32m   1730\u001b[0m         \u001b[0;32mfor\u001b[0m \u001b[0mepoch\u001b[0m \u001b[0;32min\u001b[0m \u001b[0mrange\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mepochs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m   1731\u001b[0m             \u001b[0mcbks\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mon_epoch_begin\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mepoch\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1732\u001b[0;31m             \u001b[0mlogs\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_run_one_epoch\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mtrain_loader\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mcbks\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m'train'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m   1733\u001b[0m             \u001b[0mcbks\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mon_epoch_end\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mepoch\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mlogs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m   1734\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model.py\u001b[0m in \u001b[0;36m_run_one_epoch\u001b[0;34m(self, data_loader, callbacks, mode, logs)\u001b[0m\n\u001b[1;32m   2060\u001b[0m                                    step + 1 == len(data_loader))\n\u001b[1;32m   2061\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 2062\u001b[0;31m                 \u001b[0mouts\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mgetattr\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mmode\u001b[0m \u001b[0;34m+\u001b[0m \u001b[0;34m'_batch'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0m_inputs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m   2063\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m   2064\u001b[0m                 \u001b[0;32mif\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_metrics\u001b[0m \u001b[0;32mand\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_loss\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model.py\u001b[0m in \u001b[0;36mtrain_batch\u001b[0;34m(self, inputs, labels, update)\u001b[0m\n\u001b[1;32m   1059\u001b[0m               \u001b[0mprint\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mloss\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m   1060\u001b[0m         \"\"\"\n\u001b[0;32m-> 1061\u001b[0;31m         \u001b[0mloss\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_adapter\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mtrain_batch\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mlabels\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mupdate\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m   1062\u001b[0m         \u001b[0;32mif\u001b[0m \u001b[0mfluid\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0min_dygraph_mode\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m \u001b[0;32mand\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_input_info\u001b[0m \u001b[0;32mis\u001b[0m \u001b[0;32mNone\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m   1063\u001b[0m             \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_update_inputs\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model.py\u001b[0m in \u001b[0;36mtrain_batch\u001b[0;34m(self, inputs, labels, update)\u001b[0m\n\u001b[1;32m    712\u001b[0m             \u001b[0;32melse\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    713\u001b[0m                 outputs = self.model.network.forward(\n\u001b[0;32m--> 714\u001b[0;31m                     *[to_variable(x) for x in inputs])\n\u001b[0m\u001b[1;32m    715\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    716\u001b[0m             \u001b[0mlosses\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mmodel\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_loss\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mto_list\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0moutputs\u001b[0m\u001b[0;34m)\u001b[0m \u001b[0;34m+\u001b[0m \u001b[0mlabels\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/vision/models/resnet.py\u001b[0m in \u001b[0;36mforward\u001b[0;34m(self, x)\u001b[0m\n\u001b[1;32m    240\u001b[0m         \u001b[0mx\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mmaxpool\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    241\u001b[0m         \u001b[0mx\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mlayer1\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 242\u001b[0;31m         \u001b[0mx\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mlayer2\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    243\u001b[0m         \u001b[0mx\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mlayer3\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    244\u001b[0m         \u001b[0mx\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mlayer4\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m__call__\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    915\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    916\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0m__call__\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 917\u001b[0;31m         \u001b[0;32mreturn\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_dygraph_call_func\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    918\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    919\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m_dygraph_call_func\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    905\u001b[0m             \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_built\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;32mTrue\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    906\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 907\u001b[0;31m         \u001b[0moutputs\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    908\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    909\u001b[0m         \u001b[0;32mfor\u001b[0m \u001b[0mforward_post_hook\u001b[0m \u001b[0;32min\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_forward_post_hooks\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mvalues\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/container.py\u001b[0m in \u001b[0;36mforward\u001b[0;34m(self, input)\u001b[0m\n\u001b[1;32m     96\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0minput\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m     97\u001b[0m         \u001b[0;32mfor\u001b[0m \u001b[0mlayer\u001b[0m \u001b[0;32min\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_sub_layers\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mvalues\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m---> 98\u001b[0;31m             \u001b[0minput\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mlayer\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0minput\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m     99\u001b[0m         \u001b[0;32mreturn\u001b[0m \u001b[0minput\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    100\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m__call__\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    915\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    916\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0m__call__\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 917\u001b[0;31m         \u001b[0;32mreturn\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_dygraph_call_func\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    918\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    919\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m_dygraph_call_func\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    905\u001b[0m             \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_built\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;32mTrue\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    906\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 907\u001b[0;31m         \u001b[0moutputs\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    908\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    909\u001b[0m         \u001b[0;32mfor\u001b[0m \u001b[0mforward_post_hook\u001b[0m \u001b[0;32min\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_forward_post_hooks\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mvalues\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/vision/models/resnet.py\u001b[0m in \u001b[0;36mforward\u001b[0;34m(self, x)\u001b[0m\n\u001b[1;32m    131\u001b[0m         \u001b[0mout\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mrelu\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mout\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    132\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 133\u001b[0;31m         \u001b[0mout\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mconv2\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mout\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    134\u001b[0m         \u001b[0mout\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mbn2\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mout\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    135\u001b[0m         \u001b[0mout\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mrelu\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mout\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m__call__\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    915\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    916\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0m__call__\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 917\u001b[0;31m         \u001b[0;32mreturn\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_dygraph_call_func\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    918\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    919\u001b[0m     \u001b[0;32mdef\u001b[0m \u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py\u001b[0m in \u001b[0;36m_dygraph_call_func\u001b[0;34m(self, *inputs, **kwargs)\u001b[0m\n\u001b[1;32m    905\u001b[0m             \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_built\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0;32mTrue\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    906\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 907\u001b[0;31m         \u001b[0moutputs\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mforward\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m*\u001b[0m\u001b[0minputs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m**\u001b[0m\u001b[0mkwargs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    908\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    909\u001b[0m         \u001b[0;32mfor\u001b[0m \u001b[0mforward_post_hook\u001b[0m \u001b[0;32min\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_forward_post_hooks\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mvalues\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/conv.py\u001b[0m in \u001b[0;36mforward\u001b[0;34m(self, x)\u001b[0m\n\u001b[1;32m    675\u001b[0m             \u001b[0mchannel_dim\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_channel_dim\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    676\u001b[0m             \u001b[0mop_type\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0m_op_type\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 677\u001b[0;31m             use_cudnn=self._use_cudnn)\n\u001b[0m\u001b[1;32m    678\u001b[0m         \u001b[0;32mreturn\u001b[0m \u001b[0mout\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    679\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;32m/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/functional/conv.py\u001b[0m in \u001b[0;36m_conv_nd\u001b[0;34m(x, weight, bias, stride, padding, padding_algorithm, dilation, groups, data_format, channel_dim, op_type, use_cudnn, use_mkldnn, name)\u001b[0m\n\u001b[1;32m    121\u001b[0m                  \u001b[0;34m\"padding_algorithm\"\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mpadding_algorithm\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m\"data_format\"\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    122\u001b[0m                  data_format)\n\u001b[0;32m--> 123\u001b[0;31m         \u001b[0mpre_bias\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mgetattr\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0m_C_ops\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mop_type\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mx\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mweight\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0;34m*\u001b[0m\u001b[0mattrs\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m    124\u001b[0m         \u001b[0;32mif\u001b[0m \u001b[0mbias\u001b[0m \u001b[0;32mis\u001b[0m \u001b[0;32mnot\u001b[0m \u001b[0;32mNone\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m    125\u001b[0m             \u001b[0mout\u001b[0m \u001b[0;34m=\u001b[0m \u001b[0mnn\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0melementwise_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mpre_bias\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mbias\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0maxis\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mchannel_dim\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
      "\u001b[0;31mKeyboardInterrupt\u001b[0m: "
     ]
    }
   ],
   "source": [
    "# 确保从paddle.vision.datasets.Cifar10中加载的图像数据是np.ndarray类型\n",
    "paddle.vision.set_image_backend('cv2')\n",
    "# 调用resnet50模型\n",
    "model = paddle.Model(resnet50(pretrained=False, num_classes=10))\n",
    "\n",
    "\n",
    "# 定义优化器\n",
    "optimizer = Momentum(learning_rate=0.01,\n",
    "                     momentum=0.9,\n",
    "                     weight_decay=L2Decay(1e-4),\n",
    "                     parameters=model.parameters())\n",
    "# 进行训练前准备\n",
    "model.prepare(optimizer, CrossEntropyLoss(), Accuracy(topk=(1, 5)))\n",
    "# 启动训练\n",
    "model.fit(train_dataset,\n",
    "          val_dataset,\n",
    "          epochs=50,\n",
    "          batch_size=64,\n",
    "          save_dir=\"./output\",\n",
    "          num_workers=8)"
   ]
  },
  {
   "cell_type": "code",
   "execution_count": null,
   "metadata": {
    "execution": {
     "iopub.status.busy": "2022-02-23T15:19:27.073874Z",
     "iopub.status.idle": "2022-02-23T15:19:27.150917Z",
     "shell.execute_reply": "2022-02-23T15:19:27.150403Z",
     "shell.execute_reply.started": "2022-02-23T15:19:27.150329Z"
    },
    "scrolled": true
   },
   "outputs": [],
   "source": [
    "# 模型评估\n",
    "model.evaluate(val_dataset, batch_size=64, verbose=1)"
   ]
  },
  {
   "cell_type": "code",
   "execution_count": null,
   "metadata": {
    "scrolled": true
   },
   "outputs": [],
   "source": [
    "# 保存模型参数\n",
    "model.save('RESNEGT50')  # save for training\n",
    "model.save('RESNEGT50', False)  # save for inference"
   ]
  }
 ],
 "metadata": {
  "kernelspec": {
   "display_name": "Python 3",
   "language": "python",
   "name": "py35-paddle1.2.0"
  },
  "language_info": {
   "codemirror_mode": {
    "name": "ipython",
    "version": 3
   },
   "file_extension": ".py",
   "mimetype": "text/x-python",
   "name": "python",
   "nbconvert_exporter": "python",
   "pygments_lexer": "ipython3",
   "version": "3.7.4"
  }
 },
 "nbformat": 4,
 "nbformat_minor": 4
}
