vjt commited on
Commit
34d8372
β€’
1 Parent(s): 0aa4519

update model card README.md

Browse files
Files changed (2) hide show
  1. README.md +65 -0
  2. T5Train.ipynb +365 -10
README.md ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - wikisql
7
+ model-index:
8
+ - name: T5Training
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # T5Training
16
+
17
+ This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the wikisql dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.0341
20
+ - Rouge2 Precision: 0.9368
21
+ - Rouge2 Recall: 0.8687
22
+ - Rouge2 Fmeasure: 0.896
23
+
24
+ ## Model description
25
+
26
+ More information needed
27
+
28
+ ## Intended uses & limitations
29
+
30
+ More information needed
31
+
32
+ ## Training and evaluation data
33
+
34
+ More information needed
35
+
36
+ ## Training procedure
37
+
38
+ ### Training hyperparameters
39
+
40
+ The following hyperparameters were used during training:
41
+ - learning_rate: 5e-05
42
+ - train_batch_size: 16
43
+ - eval_batch_size: 16
44
+ - seed: 42
45
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
+ - lr_scheduler_type: linear
47
+ - num_epochs: 5
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure |
52
+ |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:|
53
+ | 0.0225 | 1.0 | 4049 | 0.0355 | 0.9325 | 0.8665 | 0.8929 |
54
+ | 0.0182 | 2.0 | 8098 | 0.0359 | 0.9364 | 0.8682 | 0.8956 |
55
+ | 0.016 | 3.0 | 12147 | 0.0354 | 0.9353 | 0.868 | 0.895 |
56
+ | 0.0156 | 4.0 | 16196 | 0.0351 | 0.9366 | 0.8684 | 0.8958 |
57
+ | 0.0177 | 5.0 | 20245 | 0.0341 | 0.9368 | 0.8687 | 0.896 |
58
+
59
+
60
+ ### Framework versions
61
+
62
+ - Transformers 4.26.0
63
+ - Pytorch 1.13.1+cu117
64
+ - Datasets 2.8.0
65
+ - Tokenizers 0.13.2
T5Train.ipynb CHANGED
@@ -199,7 +199,7 @@
199
  {
200
  "cell_type": "code",
201
  "execution_count": null,
202
- "id": "b0a1835b",
203
  "metadata": {},
204
  "outputs": [],
205
  "source": []
@@ -7534,8 +7534,8 @@
7534
  "\n",
7535
  " <div>\n",
7536
  " \n",
7537
- " <progress value='20246' max='20245' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
7538
- " [20245/20245 4:07:47, Epoch 5/5]\n",
7539
  " </div>\n",
7540
  " <table border=\"1\" class=\"dataframe\">\n",
7541
  " <thead>\n",
@@ -7581,14 +7581,16 @@
7581
  " <td>0.868400</td>\n",
7582
  " <td>0.895800</td>\n",
7583
  " </tr>\n",
 
 
 
 
 
 
 
 
7584
  " </tbody>\n",
7585
- "</table><p>\n",
7586
- " <div>\n",
7587
- " \n",
7588
- " <progress value='946' max='993' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
7589
- " [946/993 09:10 < 00:27, 1.72 it/s]\n",
7590
- " </div>\n",
7591
- " "
7592
  ],
7593
  "text/plain": [
7594
  "<IPython.core.display.HTML object>"
@@ -42584,8 +42586,361 @@
42584
  " \"pad_token_id\": 0,\n",
42585
  " \"transformers_version\": \"4.26.0\"\n",
42586
  "}\n",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42587
  "\n"
42588
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42589
  }
42590
  ],
42591
  "source": [
 
199
  {
200
  "cell_type": "code",
201
  "execution_count": null,
202
+ "id": "76f3fc6f",
203
  "metadata": {},
204
  "outputs": [],
205
  "source": []
 
7534
  "\n",
7535
  " <div>\n",
7536
  " \n",
7537
+ " <progress value='20245' max='20245' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
7538
+ " [20245/20245 4:18:19, Epoch 5/5]\n",
7539
  " </div>\n",
7540
  " <table border=\"1\" class=\"dataframe\">\n",
7541
  " <thead>\n",
 
7581
  " <td>0.868400</td>\n",
7582
  " <td>0.895800</td>\n",
7583
  " </tr>\n",
7584
+ " <tr>\n",
7585
+ " <td>5</td>\n",
7586
+ " <td>0.017700</td>\n",
7587
+ " <td>0.034084</td>\n",
7588
+ " <td>0.936800</td>\n",
7589
+ " <td>0.868700</td>\n",
7590
+ " <td>0.896000</td>\n",
7591
+ " </tr>\n",
7592
  " </tbody>\n",
7593
+ "</table><p>"
 
 
 
 
 
 
7594
  ],
7595
  "text/plain": [
7596
  "<IPython.core.display.HTML object>"
 
42586
  " \"pad_token_id\": 0,\n",
42587
  " \"transformers_version\": \"4.26.0\"\n",
42588
  "}\n",
42589
+ "\n",
42590
+ "Generate config GenerationConfig {\n",
42591
+ " \"decoder_start_token_id\": 0,\n",
42592
+ " \"eos_token_id\": 1,\n",
42593
+ " \"pad_token_id\": 0,\n",
42594
+ " \"transformers_version\": \"4.26.0\"\n",
42595
+ "}\n",
42596
+ "\n",
42597
+ "Generate config GenerationConfig {\n",
42598
+ " \"decoder_start_token_id\": 0,\n",
42599
+ " \"eos_token_id\": 1,\n",
42600
+ " \"pad_token_id\": 0,\n",
42601
+ " \"transformers_version\": \"4.26.0\"\n",
42602
+ "}\n",
42603
+ "\n",
42604
+ "Generate config GenerationConfig {\n",
42605
+ " \"decoder_start_token_id\": 0,\n",
42606
+ " \"eos_token_id\": 1,\n",
42607
+ " \"pad_token_id\": 0,\n",
42608
+ " \"transformers_version\": \"4.26.0\"\n",
42609
+ "}\n",
42610
+ "\n",
42611
+ "Generate config GenerationConfig {\n",
42612
+ " \"decoder_start_token_id\": 0,\n",
42613
+ " \"eos_token_id\": 1,\n",
42614
+ " \"pad_token_id\": 0,\n",
42615
+ " \"transformers_version\": \"4.26.0\"\n",
42616
+ "}\n",
42617
+ "\n",
42618
+ "Generate config GenerationConfig {\n",
42619
+ " \"decoder_start_token_id\": 0,\n",
42620
+ " \"eos_token_id\": 1,\n",
42621
+ " \"pad_token_id\": 0,\n",
42622
+ " \"transformers_version\": \"4.26.0\"\n",
42623
+ "}\n",
42624
+ "\n",
42625
+ "Generate config GenerationConfig {\n",
42626
+ " \"decoder_start_token_id\": 0,\n",
42627
+ " \"eos_token_id\": 1,\n",
42628
+ " \"pad_token_id\": 0,\n",
42629
+ " \"transformers_version\": \"4.26.0\"\n",
42630
+ "}\n",
42631
+ "\n",
42632
+ "Generate config GenerationConfig {\n",
42633
+ " \"decoder_start_token_id\": 0,\n",
42634
+ " \"eos_token_id\": 1,\n",
42635
+ " \"pad_token_id\": 0,\n",
42636
+ " \"transformers_version\": \"4.26.0\"\n",
42637
+ "}\n",
42638
+ "\n",
42639
+ "Generate config GenerationConfig {\n",
42640
+ " \"decoder_start_token_id\": 0,\n",
42641
+ " \"eos_token_id\": 1,\n",
42642
+ " \"pad_token_id\": 0,\n",
42643
+ " \"transformers_version\": \"4.26.0\"\n",
42644
+ "}\n",
42645
+ "\n",
42646
+ "Generate config GenerationConfig {\n",
42647
+ " \"decoder_start_token_id\": 0,\n",
42648
+ " \"eos_token_id\": 1,\n",
42649
+ " \"pad_token_id\": 0,\n",
42650
+ " \"transformers_version\": \"4.26.0\"\n",
42651
+ "}\n",
42652
+ "\n",
42653
+ "Generate config GenerationConfig {\n",
42654
+ " \"decoder_start_token_id\": 0,\n",
42655
+ " \"eos_token_id\": 1,\n",
42656
+ " \"pad_token_id\": 0,\n",
42657
+ " \"transformers_version\": \"4.26.0\"\n",
42658
+ "}\n",
42659
+ "\n",
42660
+ "Generate config GenerationConfig {\n",
42661
+ " \"decoder_start_token_id\": 0,\n",
42662
+ " \"eos_token_id\": 1,\n",
42663
+ " \"pad_token_id\": 0,\n",
42664
+ " \"transformers_version\": \"4.26.0\"\n",
42665
+ "}\n",
42666
+ "\n",
42667
+ "Generate config GenerationConfig {\n",
42668
+ " \"decoder_start_token_id\": 0,\n",
42669
+ " \"eos_token_id\": 1,\n",
42670
+ " \"pad_token_id\": 0,\n",
42671
+ " \"transformers_version\": \"4.26.0\"\n",
42672
+ "}\n",
42673
+ "\n",
42674
+ "Generate config GenerationConfig {\n",
42675
+ " \"decoder_start_token_id\": 0,\n",
42676
+ " \"eos_token_id\": 1,\n",
42677
+ " \"pad_token_id\": 0,\n",
42678
+ " \"transformers_version\": \"4.26.0\"\n",
42679
+ "}\n",
42680
+ "\n",
42681
+ "Generate config GenerationConfig {\n",
42682
+ " \"decoder_start_token_id\": 0,\n",
42683
+ " \"eos_token_id\": 1,\n",
42684
+ " \"pad_token_id\": 0,\n",
42685
+ " \"transformers_version\": \"4.26.0\"\n",
42686
+ "}\n",
42687
+ "\n",
42688
+ "Generate config GenerationConfig {\n",
42689
+ " \"decoder_start_token_id\": 0,\n",
42690
+ " \"eos_token_id\": 1,\n",
42691
+ " \"pad_token_id\": 0,\n",
42692
+ " \"transformers_version\": \"4.26.0\"\n",
42693
+ "}\n",
42694
+ "\n",
42695
+ "Generate config GenerationConfig {\n",
42696
+ " \"decoder_start_token_id\": 0,\n",
42697
+ " \"eos_token_id\": 1,\n",
42698
+ " \"pad_token_id\": 0,\n",
42699
+ " \"transformers_version\": \"4.26.0\"\n",
42700
+ "}\n",
42701
+ "\n",
42702
+ "Generate config GenerationConfig {\n",
42703
+ " \"decoder_start_token_id\": 0,\n",
42704
+ " \"eos_token_id\": 1,\n",
42705
+ " \"pad_token_id\": 0,\n",
42706
+ " \"transformers_version\": \"4.26.0\"\n",
42707
+ "}\n",
42708
+ "\n",
42709
+ "Generate config GenerationConfig {\n",
42710
+ " \"decoder_start_token_id\": 0,\n",
42711
+ " \"eos_token_id\": 1,\n",
42712
+ " \"pad_token_id\": 0,\n",
42713
+ " \"transformers_version\": \"4.26.0\"\n",
42714
+ "}\n",
42715
  "\n"
42716
  ]
42717
+ },
42718
+ {
42719
+ "name": "stderr",
42720
+ "output_type": "stream",
42721
+ "text": [
42722
+ "Generate config GenerationConfig {\n",
42723
+ " \"decoder_start_token_id\": 0,\n",
42724
+ " \"eos_token_id\": 1,\n",
42725
+ " \"pad_token_id\": 0,\n",
42726
+ " \"transformers_version\": \"4.26.0\"\n",
42727
+ "}\n",
42728
+ "\n",
42729
+ "Generate config GenerationConfig {\n",
42730
+ " \"decoder_start_token_id\": 0,\n",
42731
+ " \"eos_token_id\": 1,\n",
42732
+ " \"pad_token_id\": 0,\n",
42733
+ " \"transformers_version\": \"4.26.0\"\n",
42734
+ "}\n",
42735
+ "\n",
42736
+ "Generate config GenerationConfig {\n",
42737
+ " \"decoder_start_token_id\": 0,\n",
42738
+ " \"eos_token_id\": 1,\n",
42739
+ " \"pad_token_id\": 0,\n",
42740
+ " \"transformers_version\": \"4.26.0\"\n",
42741
+ "}\n",
42742
+ "\n",
42743
+ "Generate config GenerationConfig {\n",
42744
+ " \"decoder_start_token_id\": 0,\n",
42745
+ " \"eos_token_id\": 1,\n",
42746
+ " \"pad_token_id\": 0,\n",
42747
+ " \"transformers_version\": \"4.26.0\"\n",
42748
+ "}\n",
42749
+ "\n",
42750
+ "Generate config GenerationConfig {\n",
42751
+ " \"decoder_start_token_id\": 0,\n",
42752
+ " \"eos_token_id\": 1,\n",
42753
+ " \"pad_token_id\": 0,\n",
42754
+ " \"transformers_version\": \"4.26.0\"\n",
42755
+ "}\n",
42756
+ "\n",
42757
+ "Generate config GenerationConfig {\n",
42758
+ " \"decoder_start_token_id\": 0,\n",
42759
+ " \"eos_token_id\": 1,\n",
42760
+ " \"pad_token_id\": 0,\n",
42761
+ " \"transformers_version\": \"4.26.0\"\n",
42762
+ "}\n",
42763
+ "\n",
42764
+ "Generate config GenerationConfig {\n",
42765
+ " \"decoder_start_token_id\": 0,\n",
42766
+ " \"eos_token_id\": 1,\n",
42767
+ " \"pad_token_id\": 0,\n",
42768
+ " \"transformers_version\": \"4.26.0\"\n",
42769
+ "}\n",
42770
+ "\n",
42771
+ "Generate config GenerationConfig {\n",
42772
+ " \"decoder_start_token_id\": 0,\n",
42773
+ " \"eos_token_id\": 1,\n",
42774
+ " \"pad_token_id\": 0,\n",
42775
+ " \"transformers_version\": \"4.26.0\"\n",
42776
+ "}\n",
42777
+ "\n",
42778
+ "Generate config GenerationConfig {\n",
42779
+ " \"decoder_start_token_id\": 0,\n",
42780
+ " \"eos_token_id\": 1,\n",
42781
+ " \"pad_token_id\": 0,\n",
42782
+ " \"transformers_version\": \"4.26.0\"\n",
42783
+ "}\n",
42784
+ "\n",
42785
+ "Generate config GenerationConfig {\n",
42786
+ " \"decoder_start_token_id\": 0,\n",
42787
+ " \"eos_token_id\": 1,\n",
42788
+ " \"pad_token_id\": 0,\n",
42789
+ " \"transformers_version\": \"4.26.0\"\n",
42790
+ "}\n",
42791
+ "\n",
42792
+ "Generate config GenerationConfig {\n",
42793
+ " \"decoder_start_token_id\": 0,\n",
42794
+ " \"eos_token_id\": 1,\n",
42795
+ " \"pad_token_id\": 0,\n",
42796
+ " \"transformers_version\": \"4.26.0\"\n",
42797
+ "}\n",
42798
+ "\n",
42799
+ "Generate config GenerationConfig {\n",
42800
+ " \"decoder_start_token_id\": 0,\n",
42801
+ " \"eos_token_id\": 1,\n",
42802
+ " \"pad_token_id\": 0,\n",
42803
+ " \"transformers_version\": \"4.26.0\"\n",
42804
+ "}\n",
42805
+ "\n",
42806
+ "Generate config GenerationConfig {\n",
42807
+ " \"decoder_start_token_id\": 0,\n",
42808
+ " \"eos_token_id\": 1,\n",
42809
+ " \"pad_token_id\": 0,\n",
42810
+ " \"transformers_version\": \"4.26.0\"\n",
42811
+ "}\n",
42812
+ "\n",
42813
+ "Generate config GenerationConfig {\n",
42814
+ " \"decoder_start_token_id\": 0,\n",
42815
+ " \"eos_token_id\": 1,\n",
42816
+ " \"pad_token_id\": 0,\n",
42817
+ " \"transformers_version\": \"4.26.0\"\n",
42818
+ "}\n",
42819
+ "\n",
42820
+ "Generate config GenerationConfig {\n",
42821
+ " \"decoder_start_token_id\": 0,\n",
42822
+ " \"eos_token_id\": 1,\n",
42823
+ " \"pad_token_id\": 0,\n",
42824
+ " \"transformers_version\": \"4.26.0\"\n",
42825
+ "}\n",
42826
+ "\n",
42827
+ "Generate config GenerationConfig {\n",
42828
+ " \"decoder_start_token_id\": 0,\n",
42829
+ " \"eos_token_id\": 1,\n",
42830
+ " \"pad_token_id\": 0,\n",
42831
+ " \"transformers_version\": \"4.26.0\"\n",
42832
+ "}\n",
42833
+ "\n",
42834
+ "Generate config GenerationConfig {\n",
42835
+ " \"decoder_start_token_id\": 0,\n",
42836
+ " \"eos_token_id\": 1,\n",
42837
+ " \"pad_token_id\": 0,\n",
42838
+ " \"transformers_version\": \"4.26.0\"\n",
42839
+ "}\n",
42840
+ "\n",
42841
+ "Generate config GenerationConfig {\n",
42842
+ " \"decoder_start_token_id\": 0,\n",
42843
+ " \"eos_token_id\": 1,\n",
42844
+ " \"pad_token_id\": 0,\n",
42845
+ " \"transformers_version\": \"4.26.0\"\n",
42846
+ "}\n",
42847
+ "\n",
42848
+ "Generate config GenerationConfig {\n",
42849
+ " \"decoder_start_token_id\": 0,\n",
42850
+ " \"eos_token_id\": 1,\n",
42851
+ " \"pad_token_id\": 0,\n",
42852
+ " \"transformers_version\": \"4.26.0\"\n",
42853
+ "}\n",
42854
+ "\n",
42855
+ "Generate config GenerationConfig {\n",
42856
+ " \"decoder_start_token_id\": 0,\n",
42857
+ " \"eos_token_id\": 1,\n",
42858
+ " \"pad_token_id\": 0,\n",
42859
+ " \"transformers_version\": \"4.26.0\"\n",
42860
+ "}\n",
42861
+ "\n",
42862
+ "Generate config GenerationConfig {\n",
42863
+ " \"decoder_start_token_id\": 0,\n",
42864
+ " \"eos_token_id\": 1,\n",
42865
+ " \"pad_token_id\": 0,\n",
42866
+ " \"transformers_version\": \"4.26.0\"\n",
42867
+ "}\n",
42868
+ "\n",
42869
+ "Generate config GenerationConfig {\n",
42870
+ " \"decoder_start_token_id\": 0,\n",
42871
+ " \"eos_token_id\": 1,\n",
42872
+ " \"pad_token_id\": 0,\n",
42873
+ " \"transformers_version\": \"4.26.0\"\n",
42874
+ "}\n",
42875
+ "\n",
42876
+ "Generate config GenerationConfig {\n",
42877
+ " \"decoder_start_token_id\": 0,\n",
42878
+ " \"eos_token_id\": 1,\n",
42879
+ " \"pad_token_id\": 0,\n",
42880
+ " \"transformers_version\": \"4.26.0\"\n",
42881
+ "}\n",
42882
+ "\n",
42883
+ "Generate config GenerationConfig {\n",
42884
+ " \"decoder_start_token_id\": 0,\n",
42885
+ " \"eos_token_id\": 1,\n",
42886
+ " \"pad_token_id\": 0,\n",
42887
+ " \"transformers_version\": \"4.26.0\"\n",
42888
+ "}\n",
42889
+ "\n",
42890
+ "Generate config GenerationConfig {\n",
42891
+ " \"decoder_start_token_id\": 0,\n",
42892
+ " \"eos_token_id\": 1,\n",
42893
+ " \"pad_token_id\": 0,\n",
42894
+ " \"transformers_version\": \"4.26.0\"\n",
42895
+ "}\n",
42896
+ "\n",
42897
+ "Generate config GenerationConfig {\n",
42898
+ " \"decoder_start_token_id\": 0,\n",
42899
+ " \"eos_token_id\": 1,\n",
42900
+ " \"pad_token_id\": 0,\n",
42901
+ " \"transformers_version\": \"4.26.0\"\n",
42902
+ "}\n",
42903
+ "\n",
42904
+ "Generate config GenerationConfig {\n",
42905
+ " \"decoder_start_token_id\": 0,\n",
42906
+ " \"eos_token_id\": 1,\n",
42907
+ " \"pad_token_id\": 0,\n",
42908
+ " \"transformers_version\": \"4.26.0\"\n",
42909
+ "}\n",
42910
+ "\n",
42911
+ "Generate config GenerationConfig {\n",
42912
+ " \"decoder_start_token_id\": 0,\n",
42913
+ " \"eos_token_id\": 1,\n",
42914
+ " \"pad_token_id\": 0,\n",
42915
+ " \"transformers_version\": \"4.26.0\"\n",
42916
+ "}\n",
42917
+ "\n",
42918
+ "Saving model checkpoint to C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-20245\n",
42919
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-20245\\config.json\n",
42920
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-20245\\generation_config.json\n",
42921
+ "Model weights saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-20245\\pytorch_model.bin\n",
42922
+ "Deleting older checkpoint [C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-8098] due to args.save_total_limit\n",
42923
+ "\n",
42924
+ "\n",
42925
+ "Training completed. Do not forget to share your model on huggingface.co/models =)\n",
42926
+ "\n",
42927
+ "\n",
42928
+ "Loading best model from C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\checkpoint-20245 (score: 0.03408432751893997).\n",
42929
+ "Saving model checkpoint to C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\output\n",
42930
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\output\\config.json\n",
42931
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\output\\generation_config.json\n",
42932
+ "Model weights saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\output\\pytorch_model.bin\n",
42933
+ "Saving model checkpoint to C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\n",
42934
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\config.json\n",
42935
+ "Configuration saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\generation_config.json\n",
42936
+ "Model weights saved in C:\\Users\\vjmar\\Documents\\1. Code\\0. Work-Macro4\\T5Training\\pytorch_model.bin\n",
42937
+ "Several commits (2) will be pushed upstream.\n",
42938
+ "WARNING:huggingface_hub.repository:Several commits (2) will be pushed upstream.\n",
42939
+ "The progress bars may be unreliable.\n",
42940
+ "WARNING:huggingface_hub.repository:The progress bars may be unreliable.\n",
42941
+ "Upload file output/pytorch_model.bin: 0%| | 32.0k/850M [00:00<?, ?B/s]\n",
42942
+ "Upload file output/pytorch_model.bin: 91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 777M/850M [08:12<00:37, 2.08MB/s]\u001b[A"
42943
+ ]
42944
  }
42945
  ],
42946
  "source": [