santyzenith commited on
Commit
d09282d
1 Parent(s): 02f3134

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "santyzenith/augmented_t5_pictos",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "flax-community/spanish-t5-small",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7f33493add5f761ccad62d11eab058975dedf81b122115224d6aaa0e9f21d2e
3
  size 242020886
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:038438ae865e14d40a9c34e5fc08c413e43b9d0243639464a187a81c6b315300
3
  size 242020886
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
tokenizer.json CHANGED
@@ -1,7 +1,19 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 32,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": "BatchLongest",
11
+ "direction": "Right",
12
+ "pad_to_multiple_of": null,
13
+ "pad_id": 0,
14
+ "pad_type_id": 0,
15
+ "pad_token": "<pad>"
16
+ },
17
  "added_tokens": [
18
  {
19
  "id": 0,
tokenizer_config.json CHANGED
@@ -930,15 +930,8 @@
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
- "max_length": 32,
934
  "model_max_length": 1000000000000000019884624838656,
935
- "pad_to_multiple_of": null,
936
  "pad_token": "<pad>",
937
- "pad_token_type_id": 0,
938
- "padding_side": "right",
939
- "stride": 0,
940
  "tokenizer_class": "T5Tokenizer",
941
- "truncation_side": "right",
942
- "truncation_strategy": "longest_first",
943
  "unk_token": "<unk>"
944
  }
 
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "model_max_length": 1000000000000000019884624838656,
 
934
  "pad_token": "<pad>",
 
 
 
935
  "tokenizer_class": "T5Tokenizer",
 
 
936
  "unk_token": "<unk>"
937
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7828fc709c44abee872aa93ead6134fc99aa451204bdb7b02ab3a86e08d0d572
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba527213b6e5900112f854d317602df72ab9b87e56d8e3c7ae4734cac7753d6c
3
  size 4664