owanr commited on
Commit
1136def
1 Parent(s): 11a882d

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [google/t5-v1_1-xl](https://huggingface.co/google/t5-v1_1-xl) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 7.5273
19
 
20
  ## Model description
21
 
@@ -46,15 +46,15 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 8.1957 | 1.0 | 125 | 7.5273 |
50
- | 8.6066 | 2.0 | 250 | 7.5273 |
51
- | 8.5176 | 3.0 | 375 | 7.5273 |
52
- | 8.1945 | 4.0 | 500 | 7.5273 |
53
 
54
 
55
  ### Framework versions
56
 
57
- - Transformers 4.35.2
58
- - Pytorch 2.1.1+cu121
59
- - Datasets 2.15.0
60
- - Tokenizers 0.15.0
 
15
 
16
  This model is a fine-tuned version of [google/t5-v1_1-xl](https://huggingface.co/google/t5-v1_1-xl) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: nan
19
 
20
  ## Model description
21
 
 
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
+ | 7.8207 | 1.0 | 125 | 7.625 |
50
+ | 0.0 | 2.0 | 250 | nan |
51
+ | 0.0 | 3.0 | 375 | nan |
52
+ | 0.0 | 4.0 | 500 | nan |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.34.0
58
+ - Pytorch 2.1.0+cu121
59
+ - Datasets 2.6.1
60
+ - Tokenizers 0.14.1
adapter_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "google/t5-v1_1-xl",
5
  "bias": "none",
@@ -13,7 +12,6 @@
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 8,
16
- "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
  "q",
 
1
  {
 
2
  "auto_mapping": null,
3
  "base_model_name_or_path": "google/t5-v1_1-xl",
4
  "bias": "none",
 
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 8,
 
15
  "revision": null,
16
  "target_modules": [
17
  "q",
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ad04c687a553091a026d182726c92eb7ee3bb2b2c7b98b13b219d2370419c76
3
  size 18980874
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0fa4210778510d648ad8a1f128794c74ce6b32d3212567c0f0fa25cbe6f9872
3
  size 18980874
added_tokens.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "<extra_id_0>": 32099,
3
  "<extra_id_10>": 32089,
4
  "<extra_id_11>": 32088,
@@ -98,5 +100,6 @@
98
  "<extra_id_97>": 32002,
99
  "<extra_id_98>": 32001,
100
  "<extra_id_99>": 32000,
101
- "<extra_id_9>": 32090
 
102
  }
 
1
  {
2
+ "0": 632,
3
+ "1": 536,
4
  "<extra_id_0>": 32099,
5
  "<extra_id_10>": 32089,
6
  "<extra_id_11>": 32088,
 
100
  "<extra_id_97>": 32002,
101
  "<extra_id_98>": 32001,
102
  "<extra_id_99>": 32000,
103
+ "<extra_id_9>": 32090,
104
+ "<unk>": 2
105
  }
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "1",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "0",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "1",
105
+ "pad_token": "0",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
tokenizer_config.json CHANGED
@@ -2,25 +2,25 @@
2
  "added_tokens_decoder": {
3
  "2": {
4
  "content": "<unk>",
5
- "lstrip": false,
6
  "normalized": false,
7
- "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
  "536": {
12
  "content": "1",
13
- "lstrip": false,
14
  "normalized": false,
15
- "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
  "632": {
20
  "content": "0",
21
- "lstrip": false,
22
  "normalized": false,
23
- "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
@@ -937,5 +937,6 @@
937
  "pad_token": "0",
938
  "sp_model_kwargs": {},
939
  "tokenizer_class": "T5Tokenizer",
 
940
  "unk_token": "<unk>"
941
  }
 
2
  "added_tokens_decoder": {
3
  "2": {
4
  "content": "<unk>",
5
+ "lstrip": true,
6
  "normalized": false,
7
+ "rstrip": true,
8
  "single_word": false,
9
  "special": true
10
  },
11
  "536": {
12
  "content": "1",
13
+ "lstrip": true,
14
  "normalized": false,
15
+ "rstrip": true,
16
  "single_word": false,
17
  "special": true
18
  },
19
  "632": {
20
  "content": "0",
21
+ "lstrip": true,
22
  "normalized": false,
23
+ "rstrip": true,
24
  "single_word": false,
25
  "special": true
26
  },
 
937
  "pad_token": "0",
938
  "sp_model_kwargs": {},
939
  "tokenizer_class": "T5Tokenizer",
940
+ "tokenizer_file": null,
941
  "unk_token": "<unk>"
942
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e9866dbdd5346acbb52aaca177c24a0f13e3e2790fd8565cb91bec90a8ee597
3
- size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a299ddc34167072848ac32b7ff45fb98f4be270c3f89bb6b2515c29e4c61907
3
+ size 4728