kingsleykim
commited on
End of training
Browse files- README.md +4 -4
- generation_config.json +2 -3
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- tokenizer_config.json +2 -2
- training_args.bin +2 -2
README.md
CHANGED
@@ -33,14 +33,14 @@ More information needed
|
|
33 |
The following hyperparameters were used during training:
|
34 |
- learning_rate: 2e-05
|
35 |
- train_batch_size: 2
|
36 |
-
- eval_batch_size:
|
37 |
- seed: 42
|
38 |
-
- gradient_accumulation_steps:
|
39 |
-
- total_train_batch_size:
|
40 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
41 |
- lr_scheduler_type: cosine
|
42 |
- lr_scheduler_warmup_ratio: 0.1
|
43 |
-
- num_epochs:
|
44 |
|
45 |
### Training results
|
46 |
|
|
|
33 |
The following hyperparameters were used during training:
|
34 |
- learning_rate: 2e-05
|
35 |
- train_batch_size: 2
|
36 |
+
- eval_batch_size: 4
|
37 |
- seed: 42
|
38 |
+
- gradient_accumulation_steps: 8
|
39 |
+
- total_train_batch_size: 16
|
40 |
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
41 |
- lr_scheduler_type: cosine
|
42 |
- lr_scheduler_warmup_ratio: 0.1
|
43 |
+
- num_epochs: 4
|
44 |
|
45 |
### Training results
|
46 |
|
generation_config.json
CHANGED
@@ -6,8 +6,7 @@
|
|
6 |
151643
|
7 |
],
|
8 |
"pad_token_id": 151643,
|
9 |
-
"temperature": 0.
|
10 |
-
"
|
11 |
-
"top_p": 0.001,
|
12 |
"transformers_version": "4.46.0"
|
13 |
}
|
|
|
6 |
151643
|
7 |
],
|
8 |
"pad_token_id": 151643,
|
9 |
+
"temperature": 0.8,
|
10 |
+
"top_p": 0.95,
|
|
|
11 |
"transformers_version": "4.46.0"
|
12 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4966659944
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10e7c7cebc1a39694902f8b478d0f8d9081130b5397aacd1e568b6fbf9b3f01c
|
3 |
size 4966659944
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4991495816
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38bb9e57c8e07d5d1750f39bdc822e9d45460737b3d9b35cf9640560604b29f8
|
3 |
size 4991495816
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4932751040
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a38e53ab9b1f5f501c40160d89d8b407e14771fa89354ab4ee0f4bd7bf7b2740
|
3 |
size 4932751040
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1691924384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fabc46bfd3157ce05711647ddbba2d86082f304bfa38605e1397745b46ba6bcf
|
3 |
size 1691924384
|
tokenizer_config.json
CHANGED
@@ -137,8 +137,8 @@
|
|
137 |
"model_max_length": 32768,
|
138 |
"pad_token": "<|endoftext|>",
|
139 |
"padding_side": "right",
|
140 |
-
"processor_class": "Qwen2VLProcessor",
|
141 |
"split_special_tokens": false,
|
142 |
"tokenizer_class": "Qwen2Tokenizer",
|
143 |
-
"unk_token": null
|
|
|
144 |
}
|
|
|
137 |
"model_max_length": 32768,
|
138 |
"pad_token": "<|endoftext|>",
|
139 |
"padding_side": "right",
|
|
|
140 |
"split_special_tokens": false,
|
141 |
"tokenizer_class": "Qwen2Tokenizer",
|
142 |
+
"unk_token": null,
|
143 |
+
"use_fast": true
|
144 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d231ba54507ce4fa6a6b4f166310e855afbecbd95170a261d41b7d4031caab89
|
3 |
+
size 5240
|