End of training
Browse files- README.md +32 -32
- config.json +2 -2
- final_checkpoint/config.json +2 -2
- final_checkpoint/generation_config.json +1 -1
- final_checkpoint/model-00001-of-00004.safetensors +1 -1
- final_checkpoint/model-00002-of-00004.safetensors +1 -1
- final_checkpoint/model-00003-of-00004.safetensors +1 -1
- final_checkpoint/model-00004-of-00004.safetensors +1 -1
- generation_config.json +1 -1
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- tokenizer_config.json +5 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
license: llama3
|
3 |
-
base_model:
|
4 |
tags:
|
5 |
- trl
|
6 |
- dpo
|
@@ -15,17 +15,17 @@ should probably proofread and complete it, then remove this comment. -->
|
|
15 |
|
16 |
# MedQA_L3_1000steps_1e7rate_03beta_CSFTDPO
|
17 |
|
18 |
-
This model is a fine-tuned version of [
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
- Loss: 0.
|
21 |
-
- Rewards/chosen: 0.
|
22 |
-
- Rewards/rejected:
|
23 |
-
- Rewards/accuracies: 0.
|
24 |
-
- Rewards/margins: 0.
|
25 |
-
- Logps/rejected: -
|
26 |
-
- Logps/chosen: -
|
27 |
-
- Logits/rejected: -0.
|
28 |
-
- Logits/chosen: -0.
|
29 |
|
30 |
## Model description
|
31 |
|
@@ -59,31 +59,31 @@ The following hyperparameters were used during training:
|
|
59 |
|
60 |
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
61 |
|:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
62 |
-
| 0.
|
63 |
-
| 0.
|
64 |
-
| 0.
|
65 |
-
| 0.
|
66 |
-
| 0.
|
67 |
-
| 0.
|
68 |
-
| 0.
|
69 |
-
| 0.
|
70 |
-
| 0.
|
71 |
-
| 0.
|
72 |
-
| 0.
|
73 |
-
| 0.
|
74 |
-
| 0.
|
75 |
-
| 0.
|
76 |
-
| 0.
|
77 |
-
| 0.
|
78 |
-
| 0.
|
79 |
-
| 0.
|
80 |
-
| 0.
|
81 |
-
| 0.
|
82 |
|
83 |
|
84 |
### Framework versions
|
85 |
|
86 |
-
- Transformers 4.41.
|
87 |
- Pytorch 2.0.0+cu117
|
88 |
- Datasets 2.19.1
|
89 |
- Tokenizers 0.19.1
|
|
|
1 |
---
|
2 |
license: llama3
|
3 |
+
base_model: tsavage68/MedQA_L3_1000steps_1e6rate_SFT
|
4 |
tags:
|
5 |
- trl
|
6 |
- dpo
|
|
|
15 |
|
16 |
# MedQA_L3_1000steps_1e7rate_03beta_CSFTDPO
|
17 |
|
18 |
+
This model is a fine-tuned version of [tsavage68/MedQA_L3_1000steps_1e6rate_SFT](https://huggingface.co/tsavage68/MedQA_L3_1000steps_1e6rate_SFT) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 0.6020
|
21 |
+
- Rewards/chosen: 0.7087
|
22 |
+
- Rewards/rejected: 0.4830
|
23 |
+
- Rewards/accuracies: 0.7341
|
24 |
+
- Rewards/margins: 0.2257
|
25 |
+
- Logps/rejected: -32.2447
|
26 |
+
- Logps/chosen: -28.9661
|
27 |
+
- Logits/rejected: -0.7358
|
28 |
+
- Logits/chosen: -0.7350
|
29 |
|
30 |
## Model description
|
31 |
|
|
|
59 |
|
60 |
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
61 |
|:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
62 |
+
| 0.6925 | 0.0489 | 50 | 0.6930 | -0.0016 | -0.0023 | 0.5011 | 0.0007 | -33.8624 | -31.3338 | -0.7320 | -0.7314 |
|
63 |
+
| 0.6841 | 0.0977 | 100 | 0.6807 | 0.2459 | 0.2195 | 0.6549 | 0.0264 | -33.1233 | -30.5088 | -0.7330 | -0.7323 |
|
64 |
+
| 0.6562 | 0.1466 | 150 | 0.6641 | 0.3800 | 0.3137 | 0.6791 | 0.0663 | -32.8092 | -30.0619 | -0.7310 | -0.7303 |
|
65 |
+
| 0.6334 | 0.1954 | 200 | 0.6509 | 0.1334 | 0.0355 | 0.7165 | 0.0979 | -33.7366 | -30.8837 | -0.7311 | -0.7304 |
|
66 |
+
| 0.6544 | 0.2443 | 250 | 0.6415 | 0.2943 | 0.1754 | 0.7209 | 0.1189 | -33.2701 | -30.3474 | -0.7311 | -0.7303 |
|
67 |
+
| 0.6145 | 0.2931 | 300 | 0.6304 | 0.3548 | 0.2099 | 0.7385 | 0.1448 | -33.1550 | -30.1459 | -0.7317 | -0.7310 |
|
68 |
+
| 0.6171 | 0.3420 | 350 | 0.6223 | 0.4756 | 0.3093 | 0.7341 | 0.1663 | -32.8238 | -29.7432 | -0.7336 | -0.7328 |
|
69 |
+
| 0.5911 | 0.3908 | 400 | 0.6181 | 0.6387 | 0.4602 | 0.7121 | 0.1785 | -32.3208 | -29.1996 | -0.7334 | -0.7327 |
|
70 |
+
| 0.5942 | 0.4397 | 450 | 0.6129 | 0.6839 | 0.4904 | 0.7253 | 0.1935 | -32.2203 | -29.0489 | -0.7347 | -0.7339 |
|
71 |
+
| 0.6096 | 0.4885 | 500 | 0.6090 | 0.7785 | 0.5741 | 0.7297 | 0.2044 | -31.9411 | -28.7335 | -0.7351 | -0.7343 |
|
72 |
+
| 0.5671 | 0.5374 | 550 | 0.6068 | 0.7522 | 0.5395 | 0.7275 | 0.2127 | -32.0566 | -28.8212 | -0.7355 | -0.7347 |
|
73 |
+
| 0.6066 | 0.5862 | 600 | 0.6061 | 0.7215 | 0.5067 | 0.7209 | 0.2147 | -32.1657 | -28.9236 | -0.7356 | -0.7348 |
|
74 |
+
| 0.5816 | 0.6351 | 650 | 0.6046 | 0.6882 | 0.4692 | 0.7231 | 0.2191 | -32.2910 | -29.0344 | -0.7356 | -0.7348 |
|
75 |
+
| 0.5968 | 0.6839 | 700 | 0.6030 | 0.6956 | 0.4723 | 0.7451 | 0.2233 | -32.2804 | -29.0097 | -0.7352 | -0.7344 |
|
76 |
+
| 0.6132 | 0.7328 | 750 | 0.6042 | 0.7103 | 0.4891 | 0.7297 | 0.2212 | -32.2246 | -28.9608 | -0.7354 | -0.7346 |
|
77 |
+
| 0.6133 | 0.7816 | 800 | 0.6021 | 0.6956 | 0.4697 | 0.7407 | 0.2258 | -32.2890 | -29.0099 | -0.7358 | -0.7350 |
|
78 |
+
| 0.6397 | 0.8305 | 850 | 0.6029 | 0.7027 | 0.4791 | 0.7341 | 0.2236 | -32.2579 | -28.9862 | -0.7354 | -0.7346 |
|
79 |
+
| 0.6273 | 0.8793 | 900 | 0.6030 | 0.7126 | 0.4896 | 0.7341 | 0.2230 | -32.2229 | -28.9533 | -0.7356 | -0.7348 |
|
80 |
+
| 0.5996 | 0.9282 | 950 | 0.6019 | 0.7087 | 0.4830 | 0.7341 | 0.2257 | -32.2447 | -28.9661 | -0.7358 | -0.7350 |
|
81 |
+
| 0.5319 | 0.9770 | 1000 | 0.6020 | 0.7087 | 0.4830 | 0.7341 | 0.2257 | -32.2447 | -28.9661 | -0.7358 | -0.7350 |
|
82 |
|
83 |
|
84 |
### Framework versions
|
85 |
|
86 |
+
- Transformers 4.41.1
|
87 |
- Pytorch 2.0.0+cu117
|
88 |
- Datasets 2.19.1
|
89 |
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -23,7 +23,7 @@
|
|
23 |
"rope_theta": 500000.0,
|
24 |
"tie_word_embeddings": false,
|
25 |
"torch_dtype": "float16",
|
26 |
-
"transformers_version": "4.41.
|
27 |
"use_cache": false,
|
28 |
"vocab_size": 128256
|
29 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "tsavage68/MedQA_L3_1000steps_1e6rate_SFT",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
23 |
"rope_theta": 500000.0,
|
24 |
"tie_word_embeddings": false,
|
25 |
"torch_dtype": "float16",
|
26 |
+
"transformers_version": "4.41.1",
|
27 |
"use_cache": false,
|
28 |
"vocab_size": 128256
|
29 |
}
|
final_checkpoint/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -23,7 +23,7 @@
|
|
23 |
"rope_theta": 500000.0,
|
24 |
"tie_word_embeddings": false,
|
25 |
"torch_dtype": "float16",
|
26 |
-
"transformers_version": "4.41.
|
27 |
"use_cache": false,
|
28 |
"vocab_size": 128256
|
29 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "tsavage68/MedQA_L3_1000steps_1e6rate_SFT",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
23 |
"rope_theta": 500000.0,
|
24 |
"tie_word_embeddings": false,
|
25 |
"torch_dtype": "float16",
|
26 |
+
"transformers_version": "4.41.1",
|
27 |
"use_cache": false,
|
28 |
"vocab_size": 128256
|
29 |
}
|
final_checkpoint/generation_config.json
CHANGED
@@ -8,5 +8,5 @@
|
|
8 |
"max_length": 4096,
|
9 |
"temperature": 0.6,
|
10 |
"top_p": 0.9,
|
11 |
-
"transformers_version": "4.41.
|
12 |
}
|
|
|
8 |
"max_length": 4096,
|
9 |
"temperature": 0.6,
|
10 |
"top_p": 0.9,
|
11 |
+
"transformers_version": "4.41.1"
|
12 |
}
|
final_checkpoint/model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4976698592
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2557ef54b1cb541258b6aac809e2ec44f1a879c7ab1c84166f20074fd81e1c3
|
3 |
size 4976698592
|
final_checkpoint/model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999802616
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e119d7c88d94f0a6db7e39294810bd68b32ced63b3bd7464af0a063f1c32029
|
3 |
size 4999802616
|
final_checkpoint/model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4915916080
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9525481aee3b9d44cee6ffb20de7add9a26abb8c1ff8bea8a5e260cac6fc12c4
|
3 |
size 4915916080
|
final_checkpoint/model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1168138808
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02aa661fc4a8018b889fdbfbb145ea620dbfb2ef8508c18febc2e41b8cafde1e
|
3 |
size 1168138808
|
generation_config.json
CHANGED
@@ -8,5 +8,5 @@
|
|
8 |
"max_length": 4096,
|
9 |
"temperature": 0.6,
|
10 |
"top_p": 0.9,
|
11 |
-
"transformers_version": "4.41.
|
12 |
}
|
|
|
8 |
"max_length": 4096,
|
9 |
"temperature": 0.6,
|
10 |
"top_p": 0.9,
|
11 |
+
"transformers_version": "4.41.1"
|
12 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4976698592
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2557ef54b1cb541258b6aac809e2ec44f1a879c7ab1c84166f20074fd81e1c3
|
3 |
size 4976698592
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999802616
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e119d7c88d94f0a6db7e39294810bd68b32ced63b3bd7464af0a063f1c32029
|
3 |
size 4999802616
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4915916080
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9525481aee3b9d44cee6ffb20de7add9a26abb8c1ff8bea8a5e260cac6fc12c4
|
3 |
size 4915916080
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1168138808
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02aa661fc4a8018b889fdbfbb145ea620dbfb2ef8508c18febc2e41b8cafde1e
|
3 |
size 1168138808
|
tokenizer_config.json
CHANGED
@@ -2053,11 +2053,15 @@
|
|
2053 |
"chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
|
|
2056 |
"model_input_names": [
|
2057 |
"input_ids",
|
2058 |
"attention_mask"
|
2059 |
],
|
2060 |
"model_max_length": 1000000000000000019884624838656,
|
2061 |
"pad_token": "<|eot_id|>",
|
2062 |
-
"
|
|
|
|
|
|
|
2063 |
}
|
|
|
2053 |
"chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
2056 |
+
"max_length": 1024,
|
2057 |
"model_input_names": [
|
2058 |
"input_ids",
|
2059 |
"attention_mask"
|
2060 |
],
|
2061 |
"model_max_length": 1000000000000000019884624838656,
|
2062 |
"pad_token": "<|eot_id|>",
|
2063 |
+
"stride": 0,
|
2064 |
+
"tokenizer_class": "PreTrainedTokenizerFast",
|
2065 |
+
"truncation_side": "right",
|
2066 |
+
"truncation_strategy": "longest_first"
|
2067 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4667
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3c7fa4546475a34fdf342ed9f3840232fd6ebbaace04e86920503ddffcd0bbf4
|
3 |
size 4667
|