Training in progress, epoch 1
Browse files- config.json +16 -16
- logs/events.out.tfevents.1731515433.cmc-Z790-AORUS-ELITE-DDR4.209585.5 +2 -2
- logs/events.out.tfevents.1731515550.cmc-Z790-AORUS-ELITE-DDR4.209585.6 +3 -0
- logs/events.out.tfevents.1731515779.cmc-Z790-AORUS-ELITE-DDR4.212518.0 +3 -0
- model.safetensors +1 -1
- tokenizer.json +16 -2
- training_args.bin +1 -1
config.json
CHANGED
@@ -10,30 +10,30 @@
|
|
10 |
"hidden_dropout_prob": 0.1,
|
11 |
"hidden_size": 768,
|
12 |
"id2label": {
|
13 |
-
"0": "
|
14 |
"1": "Precaution",
|
15 |
-
"2": "
|
16 |
-
"3": "
|
17 |
-
"4": "Medical
|
18 |
-
"5": "
|
19 |
-
"6": "
|
20 |
-
"7": "
|
21 |
"8": "Treatment Plan",
|
22 |
-
"9": "
|
23 |
},
|
24 |
"initializer_range": 0.02,
|
25 |
"intermediate_size": 3072,
|
26 |
"label2id": {
|
27 |
-
"Cause Analysis":
|
28 |
-
"Diagnosis":
|
29 |
-
"Disease Description":
|
30 |
-
"Effect":
|
31 |
-
"Medical Advice":
|
32 |
-
"Medical Knowledge":
|
33 |
"Precaution": 1,
|
34 |
-
"Result Description":
|
35 |
"Treatment Plan": 8,
|
36 |
-
"Unknown":
|
37 |
},
|
38 |
"layer_norm_eps": 1e-12,
|
39 |
"max_position_embeddings": 512,
|
|
|
10 |
"hidden_dropout_prob": 0.1,
|
11 |
"hidden_size": 768,
|
12 |
"id2label": {
|
13 |
+
"0": "Disease Description",
|
14 |
"1": "Precaution",
|
15 |
+
"2": "Diagnosis",
|
16 |
+
"3": "Medical Advice",
|
17 |
+
"4": "Medical Knowledge",
|
18 |
+
"5": "Cause Analysis",
|
19 |
+
"6": "Result Description",
|
20 |
+
"7": "Unknown",
|
21 |
"8": "Treatment Plan",
|
22 |
+
"9": "Effect"
|
23 |
},
|
24 |
"initializer_range": 0.02,
|
25 |
"intermediate_size": 3072,
|
26 |
"label2id": {
|
27 |
+
"Cause Analysis": 5,
|
28 |
+
"Diagnosis": 2,
|
29 |
+
"Disease Description": 0,
|
30 |
+
"Effect": 9,
|
31 |
+
"Medical Advice": 3,
|
32 |
+
"Medical Knowledge": 4,
|
33 |
"Precaution": 1,
|
34 |
+
"Result Description": 6,
|
35 |
"Treatment Plan": 8,
|
36 |
+
"Unknown": 7
|
37 |
},
|
38 |
"layer_norm_eps": 1e-12,
|
39 |
"max_position_embeddings": 512,
|
logs/events.out.tfevents.1731515433.cmc-Z790-AORUS-ELITE-DDR4.209585.5
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf2baccede39ad120a87c5f7fc44eb06f0f19217c34d4d39836df35789725b70
|
3 |
+
size 6354
|
logs/events.out.tfevents.1731515550.cmc-Z790-AORUS-ELITE-DDR4.209585.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d5e640f5b1c6d45c8a32003d7118edd4eb1c110275bed984cb8c1cfa1c03e79
|
3 |
+
size 5616
|
logs/events.out.tfevents.1731515779.cmc-Z790-AORUS-ELITE-DDR4.212518.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8dc1e4818e826b8dcd0d7a8f0aa4a12b0e98edcf8963ef7bdaf7bbcd88ccec4d
|
3 |
+
size 16895
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 437983256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9758181b3d079b3fd2e6be27655aacacf0c79210f60d8c70c834ad547238493a
|
3 |
size 437983256
|
tokenizer.json
CHANGED
@@ -1,7 +1,21 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
"added_tokens": [
|
6 |
{
|
7 |
"id": 0,
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": {
|
4 |
+
"direction": "Right",
|
5 |
+
"max_length": 512,
|
6 |
+
"strategy": "LongestFirst",
|
7 |
+
"stride": 0
|
8 |
+
},
|
9 |
+
"padding": {
|
10 |
+
"strategy": {
|
11 |
+
"Fixed": 512
|
12 |
+
},
|
13 |
+
"direction": "Right",
|
14 |
+
"pad_to_multiple_of": null,
|
15 |
+
"pad_id": 0,
|
16 |
+
"pad_type_id": 0,
|
17 |
+
"pad_token": "[PAD]"
|
18 |
+
},
|
19 |
"added_tokens": [
|
20 |
{
|
21 |
"id": 0,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5304
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a71f3681e5c2a1b82627babe6f6d96e612883ba45bec189a74b8a61dd349556
|
3 |
size 5304
|