Training complete
Browse files- README.md +12 -24
- config.json +3 -3
- model.safetensors +1 -1
- runs/Jun04_23-27-50_eefaff9f758d/events.out.tfevents.1717543689.eefaff9f758d.4259.0 +3 -0
- sentencepiece.bpe.model +2 -2
- tokenizer.json +0 -0
- tokenizer_config.json +11 -10
- training_args.bin +1 -1
README.md
CHANGED
@@ -9,18 +9,6 @@ metrics:
|
|
9 |
model-index:
|
10 |
- name: relatives_psr
|
11 |
results: []
|
12 |
-
widget:
|
13 |
-
- text: >-
|
14 |
-
Rappelez-vous que la seule loi qui compte, la grande loi de la cité, la
|
15 |
-
grande loi de la République est celle qui est forgée grâce au débat et au
|
16 |
-
dissentiment, non à la pensée unique.
|
17 |
-
- text: >-
|
18 |
-
Le Parlement prendra toute sa place dans l'effort de redressement du pays, un Parlement qui mesure à sa juste valeur l'attente de l'opinion et qui, dans le cadre de la séparation des pouvoirs, poursuivra sa rénovation
|
19 |
-
language:
|
20 |
-
- fr
|
21 |
-
pipeline_tag: token-classification
|
22 |
-
datasets:
|
23 |
-
- djamina/relatives_psr
|
24 |
---
|
25 |
|
26 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -30,11 +18,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
30 |
|
31 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
32 |
It achieves the following results on the evaluation set:
|
33 |
-
- Loss: 0.
|
34 |
-
- Precision: 0.
|
35 |
-
- Recall: 0.
|
36 |
-
- F1: 0.
|
37 |
-
- Accuracy: 0.
|
38 |
|
39 |
## Model description
|
40 |
|
@@ -65,16 +53,16 @@ The following hyperparameters were used during training:
|
|
65 |
|
66 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
67 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
68 |
-
| No log | 1.0 |
|
69 |
-
| No log | 2.0 |
|
70 |
-
| No log | 3.0 |
|
71 |
-
| No log | 4.0 |
|
72 |
-
| No log | 5.0 |
|
73 |
|
74 |
|
75 |
### Framework versions
|
76 |
|
77 |
- Transformers 4.41.2
|
78 |
- Pytorch 2.3.0+cu121
|
79 |
-
- Datasets 2.19.
|
80 |
-
- Tokenizers 0.19.1
|
|
|
9 |
model-index:
|
10 |
- name: relatives_psr
|
11 |
results: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
---
|
13 |
|
14 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
18 |
|
19 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.0536
|
22 |
+
- Precision: 0.6640
|
23 |
+
- Recall: 0.6450
|
24 |
+
- F1: 0.6483
|
25 |
+
- Accuracy: 0.9804
|
26 |
|
27 |
## Model description
|
28 |
|
|
|
53 |
|
54 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
55 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
56 |
+
| No log | 1.0 | 49 | 0.0602 | 0.7398 | 0.6309 | 0.5692 | 0.9769 |
|
57 |
+
| No log | 2.0 | 98 | 0.0583 | 0.6821 | 0.6304 | 0.5952 | 0.9789 |
|
58 |
+
| No log | 3.0 | 147 | 0.0584 | 0.6604 | 0.6098 | 0.5610 | 0.9776 |
|
59 |
+
| No log | 4.0 | 196 | 0.0564 | 0.5868 | 0.6107 | 0.5981 | 0.9761 |
|
60 |
+
| No log | 5.0 | 245 | 0.0536 | 0.6640 | 0.6450 | 0.6483 | 0.9804 |
|
61 |
|
62 |
|
63 |
### Framework versions
|
64 |
|
65 |
- Transformers 4.41.2
|
66 |
- Pytorch 2.3.0+cu121
|
67 |
+
- Datasets 2.19.2
|
68 |
+
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -18,10 +18,10 @@
|
|
18 |
"initializer_range": 0.02,
|
19 |
"intermediate_size": 3072,
|
20 |
"label2id": {
|
21 |
-
"
|
22 |
-
"DET": 1,
|
23 |
"APPO": 2,
|
24 |
-
"
|
|
|
25 |
},
|
26 |
"layer_norm_eps": 1e-12,
|
27 |
"max_position_embeddings": 512,
|
|
|
18 |
"initializer_range": 0.02,
|
19 |
"intermediate_size": 3072,
|
20 |
"label2id": {
|
21 |
+
"AMBIGUE": 3,
|
|
|
22 |
"APPO": 2,
|
23 |
+
"DET": 1,
|
24 |
+
"O": 0
|
25 |
},
|
26 |
"layer_norm_eps": 1e-12,
|
27 |
"max_position_embeddings": 512,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1877839624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c47dc4f39cc656b60e56d0eea10078a6c4674cd5d982826530df2ade940ec43c
|
3 |
size 1877839624
|
runs/Jun04_23-27-50_eefaff9f758d/events.out.tfevents.1717543689.eefaff9f758d.4259.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c22fb096491073b9e0bdaa26f0e60ae488fd9607d29ed8b46d43bd7859594ceb
|
3 |
+
size 7574
|
sentencepiece.bpe.model
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f98f266fdc548c94216aaadc13ffaaafacf0c8793303e2195322d954549ea261
|
3 |
+
size 808767
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -24,7 +24,7 @@
|
|
24 |
"single_word": false,
|
25 |
"special": true
|
26 |
},
|
27 |
-
"
|
28 |
"content": "<unk>",
|
29 |
"lstrip": false,
|
30 |
"normalized": false,
|
@@ -32,6 +32,14 @@
|
|
32 |
"single_word": false,
|
33 |
"special": true
|
34 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
"5": {
|
36 |
"content": "<s>",
|
37 |
"lstrip": false,
|
@@ -55,14 +63,6 @@
|
|
55 |
"rstrip": false,
|
56 |
"single_word": false,
|
57 |
"special": true
|
58 |
-
},
|
59 |
-
"32005": {
|
60 |
-
"content": "<unk>NOTUSED",
|
61 |
-
"lstrip": false,
|
62 |
-
"normalized": false,
|
63 |
-
"rstrip": false,
|
64 |
-
"single_word": false,
|
65 |
-
"special": true
|
66 |
}
|
67 |
},
|
68 |
"additional_special_tokens": [
|
@@ -75,9 +75,10 @@
|
|
75 |
"cls_token": "<s>",
|
76 |
"eos_token": "</s>",
|
77 |
"mask_token": "<mask>",
|
78 |
-
"model_max_length":
|
79 |
"pad_token": "<pad>",
|
80 |
"sep_token": "</s>",
|
|
|
81 |
"tokenizer_class": "CamembertTokenizer",
|
82 |
"unk_token": "<unk>"
|
83 |
}
|
|
|
24 |
"single_word": false,
|
25 |
"special": true
|
26 |
},
|
27 |
+
"3": {
|
28 |
"content": "<unk>",
|
29 |
"lstrip": false,
|
30 |
"normalized": false,
|
|
|
32 |
"single_word": false,
|
33 |
"special": true
|
34 |
},
|
35 |
+
"4": {
|
36 |
+
"content": "<unk>NOTUSED",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
},
|
43 |
"5": {
|
44 |
"content": "<s>",
|
45 |
"lstrip": false,
|
|
|
63 |
"rstrip": false,
|
64 |
"single_word": false,
|
65 |
"special": true
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
}
|
67 |
},
|
68 |
"additional_special_tokens": [
|
|
|
75 |
"cls_token": "<s>",
|
76 |
"eos_token": "</s>",
|
77 |
"mask_token": "<mask>",
|
78 |
+
"model_max_length": 1000000000000000019884624838656,
|
79 |
"pad_token": "<pad>",
|
80 |
"sep_token": "</s>",
|
81 |
+
"sp_model_kwargs": {},
|
82 |
"tokenizer_class": "CamembertTokenizer",
|
83 |
"unk_token": "<unk>"
|
84 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd2109481ecfe42e2134e8db5c5cc8b16f0d22287b04bb90d0095e1a660e425c
|
3 |
size 5112
|