Rodrigo1771 commited on
Commit
cbb688f
β€’
1 Parent(s): b08780a

End of training

Browse files
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: PlanTL-GOB-ES/bsc-bio-ehr-es
4
+ tags:
5
+ - token-classification
6
+ - generated_from_trainer
7
+ datasets:
8
+ - Rodrigo1771/drugtemist-ner
9
+ model-index:
10
+ - name: output
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # output
18
+
19
+ This model is a fine-tuned version of [PlanTL-GOB-ES/bsc-bio-ehr-es](https://huggingface.co/PlanTL-GOB-ES/bsc-bio-ehr-es) on the Rodrigo1771/drugtemist-ner dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - eval_loss: 1.2930
22
+ - eval_precision: 0.0028
23
+ - eval_recall: 0.2693
24
+ - eval_f1: 0.0056
25
+ - eval_accuracy: 0.0288
26
+ - eval_runtime: 16.5731
27
+ - eval_samples_per_second: 410.727
28
+ - eval_steps_per_second: 51.348
29
+ - step: 0
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+
43
+ ## Training procedure
44
+
45
+ ### Training hyperparameters
46
+
47
+ The following hyperparameters were used during training:
48
+ - learning_rate: 5e-05
49
+ - train_batch_size: 4
50
+ - eval_batch_size: 8
51
+ - seed: 42
52
+ - gradient_accumulation_steps: 4
53
+ - total_train_batch_size: 16
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: linear
56
+ - num_epochs: 10.0
57
+
58
+ ### Framework versions
59
+
60
+ - Transformers 4.40.2
61
+ - Pytorch 2.2.1+cu121
62
+ - Datasets 2.19.1
63
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.028794858943639246,
3
+ "eval_f1": 0.00559299062744574,
4
+ "eval_loss": 1.2929713726043701,
5
+ "eval_precision": 0.0028258395540381536,
6
+ "eval_recall": 0.2693014705882353,
7
+ "eval_runtime": 16.5731,
8
+ "eval_samples": 6807,
9
+ "eval_samples_per_second": 410.727,
10
+ "eval_steps_per_second": 51.348,
11
+ "predict_accuracy": 0.028794858943639246,
12
+ "predict_f1": 0.00559299062744574,
13
+ "predict_loss": 1.2929713726043701,
14
+ "predict_precision": 0.0028258395540381536,
15
+ "predict_recall": 0.2693014705882353,
16
+ "predict_runtime": 15.7694,
17
+ "predict_samples_per_second": 431.659,
18
+ "predict_steps_per_second": 53.965
19
+ }
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "PlanTL-GOB-ES/bsc-bio-ehr-es",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "finetuning_task": "ner",
11
+ "gradient_checkpointing": false,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 768,
15
+ "id2label": {
16
+ "0": "O",
17
+ "1": "B-FARMACO",
18
+ "2": "I-FARMACO"
19
+ },
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "B-FARMACO": 1,
24
+ "I-FARMACO": 2,
25
+ "O": 0
26
+ },
27
+ "layer_norm_eps": 1e-05,
28
+ "max_position_embeddings": 514,
29
+ "model_type": "roberta",
30
+ "num_attention_heads": 12,
31
+ "num_hidden_layers": 12,
32
+ "pad_token_id": 1,
33
+ "position_embedding_type": "absolute",
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.40.2",
36
+ "type_vocab_size": 1,
37
+ "use_cache": true,
38
+ "vocab_size": 50262
39
+ }
eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_accuracy": 0.028794858943639246,
3
+ "eval_f1": 0.00559299062744574,
4
+ "eval_loss": 1.2929713726043701,
5
+ "eval_precision": 0.0028258395540381536,
6
+ "eval_recall": 0.2693014705882353,
7
+ "eval_runtime": 16.5731,
8
+ "eval_samples": 6807,
9
+ "eval_samples_per_second": 410.727,
10
+ "eval_steps_per_second": 51.348
11
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d6afec433b240e395bd3f2e246116229e0d675eb0ae73c1d584f9ad8e27ea20
3
+ size 496244100
predict_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "predict_accuracy": 0.028794858943639246,
3
+ "predict_f1": 0.00559299062744574,
4
+ "predict_loss": 1.2929713726043701,
5
+ "predict_precision": 0.0028258395540381536,
6
+ "predict_recall": 0.2693014705882353,
7
+ "predict_runtime": 15.7694,
8
+ "predict_samples_per_second": 431.659,
9
+ "predict_steps_per_second": 53.965
10
+ }
predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tb/events.out.tfevents.1715783401.61af03e56d14.4488.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ab9da2ae3e0d6f1547437ff1ca2dccde86482093f826a321df49f7924fc92ce
3
+ size 486
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50261": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "mask_token": "<mask>",
51
+ "max_len": 512,
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
58
+ }
train.log ADDED
@@ -0,0 +1,343 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/851 [00:00<?, ?it/s]
1
  1%| | 6/851 [00:00<00:14, 56.94it/s]
2
  2%|▏ | 14/851 [00:00<00:12, 67.06it/s]
3
  3%|β–Ž | 22/851 [00:00<00:11, 69.14it/s]
4
  4%|β–Ž | 30/851 [00:00<00:11, 70.86it/s]
5
  4%|▍ | 38/851 [00:00<00:11, 72.79it/s]
6
  5%|β–Œ | 46/851 [00:00<00:10, 74.66it/s]
7
  6%|β–‹ | 54/851 [00:00<00:10, 75.69it/s]
8
  7%|β–‹ | 62/851 [00:00<00:10, 76.61it/s]
9
  8%|β–Š | 70/851 [00:00<00:10, 71.80it/s]
10
  9%|β–‰ | 78/851 [00:01<00:10, 72.55it/s]
11
  10%|β–ˆ | 86/851 [00:01<00:10, 73.09it/s]
12
  11%|β–ˆ | 94/851 [00:01<00:10, 71.17it/s]
13
  12%|β–ˆβ– | 102/851 [00:01<00:10, 72.61it/s]
14
  13%|β–ˆβ–Ž | 110/851 [00:01<00:10, 71.20it/s]
15
  14%|β–ˆβ– | 118/851 [00:01<00:10, 72.02it/s]
16
  15%|β–ˆβ– | 126/851 [00:01<00:10, 68.64it/s]
17
  16%|β–ˆβ–Œ | 134/851 [00:01<00:10, 68.76it/s]
18
  17%|β–ˆβ–‹ | 142/851 [00:01<00:10, 69.40it/s]
19
  18%|β–ˆβ–Š | 149/851 [00:02<00:10, 67.95it/s]
20
  18%|β–ˆβ–Š | 157/851 [00:02<00:09, 70.23it/s]
21
  19%|β–ˆβ–‰ | 165/851 [00:02<00:09, 71.18it/s]
22
  20%|β–ˆβ–ˆ | 173/851 [00:02<00:09, 72.70it/s]
23
  21%|β–ˆβ–ˆβ– | 181/851 [00:02<00:09, 73.97it/s]
24
  22%|β–ˆβ–ˆβ– | 189/851 [00:02<00:08, 75.32it/s]
25
  23%|β–ˆβ–ˆβ–Ž | 197/851 [00:02<00:08, 75.83it/s]
26
  24%|β–ˆβ–ˆβ– | 206/851 [00:02<00:08, 76.15it/s]
27
  25%|β–ˆβ–ˆβ–Œ | 214/851 [00:02<00:08, 71.48it/s]
28
  26%|β–ˆβ–ˆβ–Œ | 222/851 [00:03<00:08, 72.75it/s]
29
  27%|β–ˆβ–ˆβ–‹ | 230/851 [00:03<00:08, 73.89it/s]
30
  28%|β–ˆβ–ˆβ–Š | 238/851 [00:03<00:08, 72.13it/s]
31
  29%|β–ˆβ–ˆβ–‰ | 246/851 [00:03<00:08, 70.45it/s]
32
  30%|β–ˆβ–ˆβ–‰ | 255/851 [00:03<00:08, 73.88it/s]
33
  31%|β–ˆβ–ˆβ–ˆ | 264/851 [00:03<00:07, 76.63it/s]
34
  32%|β–ˆβ–ˆβ–ˆβ– | 272/851 [00:03<00:07, 74.50it/s]
35
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 281/851 [00:03<00:07, 77.22it/s]
36
  34%|β–ˆβ–ˆβ–ˆβ– | 289/851 [00:03<00:07, 74.33it/s]
37
  35%|β–ˆβ–ˆβ–ˆβ–Œ | 298/851 [00:04<00:07, 76.25it/s]
38
  36%|β–ˆβ–ˆβ–ˆβ–Œ | 307/851 [00:04<00:06, 78.12it/s]
39
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 315/851 [00:04<00:07, 73.67it/s]
40
  38%|β–ˆβ–ˆβ–ˆβ–Š | 324/851 [00:04<00:07, 74.73it/s]
41
  39%|β–ˆβ–ˆβ–ˆβ–‰ | 332/851 [00:04<00:07, 73.75it/s]
42
  40%|β–ˆβ–ˆβ–ˆβ–‰ | 340/851 [00:04<00:06, 74.75it/s]
43
  41%|β–ˆβ–ˆβ–ˆβ–ˆ | 348/851 [00:04<00:06, 75.70it/s]
44
  42%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 356/851 [00:04<00:06, 72.87it/s]
45
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 364/851 [00:04<00:06, 73.10it/s]
46
  44%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 372/851 [00:05<00:06, 71.82it/s]
47
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 380/851 [00:05<00:06, 69.17it/s]
48
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 388/851 [00:05<00:06, 72.06it/s]
49
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 396/851 [00:05<00:06, 72.74it/s]
50
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 404/851 [00:05<00:06, 68.56it/s]
51
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 412/851 [00:05<00:06, 69.70it/s]
52
  49%|β–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 420/851 [00:05<00:05, 72.48it/s]
53
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 428/851 [00:05<00:05, 70.96it/s]
54
  51%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 436/851 [00:06<00:05, 72.47it/s]
55
  52%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 444/851 [00:06<00:05, 74.07it/s]
56
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 452/851 [00:06<00:05, 75.55it/s]
57
  54%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 460/851 [00:06<00:05, 74.33it/s]
58
  55%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 468/851 [00:06<00:05, 71.78it/s]
59
  56%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 476/851 [00:06<00:05, 67.46it/s]
60
  57%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 483/851 [00:06<00:05, 67.69it/s]
61
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 491/851 [00:06<00:05, 69.09it/s]
62
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 500/851 [00:06<00:04, 73.77it/s]
63
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 508/851 [00:07<00:04, 72.84it/s]
64
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 517/851 [00:07<00:04, 74.44it/s]
65
  62%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 525/851 [00:07<00:04, 70.40it/s]
66
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 533/851 [00:07<00:04, 71.47it/s]
67
  64%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 541/851 [00:07<00:04, 73.42it/s]
68
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 549/851 [00:07<00:04, 71.38it/s]
69
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 557/851 [00:07<00:04, 73.46it/s]
70
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 566/851 [00:07<00:03, 76.65it/s]
71
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 574/851 [00:07<00:03, 76.97it/s]
72
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 582/851 [00:08<00:03, 74.62it/s]
73
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 590/851 [00:08<00:03, 71.50it/s]
74
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 598/851 [00:08<00:03, 72.36it/s]
75
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 606/851 [00:08<00:03, 70.56it/s]
76
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 614/851 [00:08<00:03, 69.62it/s]
77
  73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 622/851 [00:08<00:03, 71.02it/s]
78
  74%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 630/851 [00:08<00:03, 67.52it/s]
79
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 638/851 [00:08<00:03, 70.60it/s]
80
  76%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 646/851 [00:08<00:03, 66.60it/s]
81
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 654/851 [00:09<00:02, 69.98it/s]
82
  78%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 662/851 [00:09<00:02, 70.75it/s]
83
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 670/851 [00:09<00:02, 72.07it/s]
84
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 678/851 [00:09<00:02, 70.99it/s]
85
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 686/851 [00:09<00:02, 72.75it/s]
86
  82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 695/851 [00:09<00:02, 75.21it/s]
87
  83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 704/851 [00:09<00:01, 76.96it/s]
88
  84%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 713/851 [00:09<00:01, 78.47it/s]
89
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 721/851 [00:09<00:01, 76.12it/s]
90
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 730/851 [00:10<00:01, 78.39it/s]
91
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 739/851 [00:10<00:01, 78.90it/s]
92
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 747/851 [00:10<00:01, 78.42it/s]
93
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 755/851 [00:10<00:01, 78.48it/s]
94
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 764/851 [00:10<00:01, 78.64it/s]
95
  91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 772/851 [00:10<00:01, 74.92it/s]
96
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 780/851 [00:10<00:00, 72.75it/s]
97
  93%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 788/851 [00:10<00:00, 73.39it/s]
98
  94%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 796/851 [00:10<00:00, 75.11it/s]
99
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 805/851 [00:11<00:00, 76.84it/s]
100
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 813/851 [00:11<00:00, 73.85it/s]
101
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 821/851 [00:11<00:00, 74.79it/s]
102
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 829/851 [00:11<00:00, 75.96it/s]
103
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 837/851 [00:11<00:00, 74.64it/s]
104
  99%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 845/851 [00:11<00:00, 69.55it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
105
  0%| | 0/851 [00:00<?, ?it/s]
106
  1%| | 10/851 [00:00<00:08, 95.20it/s]
107
  2%|▏ | 20/851 [00:00<00:10, 81.67it/s]
108
  3%|β–Ž | 29/851 [00:00<00:10, 78.36it/s]
109
  4%|▍ | 37/851 [00:00<00:10, 76.45it/s]
110
  5%|β–Œ | 45/851 [00:00<00:10, 76.74it/s]
111
  6%|β–Œ | 53/851 [00:00<00:10, 77.26it/s]
112
  7%|β–‹ | 62/851 [00:00<00:10, 77.44it/s]
113
  8%|β–Š | 70/851 [00:00<00:10, 72.24it/s]
114
  9%|β–‰ | 78/851 [00:01<00:10, 73.99it/s]
115
  10%|β–ˆ | 86/851 [00:01<00:10, 73.73it/s]
116
  11%|β–ˆ | 94/851 [00:01<00:10, 72.12it/s]
117
  12%|β–ˆβ– | 102/851 [00:01<00:10, 73.14it/s]
118
  13%|β–ˆβ–Ž | 110/851 [00:01<00:10, 71.41it/s]
119
  14%|β–ˆβ– | 118/851 [00:01<00:10, 72.63it/s]
120
  15%|β–ˆβ– | 126/851 [00:01<00:10, 69.23it/s]
121
  16%|β–ˆβ–Œ | 133/851 [00:01<00:10, 69.22it/s]
122
  16%|β–ˆβ–‹ | 140/851 [00:01<00:10, 69.15it/s]
123
  17%|β–ˆβ–‹ | 148/851 [00:02<00:10, 68.62it/s]
124
  18%|β–ˆβ–Š | 156/851 [00:02<00:09, 70.75it/s]
125
  19%|β–ˆβ–‰ | 164/851 [00:02<00:09, 71.62it/s]
126
  20%|β–ˆβ–ˆ | 172/851 [00:02<00:09, 73.75it/s]
127
  21%|β–ˆβ–ˆ | 180/851 [00:02<00:09, 74.25it/s]
128
  22%|β–ˆβ–ˆβ– | 188/851 [00:02<00:08, 74.90it/s]
129
  23%|β–ˆβ–ˆβ–Ž | 196/851 [00:02<00:08, 76.03it/s]
130
  24%|β–ˆβ–ˆβ– | 205/851 [00:02<00:08, 77.93it/s]
131
  25%|β–ˆβ–ˆβ–Œ | 213/851 [00:02<00:08, 72.81it/s]
132
  26%|β–ˆβ–ˆβ–Œ | 221/851 [00:03<00:08, 71.31it/s]
133
  27%|β–ˆβ–ˆβ–‹ | 229/851 [00:03<00:08, 72.60it/s]
134
  28%|β–ˆβ–ˆβ–Š | 237/851 [00:03<00:08, 72.11it/s]
135
  29%|β–ˆβ–ˆβ–‰ | 245/851 [00:03<00:08, 69.33it/s]
136
  30%|β–ˆβ–ˆβ–‰ | 254/851 [00:03<00:08, 72.97it/s]
137
  31%|β–ˆβ–ˆβ–ˆ | 263/851 [00:03<00:07, 75.87it/s]
138
  32%|β–ˆβ–ˆβ–ˆβ– | 271/851 [00:03<00:07, 74.16it/s]
139
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 280/851 [00:03<00:07, 76.76it/s]
140
  34%|β–ˆβ–ˆβ–ˆβ– | 288/851 [00:03<00:07, 74.29it/s]
141
  35%|β–ˆβ–ˆβ–ˆβ– | 296/851 [00:04<00:07, 75.23it/s]
142
  36%|β–ˆβ–ˆβ–ˆβ–Œ | 305/851 [00:04<00:07, 77.73it/s]
143
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 313/851 [00:04<00:07, 72.60it/s]
144
  38%|β–ˆβ–ˆβ–ˆβ–Š | 322/851 [00:04<00:06, 76.24it/s]
145
  39%|β–ˆβ–ˆβ–ˆβ–‰ | 330/851 [00:04<00:06, 74.73it/s]
146
  40%|β–ˆβ–ˆβ–ˆβ–‰ | 338/851 [00:04<00:06, 74.43it/s]
147
  41%|β–ˆβ–ˆβ–ˆβ–ˆ | 346/851 [00:04<00:06, 74.71it/s]
148
  42%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 354/851 [00:04<00:06, 72.06it/s]
149
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 362/851 [00:04<00:06, 71.45it/s]
150
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 370/851 [00:05<00:06, 71.67it/s]
151
  44%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 378/851 [00:05<00:06, 71.58it/s]
152
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 386/851 [00:05<00:06, 71.06it/s]
153
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 394/851 [00:05<00:06, 71.06it/s]
154
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 402/851 [00:05<00:06, 71.74it/s]
155
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 410/851 [00:05<00:06, 68.91it/s]
156
  49%|β–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 418/851 [00:05<00:06, 71.25it/s]
157
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 426/851 [00:05<00:06, 69.84it/s]
158
  51%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 434/851 [00:05<00:05, 72.02it/s]
159
  52%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 443/851 [00:06<00:05, 73.38it/s]
160
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 451/851 [00:06<00:05, 74.14it/s]
161
  54%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 459/851 [00:06<00:05, 75.25it/s]
162
  55%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 467/851 [00:06<00:05, 70.07it/s]
163
  56%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 475/851 [00:06<00:05, 65.66it/s]
164
  57%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 482/851 [00:06<00:05, 66.58it/s]
165
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 490/851 [00:06<00:05, 68.00it/s]
166
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 499/851 [00:06<00:04, 72.85it/s]
167
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 507/851 [00:06<00:04, 72.58it/s]
168
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 516/851 [00:07<00:04, 75.12it/s]
169
  62%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 524/851 [00:07<00:04, 71.11it/s]
170
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 532/851 [00:07<00:04, 70.85it/s]
171
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 540/851 [00:07<00:04, 73.05it/s]
172
  64%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 548/851 [00:07<00:04, 71.10it/s]
173
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 556/851 [00:07<00:04, 73.24it/s]
174
  66%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 565/851 [00:07<00:03, 76.58it/s]
175
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 573/851 [00:07<00:03, 76.46it/s]
176
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 581/851 [00:07<00:03, 75.06it/s]
177
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 589/851 [00:08<00:03, 70.81it/s]
178
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 597/851 [00:08<00:03, 71.95it/s]
179
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 605/851 [00:08<00:03, 71.62it/s]
180
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 613/851 [00:08<00:03, 69.39it/s]
181
  73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 621/851 [00:08<00:03, 70.18it/s]
182
  74%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 629/851 [00:08<00:03, 67.13it/s]
183
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 637/851 [00:08<00:03, 69.79it/s]
184
  76%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 645/851 [00:08<00:03, 65.67it/s]
185
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 653/851 [00:09<00:02, 68.49it/s]
186
  78%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 661/851 [00:09<00:02, 70.37it/s]
187
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 669/851 [00:09<00:02, 71.13it/s]
188
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 677/851 [00:09<00:02, 70.51it/s]
189
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 685/851 [00:09<00:02, 72.36it/s]
190
  82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 694/851 [00:09<00:02, 74.84it/s]
191
  83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 703/851 [00:09<00:01, 76.51it/s]
192
  84%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 712/851 [00:09<00:01, 78.46it/s]
193
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 720/851 [00:09<00:01, 76.13it/s]
194
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 729/851 [00:10<00:01, 78.36it/s]
195
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 737/851 [00:10<00:01, 78.39it/s]
196
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 745/851 [00:10<00:01, 78.52it/s]
197
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 753/851 [00:10<00:01, 77.70it/s]
198
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 762/851 [00:10<00:01, 79.73it/s]
199
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 770/851 [00:10<00:01, 74.82it/s]
200
  91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 778/851 [00:10<00:00, 74.26it/s]
201
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 786/851 [00:10<00:00, 73.12it/s]
202
  93%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 794/851 [00:10<00:00, 73.97it/s]
203
  94%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 803/851 [00:10<00:00, 76.42it/s]
204
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 811/851 [00:11<00:00, 73.67it/s]
205
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 819/851 [00:11<00:00, 74.38it/s]
206
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 828/851 [00:11<00:00, 76.43it/s]
207
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 836/851 [00:11<00:00, 76.27it/s]
208
  99%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 844/851 [00:11<00:00, 73.07it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-15 14:29:25.440457: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2
+ 2024-05-15 14:29:25.440508: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
3
+ 2024-05-15 14:29:25.442473: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
4
+ 2024-05-15 14:29:26.567248: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
5
+ 05/15/2024 14:29:28 - WARNING - __main__ - Process rank: 0, device: cuda:0, n_gpu: 1distributed training: True, 16-bits training: False
6
+ 05/15/2024 14:29:28 - INFO - __main__ - Training/evaluation parameters TrainingArguments(
7
+ _n_gpu=1,
8
+ accelerator_config={'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None},
9
+ adafactor=False,
10
+ adam_beta1=0.9,
11
+ adam_beta2=0.999,
12
+ adam_epsilon=1e-08,
13
+ auto_find_batch_size=False,
14
+ bf16=False,
15
+ bf16_full_eval=False,
16
+ data_seed=None,
17
+ dataloader_drop_last=False,
18
+ dataloader_num_workers=0,
19
+ dataloader_persistent_workers=False,
20
+ dataloader_pin_memory=True,
21
+ dataloader_prefetch_factor=None,
22
+ ddp_backend=None,
23
+ ddp_broadcast_buffers=None,
24
+ ddp_bucket_cap_mb=None,
25
+ ddp_find_unused_parameters=None,
26
+ ddp_timeout=1800,
27
+ debug=[],
28
+ deepspeed=None,
29
+ disable_tqdm=False,
30
+ dispatch_batches=None,
31
+ do_eval=True,
32
+ do_predict=True,
33
+ do_train=False,
34
+ eval_accumulation_steps=None,
35
+ eval_delay=0,
36
+ eval_do_concat_batches=True,
37
+ eval_steps=None,
38
+ evaluation_strategy=epoch,
39
+ fp16=False,
40
+ fp16_backend=auto,
41
+ fp16_full_eval=False,
42
+ fp16_opt_level=O1,
43
+ fsdp=[],
44
+ fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False},
45
+ fsdp_min_num_params=0,
46
+ fsdp_transformer_layer_cls_to_wrap=None,
47
+ full_determinism=False,
48
+ gradient_accumulation_steps=4,
49
+ gradient_checkpointing=False,
50
+ gradient_checkpointing_kwargs=None,
51
+ greater_is_better=True,
52
+ group_by_length=False,
53
+ half_precision_backend=auto,
54
+ hub_always_push=False,
55
+ hub_model_id=None,
56
+ hub_private_repo=False,
57
+ hub_strategy=every_save,
58
+ hub_token=<HUB_TOKEN>,
59
+ ignore_data_skip=False,
60
+ include_inputs_for_metrics=False,
61
+ include_num_input_tokens_seen=False,
62
+ include_tokens_per_second=False,
63
+ jit_mode_eval=False,
64
+ label_names=None,
65
+ label_smoothing_factor=0.0,
66
+ learning_rate=5e-05,
67
+ length_column_name=length,
68
+ load_best_model_at_end=True,
69
+ local_rank=0,
70
+ log_level=passive,
71
+ log_level_replica=warning,
72
+ log_on_each_node=True,
73
+ logging_dir=/content/dissertation/scripts/ner/output/tb,
74
+ logging_first_step=False,
75
+ logging_nan_inf_filter=True,
76
+ logging_steps=500,
77
+ logging_strategy=steps,
78
+ lr_scheduler_kwargs={},
79
+ lr_scheduler_type=linear,
80
+ max_grad_norm=1.0,
81
+ max_steps=-1,
82
+ metric_for_best_model=f1,
83
+ mp_parameters=,
84
+ neftune_noise_alpha=None,
85
+ no_cuda=False,
86
+ num_train_epochs=10.0,
87
+ optim=adamw_torch,
88
+ optim_args=None,
89
+ optim_target_modules=None,
90
+ output_dir=/content/dissertation/scripts/ner/output,
91
+ overwrite_output_dir=True,
92
+ past_index=-1,
93
+ per_device_eval_batch_size=8,
94
+ per_device_train_batch_size=4,
95
+ prediction_loss_only=False,
96
+ push_to_hub=True,
97
+ push_to_hub_model_id=None,
98
+ push_to_hub_organization=None,
99
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
100
+ ray_scope=last,
101
+ remove_unused_columns=True,
102
+ report_to=['tensorboard'],
103
+ resume_from_checkpoint=None,
104
+ run_name=/content/dissertation/scripts/ner/output,
105
+ save_on_each_node=False,
106
+ save_only_model=False,
107
+ save_safetensors=True,
108
+ save_steps=500,
109
+ save_strategy=epoch,
110
+ save_total_limit=None,
111
+ seed=42,
112
+ skip_memory_metrics=True,
113
+ split_batches=None,
114
+ tf32=None,
115
+ torch_compile=False,
116
+ torch_compile_backend=None,
117
+ torch_compile_mode=None,
118
+ torchdynamo=None,
119
+ tpu_metrics_debug=False,
120
+ tpu_num_cores=None,
121
+ use_cpu=False,
122
+ use_ipex=False,
123
+ use_legacy_prediction_loop=False,
124
+ use_mps_device=False,
125
+ warmup_ratio=0.0,
126
+ warmup_steps=0,
127
+ weight_decay=0.0,
128
+ )
129
+ /usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for Rodrigo1771/drugtemist-ner contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/Rodrigo1771/drugtemist-ner
130
+ You can avoid this message in future by passing the argument `trust_remote_code=True`.
131
+ Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`.
132
+ warnings.warn(
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+ /usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
141
+ warnings.warn(
142
+ [INFO|configuration_utils.py:726] 2024-05-15 14:29:40,285 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/config.json
143
+ [INFO|configuration_utils.py:789] 2024-05-15 14:29:40,289 >> Model config RobertaConfig {
144
+ "_name_or_path": "PlanTL-GOB-ES/bsc-bio-ehr-es",
145
+ "architectures": [
146
+ "RobertaForMaskedLM"
147
+ ],
148
+ "attention_probs_dropout_prob": 0.1,
149
+ "bos_token_id": 0,
150
+ "classifier_dropout": null,
151
+ "eos_token_id": 2,
152
+ "finetuning_task": "ner",
153
+ "gradient_checkpointing": false,
154
+ "hidden_act": "gelu",
155
+ "hidden_dropout_prob": 0.1,
156
+ "hidden_size": 768,
157
+ "id2label": {
158
+ "0": "O",
159
+ "1": "B-FARMACO",
160
+ "2": "I-FARMACO"
161
+ },
162
+ "initializer_range": 0.02,
163
+ "intermediate_size": 3072,
164
+ "label2id": {
165
+ "B-FARMACO": 1,
166
+ "I-FARMACO": 2,
167
+ "O": 0
168
+ },
169
+ "layer_norm_eps": 1e-05,
170
+ "max_position_embeddings": 514,
171
+ "model_type": "roberta",
172
+ "num_attention_heads": 12,
173
+ "num_hidden_layers": 12,
174
+ "pad_token_id": 1,
175
+ "position_embedding_type": "absolute",
176
+ "transformers_version": "4.40.2",
177
+ "type_vocab_size": 1,
178
+ "use_cache": true,
179
+ "vocab_size": 50262
180
+ }
181
+
182
+ [INFO|configuration_utils.py:726] 2024-05-15 14:29:40,381 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/config.json
183
+ [INFO|configuration_utils.py:789] 2024-05-15 14:29:40,382 >> Model config RobertaConfig {
184
+ "_name_or_path": "PlanTL-GOB-ES/bsc-bio-ehr-es",
185
+ "architectures": [
186
+ "RobertaForMaskedLM"
187
+ ],
188
+ "attention_probs_dropout_prob": 0.1,
189
+ "bos_token_id": 0,
190
+ "classifier_dropout": null,
191
+ "eos_token_id": 2,
192
+ "gradient_checkpointing": false,
193
+ "hidden_act": "gelu",
194
+ "hidden_dropout_prob": 0.1,
195
+ "hidden_size": 768,
196
+ "initializer_range": 0.02,
197
+ "intermediate_size": 3072,
198
+ "layer_norm_eps": 1e-05,
199
+ "max_position_embeddings": 514,
200
+ "model_type": "roberta",
201
+ "num_attention_heads": 12,
202
+ "num_hidden_layers": 12,
203
+ "pad_token_id": 1,
204
+ "position_embedding_type": "absolute",
205
+ "transformers_version": "4.40.2",
206
+ "type_vocab_size": 1,
207
+ "use_cache": true,
208
+ "vocab_size": 50262
209
+ }
210
+
211
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file vocab.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/vocab.json
212
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file merges.txt from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/merges.txt
213
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file tokenizer.json from cache at None
214
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file added_tokens.json from cache at None
215
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/special_tokens_map.json
216
+ [INFO|tokenization_utils_base.py:2087] 2024-05-15 14:29:40,392 >> loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/tokenizer_config.json
217
+ [INFO|configuration_utils.py:726] 2024-05-15 14:29:40,392 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/config.json
218
+ [INFO|configuration_utils.py:789] 2024-05-15 14:29:40,393 >> Model config RobertaConfig {
219
+ "_name_or_path": "PlanTL-GOB-ES/bsc-bio-ehr-es",
220
+ "architectures": [
221
+ "RobertaForMaskedLM"
222
+ ],
223
+ "attention_probs_dropout_prob": 0.1,
224
+ "bos_token_id": 0,
225
+ "classifier_dropout": null,
226
+ "eos_token_id": 2,
227
+ "gradient_checkpointing": false,
228
+ "hidden_act": "gelu",
229
+ "hidden_dropout_prob": 0.1,
230
+ "hidden_size": 768,
231
+ "initializer_range": 0.02,
232
+ "intermediate_size": 3072,
233
+ "layer_norm_eps": 1e-05,
234
+ "max_position_embeddings": 514,
235
+ "model_type": "roberta",
236
+ "num_attention_heads": 12,
237
+ "num_hidden_layers": 12,
238
+ "pad_token_id": 1,
239
+ "position_embedding_type": "absolute",
240
+ "transformers_version": "4.40.2",
241
+ "type_vocab_size": 1,
242
+ "use_cache": true,
243
+ "vocab_size": 50262
244
+ }
245
+
246
+ [INFO|configuration_utils.py:726] 2024-05-15 14:29:40,477 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/config.json
247
+ [INFO|configuration_utils.py:789] 2024-05-15 14:29:40,478 >> Model config RobertaConfig {
248
+ "_name_or_path": "PlanTL-GOB-ES/bsc-bio-ehr-es",
249
+ "architectures": [
250
+ "RobertaForMaskedLM"
251
+ ],
252
+ "attention_probs_dropout_prob": 0.1,
253
+ "bos_token_id": 0,
254
+ "classifier_dropout": null,
255
+ "eos_token_id": 2,
256
+ "gradient_checkpointing": false,
257
+ "hidden_act": "gelu",
258
+ "hidden_dropout_prob": 0.1,
259
+ "hidden_size": 768,
260
+ "initializer_range": 0.02,
261
+ "intermediate_size": 3072,
262
+ "layer_norm_eps": 1e-05,
263
+ "max_position_embeddings": 514,
264
+ "model_type": "roberta",
265
+ "num_attention_heads": 12,
266
+ "num_hidden_layers": 12,
267
+ "pad_token_id": 1,
268
+ "position_embedding_type": "absolute",
269
+ "transformers_version": "4.40.2",
270
+ "type_vocab_size": 1,
271
+ "use_cache": true,
272
+ "vocab_size": 50262
273
+ }
274
+
275
+ [INFO|modeling_utils.py:3429] 2024-05-15 14:29:40,721 >> loading weights file pytorch_model.bin from cache at /root/.cache/huggingface/hub/models--PlanTL-GOB-ES--bsc-bio-ehr-es/snapshots/1e543adb2d21f19d85a89305eebdbd64ab656b99/pytorch_model.bin
276
+ [INFO|modeling_utils.py:4160] 2024-05-15 14:29:40,847 >> Some weights of the model checkpoint at PlanTL-GOB-ES/bsc-bio-ehr-es were not used when initializing RobertaForTokenClassification: ['lm_head.bias', 'lm_head.decoder.bias', 'lm_head.decoder.weight', 'lm_head.dense.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.bias', 'lm_head.layer_norm.weight']
277
+ - This IS expected if you are initializing RobertaForTokenClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
278
+ - This IS NOT expected if you are initializing RobertaForTokenClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
279
+ [WARNING|modeling_utils.py:4172] 2024-05-15 14:29:40,847 >> Some weights of RobertaForTokenClassification were not initialized from the model checkpoint at PlanTL-GOB-ES/bsc-bio-ehr-es and are newly initialized: ['classifier.bias', 'classifier.weight']
280
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
281
+
282
+
283
+ /content/dissertation/scripts/ner/run_ner.py:397: FutureWarning: load_metric is deprecated and will be removed in the next major version of datasets. Use 'evaluate.load' instead, from the new library πŸ€— Evaluate: https://huggingface.co/docs/evaluate
284
+ metric = load_metric("seqeval")
285
+ /usr/local/lib/python3.10/dist-packages/datasets/load.py:759: FutureWarning: The repository for seqeval contains custom code which must be executed to correctly load the metric. You can inspect the repository content at https://raw.githubusercontent.com/huggingface/datasets/2.19.1/metrics/seqeval/seqeval.py
286
+ You can avoid this message in future by passing the argument `trust_remote_code=True`.
287
+ Passing `trust_remote_code=True` will be mandatory to load this metric from the next major release of `datasets`.
288
+ warnings.warn(
289
+
290
+ 05/15/2024 14:29:44 - INFO - __main__ - *** Evaluate ***
291
+ [INFO|trainer.py:786] 2024-05-15 14:29:44,602 >> The following columns in the evaluation set don't have a corresponding argument in `RobertaForTokenClassification.forward` and have been ignored: id, ner_tags, tokens. If id, ner_tags, tokens are not expected by `RobertaForTokenClassification.forward`, you can safely ignore this message.
292
+ [INFO|trainer.py:3614] 2024-05-15 14:29:44,607 >> ***** Running Evaluation *****
293
+ [INFO|trainer.py:3616] 2024-05-15 14:29:44,608 >> Num examples = 6807
294
+ [INFO|trainer.py:3619] 2024-05-15 14:29:44,608 >> Batch size = 8
295
+
296
  0%| | 0/851 [00:00<?, ?it/s]
297
  1%| | 6/851 [00:00<00:14, 56.94it/s]
298
  2%|▏ | 14/851 [00:00<00:12, 67.06it/s]
299
  3%|β–Ž | 22/851 [00:00<00:11, 69.14it/s]
300
  4%|β–Ž | 30/851 [00:00<00:11, 70.86it/s]
301
  4%|▍ | 38/851 [00:00<00:11, 72.79it/s]
302
  5%|β–Œ | 46/851 [00:00<00:10, 74.66it/s]
303
  6%|β–‹ | 54/851 [00:00<00:10, 75.69it/s]
304
  7%|β–‹ | 62/851 [00:00<00:10, 76.61it/s]
305
  8%|β–Š | 70/851 [00:00<00:10, 71.80it/s]
306
  9%|β–‰ | 78/851 [00:01<00:10, 72.55it/s]
307
  10%|β–ˆ | 86/851 [00:01<00:10, 73.09it/s]
308
  11%|β–ˆ | 94/851 [00:01<00:10, 71.17it/s]
309
  12%|β–ˆβ– | 102/851 [00:01<00:10, 72.61it/s]
310
  13%|β–ˆβ–Ž | 110/851 [00:01<00:10, 71.20it/s]
311
  14%|β–ˆβ– | 118/851 [00:01<00:10, 72.02it/s]
312
  15%|β–ˆβ– | 126/851 [00:01<00:10, 68.64it/s]
313
  16%|β–ˆβ–Œ | 134/851 [00:01<00:10, 68.76it/s]
314
  17%|β–ˆβ–‹ | 142/851 [00:01<00:10, 69.40it/s]
315
  18%|β–ˆβ–Š | 149/851 [00:02<00:10, 67.95it/s]
316
  18%|β–ˆβ–Š | 157/851 [00:02<00:09, 70.23it/s]
317
  19%|β–ˆβ–‰ | 165/851 [00:02<00:09, 71.18it/s]
318
  20%|β–ˆβ–ˆ | 173/851 [00:02<00:09, 72.70it/s]
319
  21%|β–ˆβ–ˆβ– | 181/851 [00:02<00:09, 73.97it/s]
320
  22%|β–ˆβ–ˆβ– | 189/851 [00:02<00:08, 75.32it/s]
321
  23%|β–ˆβ–ˆβ–Ž | 197/851 [00:02<00:08, 75.83it/s]
322
  24%|β–ˆβ–ˆβ– | 206/851 [00:02<00:08, 76.15it/s]
323
  25%|β–ˆβ–ˆβ–Œ | 214/851 [00:02<00:08, 71.48it/s]
324
  26%|β–ˆβ–ˆβ–Œ | 222/851 [00:03<00:08, 72.75it/s]
325
  27%|β–ˆβ–ˆβ–‹ | 230/851 [00:03<00:08, 73.89it/s]
326
  28%|β–ˆβ–ˆβ–Š | 238/851 [00:03<00:08, 72.13it/s]
327
  29%|β–ˆβ–ˆβ–‰ | 246/851 [00:03<00:08, 70.45it/s]
328
  30%|β–ˆβ–ˆβ–‰ | 255/851 [00:03<00:08, 73.88it/s]
329
  31%|β–ˆβ–ˆβ–ˆ | 264/851 [00:03<00:07, 76.63it/s]
330
  32%|β–ˆβ–ˆβ–ˆβ– | 272/851 [00:03<00:07, 74.50it/s]
331
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 281/851 [00:03<00:07, 77.22it/s]
332
  34%|β–ˆβ–ˆβ–ˆβ– | 289/851 [00:03<00:07, 74.33it/s]
333
  35%|β–ˆβ–ˆβ–ˆβ–Œ | 298/851 [00:04<00:07, 76.25it/s]
334
  36%|β–ˆβ–ˆβ–ˆβ–Œ | 307/851 [00:04<00:06, 78.12it/s]
335
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 315/851 [00:04<00:07, 73.67it/s]
336
  38%|β–ˆβ–ˆβ–ˆβ–Š | 324/851 [00:04<00:07, 74.73it/s]
337
  39%|β–ˆβ–ˆβ–ˆβ–‰ | 332/851 [00:04<00:07, 73.75it/s]
338
  40%|β–ˆβ–ˆβ–ˆβ–‰ | 340/851 [00:04<00:06, 74.75it/s]
339
  41%|β–ˆβ–ˆβ–ˆβ–ˆ | 348/851 [00:04<00:06, 75.70it/s]
340
  42%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 356/851 [00:04<00:06, 72.87it/s]
341
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 364/851 [00:04<00:06, 73.10it/s]
342
  44%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 372/851 [00:05<00:06, 71.82it/s]
343
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 380/851 [00:05<00:06, 69.17it/s]
344
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 388/851 [00:05<00:06, 72.06it/s]
345
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 396/851 [00:05<00:06, 72.74it/s]
346
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 404/851 [00:05<00:06, 68.56it/s]
347
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 412/851 [00:05<00:06, 69.70it/s]
348
  49%|β–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 420/851 [00:05<00:05, 72.48it/s]
349
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 428/851 [00:05<00:05, 70.96it/s]
350
  51%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 436/851 [00:06<00:05, 72.47it/s]
351
  52%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 444/851 [00:06<00:05, 74.07it/s]
352
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 452/851 [00:06<00:05, 75.55it/s]
353
  54%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 460/851 [00:06<00:05, 74.33it/s]
354
  55%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 468/851 [00:06<00:05, 71.78it/s]
355
  56%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 476/851 [00:06<00:05, 67.46it/s]
356
  57%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 483/851 [00:06<00:05, 67.69it/s]
357
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 491/851 [00:06<00:05, 69.09it/s]
358
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 500/851 [00:06<00:04, 73.77it/s]
359
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 508/851 [00:07<00:04, 72.84it/s]
360
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 517/851 [00:07<00:04, 74.44it/s]
361
  62%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 525/851 [00:07<00:04, 70.40it/s]
362
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 533/851 [00:07<00:04, 71.47it/s]
363
  64%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 541/851 [00:07<00:04, 73.42it/s]
364
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 549/851 [00:07<00:04, 71.38it/s]
365
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 557/851 [00:07<00:04, 73.46it/s]
366
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 566/851 [00:07<00:03, 76.65it/s]
367
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 574/851 [00:07<00:03, 76.97it/s]
368
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 582/851 [00:08<00:03, 74.62it/s]
369
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 590/851 [00:08<00:03, 71.50it/s]
370
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 598/851 [00:08<00:03, 72.36it/s]
371
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 606/851 [00:08<00:03, 70.56it/s]
372
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 614/851 [00:08<00:03, 69.62it/s]
373
  73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 622/851 [00:08<00:03, 71.02it/s]
374
  74%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 630/851 [00:08<00:03, 67.52it/s]
375
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 638/851 [00:08<00:03, 70.60it/s]
376
  76%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 646/851 [00:08<00:03, 66.60it/s]
377
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 654/851 [00:09<00:02, 69.98it/s]
378
  78%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 662/851 [00:09<00:02, 70.75it/s]
379
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 670/851 [00:09<00:02, 72.07it/s]
380
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 678/851 [00:09<00:02, 70.99it/s]
381
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 686/851 [00:09<00:02, 72.75it/s]
382
  82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 695/851 [00:09<00:02, 75.21it/s]
383
  83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 704/851 [00:09<00:01, 76.96it/s]
384
  84%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 713/851 [00:09<00:01, 78.47it/s]
385
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 721/851 [00:09<00:01, 76.12it/s]
386
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 730/851 [00:10<00:01, 78.39it/s]
387
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 739/851 [00:10<00:01, 78.90it/s]
388
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 747/851 [00:10<00:01, 78.42it/s]
389
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 755/851 [00:10<00:01, 78.48it/s]
390
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 764/851 [00:10<00:01, 78.64it/s]
391
  91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 772/851 [00:10<00:01, 74.92it/s]
392
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 780/851 [00:10<00:00, 72.75it/s]
393
  93%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 788/851 [00:10<00:00, 73.39it/s]
394
  94%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 796/851 [00:10<00:00, 75.11it/s]
395
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 805/851 [00:11<00:00, 76.84it/s]
396
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 813/851 [00:11<00:00, 73.85it/s]
397
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 821/851 [00:11<00:00, 74.79it/s]
398
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 829/851 [00:11<00:00, 75.96it/s]
399
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 837/851 [00:11<00:00, 74.64it/s]
400
  99%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 845/851 [00:11<00:00, 69.55it/s]
401
+ ***** eval metrics *****
402
+ eval_accuracy = 0.0288
403
+ eval_f1 = 0.0056
404
+ eval_loss = 1.293
405
+ eval_precision = 0.0028
406
+ eval_recall = 0.2693
407
+ eval_runtime = 0:00:16.57
408
+ eval_samples = 6807
409
+ eval_samples_per_second = 410.727
410
+ eval_steps_per_second = 51.348
411
+ 05/15/2024 14:30:01 - INFO - __main__ - *** Predict ***
412
+ [INFO|trainer.py:786] 2024-05-15 14:30:01,182 >> The following columns in the test set don't have a corresponding argument in `RobertaForTokenClassification.forward` and have been ignored: id, ner_tags, tokens. If id, ner_tags, tokens are not expected by `RobertaForTokenClassification.forward`, you can safely ignore this message.
413
+ [INFO|trainer.py:3614] 2024-05-15 14:30:01,184 >> ***** Running Prediction *****
414
+ [INFO|trainer.py:3616] 2024-05-15 14:30:01,184 >> Num examples = 6807
415
+ [INFO|trainer.py:3619] 2024-05-15 14:30:01,184 >> Batch size = 8
416
+
417
  0%| | 0/851 [00:00<?, ?it/s]
418
  1%| | 10/851 [00:00<00:08, 95.20it/s]
419
  2%|▏ | 20/851 [00:00<00:10, 81.67it/s]
420
  3%|β–Ž | 29/851 [00:00<00:10, 78.36it/s]
421
  4%|▍ | 37/851 [00:00<00:10, 76.45it/s]
422
  5%|β–Œ | 45/851 [00:00<00:10, 76.74it/s]
423
  6%|β–Œ | 53/851 [00:00<00:10, 77.26it/s]
424
  7%|β–‹ | 62/851 [00:00<00:10, 77.44it/s]
425
  8%|β–Š | 70/851 [00:00<00:10, 72.24it/s]
426
  9%|β–‰ | 78/851 [00:01<00:10, 73.99it/s]
427
  10%|β–ˆ | 86/851 [00:01<00:10, 73.73it/s]
428
  11%|β–ˆ | 94/851 [00:01<00:10, 72.12it/s]
429
  12%|β–ˆβ– | 102/851 [00:01<00:10, 73.14it/s]
430
  13%|β–ˆβ–Ž | 110/851 [00:01<00:10, 71.41it/s]
431
  14%|β–ˆβ– | 118/851 [00:01<00:10, 72.63it/s]
432
  15%|β–ˆβ– | 126/851 [00:01<00:10, 69.23it/s]
433
  16%|β–ˆβ–Œ | 133/851 [00:01<00:10, 69.22it/s]
434
  16%|β–ˆβ–‹ | 140/851 [00:01<00:10, 69.15it/s]
435
  17%|β–ˆβ–‹ | 148/851 [00:02<00:10, 68.62it/s]
436
  18%|β–ˆβ–Š | 156/851 [00:02<00:09, 70.75it/s]
437
  19%|β–ˆβ–‰ | 164/851 [00:02<00:09, 71.62it/s]
438
  20%|β–ˆβ–ˆ | 172/851 [00:02<00:09, 73.75it/s]
439
  21%|β–ˆβ–ˆ | 180/851 [00:02<00:09, 74.25it/s]
440
  22%|β–ˆβ–ˆβ– | 188/851 [00:02<00:08, 74.90it/s]
441
  23%|β–ˆβ–ˆβ–Ž | 196/851 [00:02<00:08, 76.03it/s]
442
  24%|β–ˆβ–ˆβ– | 205/851 [00:02<00:08, 77.93it/s]
443
  25%|β–ˆβ–ˆβ–Œ | 213/851 [00:02<00:08, 72.81it/s]
444
  26%|β–ˆβ–ˆβ–Œ | 221/851 [00:03<00:08, 71.31it/s]
445
  27%|β–ˆβ–ˆβ–‹ | 229/851 [00:03<00:08, 72.60it/s]
446
  28%|β–ˆβ–ˆβ–Š | 237/851 [00:03<00:08, 72.11it/s]
447
  29%|β–ˆβ–ˆβ–‰ | 245/851 [00:03<00:08, 69.33it/s]
448
  30%|β–ˆβ–ˆβ–‰ | 254/851 [00:03<00:08, 72.97it/s]
449
  31%|β–ˆβ–ˆβ–ˆ | 263/851 [00:03<00:07, 75.87it/s]
450
  32%|β–ˆβ–ˆβ–ˆβ– | 271/851 [00:03<00:07, 74.16it/s]
451
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 280/851 [00:03<00:07, 76.76it/s]
452
  34%|β–ˆβ–ˆβ–ˆβ– | 288/851 [00:03<00:07, 74.29it/s]
453
  35%|β–ˆβ–ˆβ–ˆβ– | 296/851 [00:04<00:07, 75.23it/s]
454
  36%|β–ˆβ–ˆβ–ˆβ–Œ | 305/851 [00:04<00:07, 77.73it/s]
455
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 313/851 [00:04<00:07, 72.60it/s]
456
  38%|β–ˆβ–ˆβ–ˆβ–Š | 322/851 [00:04<00:06, 76.24it/s]
457
  39%|β–ˆβ–ˆβ–ˆβ–‰ | 330/851 [00:04<00:06, 74.73it/s]
458
  40%|β–ˆβ–ˆβ–ˆβ–‰ | 338/851 [00:04<00:06, 74.43it/s]
459
  41%|β–ˆβ–ˆβ–ˆβ–ˆ | 346/851 [00:04<00:06, 74.71it/s]
460
  42%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 354/851 [00:04<00:06, 72.06it/s]
461
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 362/851 [00:04<00:06, 71.45it/s]
462
  43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 370/851 [00:05<00:06, 71.67it/s]
463
  44%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 378/851 [00:05<00:06, 71.58it/s]
464
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 386/851 [00:05<00:06, 71.06it/s]
465
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 394/851 [00:05<00:06, 71.06it/s]
466
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 402/851 [00:05<00:06, 71.74it/s]
467
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 410/851 [00:05<00:06, 68.91it/s]
468
  49%|β–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 418/851 [00:05<00:06, 71.25it/s]
469
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 426/851 [00:05<00:06, 69.84it/s]
470
  51%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 434/851 [00:05<00:05, 72.02it/s]
471
  52%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 443/851 [00:06<00:05, 73.38it/s]
472
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 451/851 [00:06<00:05, 74.14it/s]
473
  54%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 459/851 [00:06<00:05, 75.25it/s]
474
  55%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 467/851 [00:06<00:05, 70.07it/s]
475
  56%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 475/851 [00:06<00:05, 65.66it/s]
476
  57%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 482/851 [00:06<00:05, 66.58it/s]
477
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 490/851 [00:06<00:05, 68.00it/s]
478
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 499/851 [00:06<00:04, 72.85it/s]
479
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 507/851 [00:06<00:04, 72.58it/s]
480
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 516/851 [00:07<00:04, 75.12it/s]
481
  62%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 524/851 [00:07<00:04, 71.11it/s]
482
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 532/851 [00:07<00:04, 70.85it/s]
483
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 540/851 [00:07<00:04, 73.05it/s]
484
  64%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 548/851 [00:07<00:04, 71.10it/s]
485
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 556/851 [00:07<00:04, 73.24it/s]
486
  66%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 565/851 [00:07<00:03, 76.58it/s]
487
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 573/851 [00:07<00:03, 76.46it/s]
488
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 581/851 [00:07<00:03, 75.06it/s]
489
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 589/851 [00:08<00:03, 70.81it/s]
490
  70%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 597/851 [00:08<00:03, 71.95it/s]
491
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 605/851 [00:08<00:03, 71.62it/s]
492
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 613/851 [00:08<00:03, 69.39it/s]
493
  73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 621/851 [00:08<00:03, 70.18it/s]
494
  74%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 629/851 [00:08<00:03, 67.13it/s]
495
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 637/851 [00:08<00:03, 69.79it/s]
496
  76%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 645/851 [00:08<00:03, 65.67it/s]
497
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 653/851 [00:09<00:02, 68.49it/s]
498
  78%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 661/851 [00:09<00:02, 70.37it/s]
499
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 669/851 [00:09<00:02, 71.13it/s]
500
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 677/851 [00:09<00:02, 70.51it/s]
501
  80%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 685/851 [00:09<00:02, 72.36it/s]
502
  82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 694/851 [00:09<00:02, 74.84it/s]
503
  83%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 703/851 [00:09<00:01, 76.51it/s]
504
  84%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 712/851 [00:09<00:01, 78.46it/s]
505
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 720/851 [00:09<00:01, 76.13it/s]
506
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 729/851 [00:10<00:01, 78.36it/s]
507
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 737/851 [00:10<00:01, 78.39it/s]
508
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 745/851 [00:10<00:01, 78.52it/s]
509
  88%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 753/851 [00:10<00:01, 77.70it/s]
510
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 762/851 [00:10<00:01, 79.73it/s]
511
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 770/851 [00:10<00:01, 74.82it/s]
512
  91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 778/851 [00:10<00:00, 74.26it/s]
513
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 786/851 [00:10<00:00, 73.12it/s]
514
  93%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž| 794/851 [00:10<00:00, 73.97it/s]
515
  94%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 803/851 [00:10<00:00, 76.42it/s]
516
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 811/851 [00:11<00:00, 73.67it/s]
517
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 819/851 [00:11<00:00, 74.38it/s]
518
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 828/851 [00:11<00:00, 76.43it/s]
519
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 836/851 [00:11<00:00, 76.27it/s]
520
  99%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 844/851 [00:11<00:00, 73.07it/s]
521
+ [INFO|trainer.py:3305] 2024-05-15 14:30:17,255 >> Saving model checkpoint to /content/dissertation/scripts/ner/output
522
+ [INFO|configuration_utils.py:471] 2024-05-15 14:30:17,256 >> Configuration saved in /content/dissertation/scripts/ner/output/config.json
523
+ [INFO|modeling_utils.py:2590] 2024-05-15 14:30:18,218 >> Model weights saved in /content/dissertation/scripts/ner/output/model.safetensors
524
+ [INFO|tokenization_utils_base.py:2488] 2024-05-15 14:30:18,219 >> tokenizer config file saved in /content/dissertation/scripts/ner/output/tokenizer_config.json
525
+ [INFO|tokenization_utils_base.py:2497] 2024-05-15 14:30:18,219 >> Special tokens file saved in /content/dissertation/scripts/ner/output/special_tokens_map.json
526
+ [INFO|modelcard.py:450] 2024-05-15 14:30:18,357 >> Dropping the following result as it does not have all the necessary fields:
527
+ {'task': {'name': 'Token Classification', 'type': 'token-classification'}, 'dataset': {'name': 'Rodrigo1771/drugtemist-ner', 'type': 'Rodrigo1771/drugtemist-ner', 'config': 'DrugTEMIST NER', 'split': 'validation', 'args': 'DrugTEMIST NER'}}
528
+ ***** predict metrics *****
529
+ predict_accuracy = 0.0288
530
+ predict_f1 = 0.0056
531
+ predict_loss = 1.293
532
+ predict_precision = 0.0028
533
+ predict_recall = 0.2693
534
+ predict_runtime = 0:00:15.76
535
+ predict_samples_per_second = 431.659
536
+ predict_steps_per_second = 53.965
537
+
538
+
539
+
540
+
541
+
542
+
543
+
544
+
545
+
546
+
547
+
548
+
549
+
550
+
551
+
552
+
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18cce49dc172023921a8c234d2d643afeaa0b4f8f209fd2e1d76d267cbbe3c95
3
+ size 5048
vocab.json ADDED
The diff for this file is too large to render. See raw diff