arnabdhar commited on
Commit
46a9363
1 Parent(s): 313e861

End of training

Browse files
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: xlm-roberta-base
4
+ tags:
5
+ - pytorch
6
+ - XLMRobertaForTokenClassification
7
+ - named-entity-recognition
8
+ - wikipedia
9
+ - generated_from_trainer
10
+ model-index:
11
+ - name: xlm-roberta-base-wikineural
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # xlm-roberta-base-wikineural
19
+
20
+ This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 0.0467
23
+
24
+ ## Model description
25
+
26
+ More information needed
27
+
28
+ ## Intended uses & limitations
29
+
30
+ More information needed
31
+
32
+ ## Training and evaluation data
33
+
34
+ More information needed
35
+
36
+ ## Training procedure
37
+
38
+ ### Training hyperparameters
39
+
40
+ The following hyperparameters were used during training:
41
+ - learning_rate: 3e-05
42
+ - train_batch_size: 32
43
+ - eval_batch_size: 128
44
+ - seed: 37912547
45
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
+ - lr_scheduler_type: cosine
47
+ - lr_scheduler_warmup_ratio: 0.1
48
+ - training_steps: 100000
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:-----:|:------:|:---------------:|
54
+ | 0.0858 | 0.14 | 10000 | 0.0817 |
55
+ | 0.0719 | 0.28 | 20000 | 0.0660 |
56
+ | 0.0656 | 0.43 | 30000 | 0.0631 |
57
+ | 0.0598 | 0.57 | 40000 | 0.0574 |
58
+ | 0.0551 | 0.71 | 50000 | 0.0534 |
59
+ | 0.0523 | 0.85 | 60000 | 0.0512 |
60
+ | 0.0519 | 0.99 | 70000 | 0.0484 |
61
+ | 0.0418 | 1.13 | 80000 | 0.0480 |
62
+ | 0.042 | 1.28 | 90000 | 0.0469 |
63
+ | 0.041 | 1.42 | 100000 | 0.0467 |
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - Transformers 4.35.2
69
+ - Pytorch 2.1.0+cu118
70
+ - Datasets 2.15.0
71
+ - Tokenizers 0.15.0
runs/Dec09_08-54-55_584b4064680c/events.out.tfevents.1702165372.584b4064680c.508.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a1a7bac1e745b82d0e222c9040212d6476b9ae561e9d224b6b741d42a752841
3
+ size 364