lakshaywadhwa1993 commited on
Commit
b428f18
1 Parent(s): 53a12a5

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -24
README.md CHANGED
@@ -1,40 +1,66 @@
1
  ---
2
- language: mr
 
 
3
  datasets:
4
  - wikiann
5
- examples:
6
- widget:
7
- - text: "राज्यसभा निवडणुकांसाठी उद्या मुंबईत मतदान होणार आहे."
8
- example_title: "Sentence_1"
9
- - text: "विराट कोहली भारताकडून खेळतो."
10
- example_title: "Sentence_2"
11
- - text: "नवी दिल्ली ही भारताची राजधानी आहे"
12
- example_title: "Sentence_3"
13
  ---
14
 
15
- <h1>Marathi Named Entity Recognition Model trained using transfer learning</h1>
16
- Fine-tuning bert-base-multilingual-cased on Wikiann dataset for performing NER on Marathi language.
17
 
 
18
 
19
- ## Label ID and its corresponding label name
 
 
 
 
 
 
 
 
 
20
 
21
- Label list: (0), B-PER (1), I-PER (2), B-ORG (3), I-ORG (4), B-LOC (5), I-LOC (6)
22
 
 
23
 
24
- Example
25
- ```py
26
- from transformers import AutoTokenizer, AutoModelForTokenClassification
27
- from transformers import pipeline
28
 
29
- tokenizer = AutoTokenizer.from_pretrained("lakshaywadhwa1993/ner_marathi_bert")
30
 
31
- model = AutoModelForTokenClassification.from_pretrained("lakshaywadhwa1993/ner_marathi_bert")
32
 
33
- nlp = pipeline("ner", model=model, tokenizer=tokenizer)
34
 
35
- example = ["राज्यसभा","निवडणुकांसाठी","मुंबईत","भाजपचे" ,"चिंचवडचे", "आमदार", "लक्ष्मण", "जगताप"]
36
 
37
- results = nlp(example)
38
 
39
- results
40
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
  datasets:
6
  - wikiann
7
+ model-index:
8
+ - name: ner_marathi_bert
9
+ results: []
 
 
 
 
 
10
  ---
11
 
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # ner_marathi_bert
16
 
17
+ This model is a fine-tuned version of [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased) on the wikiann dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 0.3606
20
+ - Overall Precision: 0.8939
21
+ - Overall Recall: 0.9030
22
+ - Overall F1: 0.8984
23
+ - Overall Accuracy: 0.9347
24
+ - Loc F1: 0.8823
25
+ - Org F1: 0.8555
26
+ - Per F1: 0.9435
27
 
28
+ ## Model description
29
 
30
+ More information needed
31
 
32
+ ## Intended uses & limitations
 
 
 
33
 
34
+ More information needed
35
 
36
+ ## Training and evaluation data
37
 
38
+ More information needed
39
 
40
+ ## Training procedure
41
 
42
+ ### Training hyperparameters
43
 
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 2e-05
46
+ - train_batch_size: 16
47
+ - eval_batch_size: 16
48
+ - seed: 42
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - num_epochs: 7
52
+
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy | Loc F1 | Org F1 | Per F1 |
56
+ |:-------------:|:-----:|:----:|:---------------:|:-----------------:|:--------------:|:----------:|:----------------:|:------:|:------:|:------:|
57
+ | 0.2961 | 3.19 | 1000 | 0.3496 | 0.8720 | 0.8841 | 0.8780 | 0.9229 | 0.8599 | 0.8210 | 0.9343 |
58
+ | 0.0613 | 6.39 | 2000 | 0.3606 | 0.8939 | 0.9030 | 0.8984 | 0.9347 | 0.8823 | 0.8555 | 0.9435 |
59
+
60
+
61
+ ### Framework versions
62
+
63
+ - Transformers 4.21.0
64
+ - Pytorch 1.12.0+cu113
65
+ - Datasets 2.4.0
66
+ - Tokenizers 0.12.1