JeswinMS4 commited on
Commit
766ae09
1 Parent(s): 8902020

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -8
README.md CHANGED
@@ -8,14 +8,30 @@ tags:
8
  language:
9
  - en
10
  ---
11
- # BERT base model (uncased)
12
-
13
  Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in
14
  [this paper](https://arxiv.org/abs/1810.04805) and first released in
15
  [this repository](https://github.com/google-research/bert). This model is uncased: it does not make a difference
16
  between english and English.
17
 
18
- ## Model Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  ### Model Description
21
 
@@ -37,9 +53,8 @@ This way, the model learns an inner representation of the English language that
37
  useful for downstream tasks: if you have a dataset of labeled sentences, for instance, you can train a standard
38
  classifier using the features produced by the BERT model as inputs.
39
 
40
- - **Developed by:** [Jeswin MS, Venkatesh R, Kushal S Ballari]
41
- - **Model type:** [Intent Classification]
42
- - **Language(s) (NLP):** [English]
43
- - **License:** []
44
- - **Finetuned from model [optional]:** [Bert-base-uncased]
45
 
 
8
  language:
9
  - en
10
  ---
11
+ # BERT Base Intent model
12
+ This is a fine tuned model based on Bert-Base-Uncased model. This is model is used to classify intent into 3 categories- Fintech, Out of Scope and Abusive.
13
  Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in
14
  [this paper](https://arxiv.org/abs/1810.04805) and first released in
15
  [this repository](https://github.com/google-research/bert). This model is uncased: it does not make a difference
16
  between english and English.
17
 
18
+ ## Training procedure
19
+
20
+ ### Training hyperparameters
21
+
22
+ The following hyperparameters were used during training:
23
+ - learning_rate: 2e-5
24
+ - num_epochs: 3
25
+ - weight_decay:0.01
26
+
27
+ ### Training results
28
+
29
+ | Training Loss | Epoch | Validation Loss | Accuracy | F1 |
30
+ |:-------------:|:-----:|:----------------:|:---------------:|:--------:|
31
+ | 0.114200 | 1.0 | 0.034498 | 0.991351 | 0.991346 |
32
+ | 0.024100 | 2.0 | 0.037945 | 0.992349 | 0.992355 |
33
+ | 0.009800 | 3.0 | 0.034846 | 0.993347 | 0.993345 |
34
+
35
 
36
  ### Model Description
37
 
 
53
  useful for downstream tasks: if you have a dataset of labeled sentences, for instance, you can train a standard
54
  classifier using the features produced by the BERT model as inputs.
55
 
56
+ - **Developed by:** Jeswin MS, Venkatesh R, Kushal S Ballari
57
+ - **Model type:** Intent Classification
58
+ - **Language(s) (NLP):** English
59
+ - **Finetuned from model:** Bert-base-uncased
 
60