LGS-Mercy commited on
Commit
a333b9b
1 Parent(s): dc7c26f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -0
README.md CHANGED
@@ -50,3 +50,21 @@ from transformers import pipeline
50
  unmasker = pipeline('fill-mask', model='EnergyBERT')
51
  unmasker("Hello I'm a <mask> model.")
52
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  unmasker = pipeline('fill-mask', model='EnergyBERT')
51
  unmasker("Hello I'm a <mask> model.")
52
  ```
53
+
54
+ Training Details
55
+
56
+ ## Training Data
57
+
58
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
59
+
60
+ 1.2M Published full-text literature corpus from 2000 to 2021.
61
+
62
+ ## Training Procedure
63
+
64
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
65
+
66
+ BERT is trained on two unsupervised tasks during its pre-training period: masked language modeling and next sentence prediction. A masked language model involves masking some of the input tokens at random and training the model to predict the masked tokens based on the context surrounding the input tokens. Next sentence prediction involves training the model to predict whether two sentences follow each other logically.
67
+
68
+ ### Training Hyperparameters
69
+
70
+ - **Training regime:**