nlpaueb commited on
Commit
4504270
1 Parent(s): d80e499

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -7
README.md CHANGED
@@ -21,10 +21,9 @@ widget:
21
 
22
  SEC-BERT is a family of BERT models for the financial domain, intended to assist financial NLP research and FinTech applications.
23
  SEC-BERT consists of the following models:
24
- * SEC-BERT-BASE (this model)
25
- * [SEC-BERT-NUM](https://huggingface.co/nlpaueb/sec-bert-num): We replace every number token with a [NUM] pseudo-token handling all numeric expressions in a uniform manner, disallowing their fragmentation)
26
- * [SEC-BERT-SHAPE](https://huggingface.co/nlpaueb/sec-bert-shape): We replace numbers with pseudo-tokens that represent the number’s shape, so numeric expressions (of known shapes) are no longer fragmented.<br>
27
- (e.g. , '53.2' becomes '[XX.X]' and '40,200.5' becomes '[XX,XXX.X]').<br>
28
 
29
  ## Pre-training corpus
30
 
@@ -37,7 +36,7 @@ The model was pre-trained on 260,773 10-K filings from 1993-2019, publicly avail
37
  * We then used [Hugging Face](https://huggingface.co)'s [Transformers](https://github.com/huggingface/transformers) conversion script to convert the TF checkpoint in the desired format in order to be able to load the model in two lines of code for both PyTorch and TF2 users.
38
  * We release a model similar to the English BERT-BASE model (12-layer, 768-hidden, 12-heads, 110M parameters).
39
  * We chose to follow the same training set-up: 1 million training steps with batches of 256 sequences of length 512 with an initial learning rate 1e-4.
40
- * We were able to use a single Google Cloud TPU v3-8 provided for free from [TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc), while also utilizing [GCP research credits](https://edu.google.com/programs/credits/research). Huge thanks to both Google programs for supporting us!
41
 
42
  ## Load Pretrained Model
43
 
@@ -48,7 +47,7 @@ tokenizer = AutoTokenizer.from_pretrained("nlpaueb/sec-bert-base")
48
  model = AutoModel.from_pretrained("nlpaueb/sec-bert-base")
49
  ```
50
 
51
- ## Use SEC-BERT variants as Language Models
52
 
53
  | Sample | Masked Token |
54
  | --------------------------------------------------- | ------------ |
@@ -190,7 +189,7 @@ In the Proceedings of the 60th Annual Meeting of the Association for Computation
190
 
191
  ```
192
  @inproceedings{loukas-etal-2022-finer,
193
- title = "{FiNER}: Financial Numeric Entity Recognition for {XBRL} Tagging",
194
  author = "Loukas, Lefteris and
195
  Fergadiotis, Manos and
196
  Chalkidis, Ilias and
 
21
 
22
  SEC-BERT is a family of BERT models for the financial domain, intended to assist financial NLP research and FinTech applications.
23
  SEC-BERT consists of the following models:
24
+ * SEC-BERT-BASE (this model): Same architecture as BERT-BASE trained on financial documents.
25
+ * [SEC-BERT-NUM](https://huggingface.co/nlpaueb/sec-bert-num): Same as SEC-BERT-BASE but we replace every number token with a [NUM] pseudo-token handling all numeric expressions in a uniform manner, disallowing their fragmentation
26
+ * [SEC-BERT-SHAPE](https://huggingface.co/nlpaueb/sec-bert-shape): Same as SEC-BERT-BASE but we replace numbers with pseudo-tokens that represent the number’s shape, so numeric expressions (of known shapes) are no longer fragmented, e.g., '53.2' becomes '[XX.X]' and '40,200.5' becomes '[XX,XXX.X]'.
 
27
 
28
  ## Pre-training corpus
29
 
 
36
  * We then used [Hugging Face](https://huggingface.co)'s [Transformers](https://github.com/huggingface/transformers) conversion script to convert the TF checkpoint in the desired format in order to be able to load the model in two lines of code for both PyTorch and TF2 users.
37
  * We release a model similar to the English BERT-BASE model (12-layer, 768-hidden, 12-heads, 110M parameters).
38
  * We chose to follow the same training set-up: 1 million training steps with batches of 256 sequences of length 512 with an initial learning rate 1e-4.
39
+ * We were able to use a single Google Cloud TPU v3-8 provided for free from [TensorFlow Research Cloud (TRC)]((https://sites.research.google/trc), while also utilizing [GCP research credits](https://edu.google.com/programs/credits/research). Huge thanks to both Google programs for supporting us!
40
 
41
  ## Load Pretrained Model
42
 
 
47
  model = AutoModel.from_pretrained("nlpaueb/sec-bert-base")
48
  ```
49
 
50
+ ## Using SEC-BERT variants as Language Models
51
 
52
  | Sample | Masked Token |
53
  | --------------------------------------------------- | ------------ |
 
189
 
190
  ```
191
  @inproceedings{loukas-etal-2022-finer,
192
+ title = "{FiNER: Financial Numeric Entity Recognition for XBRL Tagging}",
193
  author = "Loukas, Lefteris and
194
  Fergadiotis, Manos and
195
  Chalkidis, Ilias and