doberst commited on
Commit
e81f464
1 Parent(s): 98fbfb0

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -43
README.md CHANGED
@@ -13,66 +13,41 @@ industry-bert-sec-v0.1 is part of a series of industry-fine-tuned sentence_trans
13
 
14
  <!-- Provide a longer summary of what this model is. -->
15
 
16
- BERT-based 768-parameter drop-in substitute for non-industry-specific embeddings model. This model was trained on a wide range of
17
- publicly available U.S. Securities and Exchange Commission (SEC) regulatory filings and related documents.
18
 
19
  - **Developed by:** llmware
20
- - **Shared by [optional]:** Darren Oberst
21
  - **Model type:** BERT-based Industry domain fine-tuned Sentence Transformer architecture
22
  - **Language(s) (NLP):** English
23
  - **License:** Apache 2.0
24
  - **Finetuned from model [optional]:** BERT-based model, fine-tuning methodology described below.
25
 
26
- ### Model Sources [optional]
27
 
28
- <!-- Provide the basic links for the model. -->
29
 
30
- - **Repository:** [More Information Needed]
31
- - **Paper [optional]:** [More Information Needed]
32
- - **Demo [optional]:** [More Information Needed]
33
 
34
- ## Uses
35
-
36
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
37
-
38
- ### Direct Use
39
-
40
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
41
-
42
- This model is intended to be used as a sentence embedding model, specifically for financial services and use cases involving regulatory and financial filing documents.
43
-
44
- [More Information Needed]
45
-
46
- ### Downstream Use [optional]
47
-
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
-
50
- [More Information Needed]
51
-
52
- ### Out-of-Scope Use
53
-
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
-
56
- [More Information Needed]
57
 
58
  ## Bias, Risks, and Limitations
59
 
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
-
62
- [More Information Needed]
63
 
64
  ### Training Procedure
65
 
66
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
67
 
68
- This model was fine-tuned using a custom self-supervised procedure that combined contrastive techniques with stochastic injections of
69
- distortions in the samples. The methodology was derived, adapted and inspired primarily from three research papers cited below:
70
- TSDAE (Reimers), DeClutr (Giorgi), and Contrastive Tension (Carlsson).
71
 
72
 
73
  ## Citation [optional]
74
 
75
- Custom training protocol used to train the model, which was derived and inspired by the following papers:
76
 
77
  @article{wang-2021-TSDAE,
78
  title = "TSDAE: Using Transformer-based Sequential Denoising Auto-Encoderfor Unsupervised Sentence Embedding Learning",
@@ -107,12 +82,8 @@ Custom training protocol used to train the model, which was derived and inspired
107
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
108
 
109
 
110
- ## Model Card Authors [optional]
111
-
112
- [More Information Needed]
113
-
114
  ## Model Card Contact
115
 
116
- [More Information Needed]
117
 
118
 
 
13
 
14
  <!-- Provide a longer summary of what this model is. -->
15
 
16
+ industry-bert-sec-v0.1 is a domain fine-tuned BERT-based 768-parameter Sentence Transformer model, intended to as a "drop-in"
17
+ substitute for embeddings in financial and regulatory domains. This model was trained on a wide range of publicly available U.S. Securities and Exchange Commission (SEC) regulatory filings and related documents.
18
 
19
  - **Developed by:** llmware
 
20
  - **Model type:** BERT-based Industry domain fine-tuned Sentence Transformer architecture
21
  - **Language(s) (NLP):** English
22
  - **License:** Apache 2.0
23
  - **Finetuned from model [optional]:** BERT-based model, fine-tuning methodology described below.
24
 
25
+ ## Model Use
26
 
27
+ from transformers import AutoTokenizer, AutoModel
28
 
29
+ tokenizer = AutoTokenizer.from_pretrained("llmware/industry-bert-sec-v0.1")
30
+ model = AutoModel.from_pretrained("llmware/industry-bert-sec-v0.1")
 
31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
  ## Bias, Risks, and Limitations
34
 
35
+ This is a semantic embedding model, fine-tuned on public domain SEC filings and regulatory documents. Results may vary if used outside of this
36
+ domain, and like any embedding model, there is always the potential for anomalies in the vector embedding space. No specific safeguards have
37
+ put in place for safety or mitigate potential bias in the dataset.
38
 
39
  ### Training Procedure
40
 
41
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
42
 
43
+ This model was fine-tuned using a custom self-supervised procedure and custom dataset that combined contrastive techniques
44
+ with stochastic injections of distortions in the samples. The methodology was derived, adapted and inspired primarily from
45
+ three research papers cited below: TSDAE (Reimers), DeClutr (Giorgi), and Contrastive Tension (Carlsson).
46
 
47
 
48
  ## Citation [optional]
49
 
50
+ Custom self-supervised training protocol used to train the model, which was derived and inspired by the following papers:
51
 
52
  @article{wang-2021-TSDAE,
53
  title = "TSDAE: Using Transformer-based Sequential Denoising Auto-Encoderfor Unsupervised Sentence Embedding Learning",
 
82
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
83
 
84
 
 
 
 
 
85
  ## Model Card Contact
86
 
87
+ Darren Oberst @ llmware
88
 
89