Adapters
medical
Laurent1 commited on
Commit
883658c
1 Parent(s): 032b543

Upload 11 files

Browse files
README.md CHANGED
@@ -1,24 +1,28 @@
1
  ---
2
- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
3
- # Doc / guide: https://huggingface.co/docs/hub/model-cards
4
- {}
5
  ---
6
 
7
  # Model Card for Model ID
8
 
9
  <!-- Provide a quick summary of what the model is/does. -->
10
 
11
- This model aims to provide a Question Answering model tuned with a short (128 tokens per row) Question Answering dataset
12
- The dataset enables fine tuning in local with small HW, such as 1 GPU with 16 Go RAM
13
 
14
 
15
  ## Model Details
16
 
17
- The model has been dowloaded from ibm/mpt-7b-instruct2 (Apache 2.0 License.) and tuned with Supervised Fine-tuning Trainer and PEFT LoRa
18
-
19
  ### Model Description
20
 
 
 
 
21
 
 
 
 
 
 
 
22
 
23
  ### Model Sources [optional]
24
 
@@ -28,91 +32,183 @@ The model has been dowloaded from ibm/mpt-7b-instruct2 (Apache 2.0 License.) and
28
  - **Paper [optional]:** [More Information Needed]
29
  - **Demo [optional]:** [More Information Needed]
30
 
 
 
 
31
 
32
  ### Direct Use
33
 
34
- text = "Below is an instruction from Human. Write a response.\n ### Instruction:\n How to diagnose Parasites - Baylisascaris infection ?\n ### Response:"
35
- inputs = tokenizer(text, return_tensors="pt").to('cuda')
36
- out = model.generate(**inputs, max_new_tokens=100)
37
 
38
- print(tokenizer.decode(out[0]))
39
 
 
 
 
 
 
 
 
 
 
40
 
41
  ## Bias, Risks, and Limitations
42
 
43
- In order to reduce training duration, the model has been trained only with the first 5100 rows of the 15500 rows dataset
44
 
45
- ### Recommendations
46
 
 
47
 
48
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
49
- Generation of plausible yet incorrect factual information, termed hallucination, is an unsolved issue in large language models.
50
 
 
51
 
52
  ## How to Get Started with the Model
53
 
54
  Use the code below to get started with the model.
55
 
56
- ## Training Details
57
 
58
- per_device_train_batch_size = 1
59
- gradient_accumulation_steps = 16
60
- epoch = 5
61
-
62
- Step Training Loss
63
- 64 1.618400
64
- 128 1.084200
65
- 192 1.021800
66
- 256 1.014300
67
- 320 0.960500
68
- 384 0.905900
69
- 448 0.885200
70
- 512 0.847400
71
- 576 0.889400
72
- 640 0.861000
73
- 704 0.800400
74
- 768 0.768600
75
- 832 0.750300
76
- 896 0.780200
77
- 960 0.762700
78
- 1024 0.698600
79
- 1088 0.672600
80
- 1152 0.693100
81
- 1216 0.708900
82
- 1280 0.662700
83
- 1344 0.630400
84
- 1408 0.624600
85
- 1472 0.627200
86
- 1536 0.628000
87
- 1600 0.603300
88
 
89
  ### Training Data
90
 
91
- Laurent1/MedQuad-MedicalQnADataset_128tokens_max
 
 
 
 
 
 
92
 
93
  #### Preprocessing [optional]
94
 
95
- Dataset already preprocessed (128 tokens max and truncated at a sentence end to keep meaning)
96
 
97
 
98
  #### Training Hyperparameters
99
 
 
 
 
100
 
101
- bnb_config = BitsAndBytesConfig(
102
- load_in_4bit=True,
103
- bnb_4bit_quant_type="nf4",
104
- bnb_4bit_compute_dtype=torch.float16,
105
- )
106
 
107
- model = AutoModelForCausalLM.from_pretrained(
108
- "ibm/mpt-7b-instruct2",
109
- device_map="auto",
110
- torch_dtype=torch.float16, #torch.bfloat16,
111
- trust_remote_code=True
112
- )
113
 
114
- #### Speeds, Sizes, Times [optional]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
115
 
116
- Training :
117
 
118
- 6287.4s - GPU T4 x2
 
1
  ---
2
+ library_name: peft
3
+ base_model: ibm/mpt-7b-instruct2
 
4
  ---
5
 
6
  # Model Card for Model ID
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
 
 
10
 
11
 
12
  ## Model Details
13
 
 
 
14
  ### Model Description
15
 
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
 
20
+ - **Developed by:** [More Information Needed]
21
+ - **Shared by [optional]:** [More Information Needed]
22
+ - **Model type:** [More Information Needed]
23
+ - **Language(s) (NLP):** [More Information Needed]
24
+ - **License:** [More Information Needed]
25
+ - **Finetuned from model [optional]:** [More Information Needed]
26
 
27
  ### Model Sources [optional]
28
 
 
32
  - **Paper [optional]:** [More Information Needed]
33
  - **Demo [optional]:** [More Information Needed]
34
 
35
+ ## Uses
36
+
37
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
38
 
39
  ### Direct Use
40
 
41
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
42
+
43
+ [More Information Needed]
44
 
45
+ ### Downstream Use [optional]
46
 
47
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Out-of-Scope Use
52
+
53
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
54
+
55
+ [More Information Needed]
56
 
57
  ## Bias, Risks, and Limitations
58
 
59
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
60
 
61
+ [More Information Needed]
62
 
63
+ ### Recommendations
64
 
65
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
 
66
 
67
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
68
 
69
  ## How to Get Started with the Model
70
 
71
  Use the code below to get started with the model.
72
 
73
+ [More Information Needed]
74
 
75
+ ## Training Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
 
77
  ### Training Data
78
 
79
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
80
+
81
+ [More Information Needed]
82
+
83
+ ### Training Procedure
84
+
85
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
86
 
87
  #### Preprocessing [optional]
88
 
89
+ [More Information Needed]
90
 
91
 
92
  #### Training Hyperparameters
93
 
94
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
95
+
96
+ #### Speeds, Sizes, Times [optional]
97
 
98
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
 
 
 
 
99
 
100
+ [More Information Needed]
 
 
 
 
 
101
 
102
+ ## Evaluation
103
+
104
+ <!-- This section describes the evaluation protocols and provides the results. -->
105
+
106
+ ### Testing Data, Factors & Metrics
107
+
108
+ #### Testing Data
109
+
110
+ <!-- This should link to a Data Card if possible. -->
111
+
112
+ [More Information Needed]
113
+
114
+ #### Factors
115
+
116
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Metrics
121
+
122
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
123
+
124
+ [More Information Needed]
125
+
126
+ ### Results
127
+
128
+ [More Information Needed]
129
+
130
+ #### Summary
131
+
132
+
133
+
134
+ ## Model Examination [optional]
135
+
136
+ <!-- Relevant interpretability work for the model goes here -->
137
+
138
+ [More Information Needed]
139
+
140
+ ## Environmental Impact
141
+
142
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
143
+
144
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
145
+
146
+ - **Hardware Type:** [More Information Needed]
147
+ - **Hours used:** [More Information Needed]
148
+ - **Cloud Provider:** [More Information Needed]
149
+ - **Compute Region:** [More Information Needed]
150
+ - **Carbon Emitted:** [More Information Needed]
151
+
152
+ ## Technical Specifications [optional]
153
+
154
+ ### Model Architecture and Objective
155
+
156
+ [More Information Needed]
157
+
158
+ ### Compute Infrastructure
159
+
160
+ [More Information Needed]
161
+
162
+ #### Hardware
163
+
164
+ [More Information Needed]
165
+
166
+ #### Software
167
+
168
+ [More Information Needed]
169
+
170
+ ## Citation [optional]
171
+
172
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
173
+
174
+ **BibTeX:**
175
+
176
+ [More Information Needed]
177
+
178
+ **APA:**
179
+
180
+ [More Information Needed]
181
+
182
+ ## Glossary [optional]
183
+
184
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
185
+
186
+ [More Information Needed]
187
+
188
+ ## More Information [optional]
189
+
190
+ [More Information Needed]
191
+
192
+ ## Model Card Authors [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Contact
197
+
198
+ [More Information Needed]
199
+
200
+
201
+ ## Training procedure
202
+
203
+
204
+ ### Framework versions
205
+
206
+
207
+ - PEFT 0.6.0
208
+ ## Training procedure
209
+
210
+
211
+ ### Framework versions
212
 
 
213
 
214
+ - PEFT 0.6.0
adapter_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "ibm/mpt-7b-instruct2",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "lora_alpha": 16,
12
+ "lora_dropout": 0.1,
13
+ "modules_to_save": null,
14
+ "peft_type": "LORA",
15
+ "r": 32,
16
+ "rank_pattern": {},
17
+ "revision": null,
18
+ "target_modules": [
19
+ "Wqkv",
20
+ "up_proj",
21
+ "out_proj",
22
+ "down_proj"
23
+ ],
24
+ "task_type": "CAUSAL_LM"
25
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca46e1175b341953e55194e308775bc2b92762c34d878554c1cf5908a0250a93
3
+ size 268528333
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:409a3e0c87eca2838b4e3b24404caad915c761ba573cdb33903a5474152df9ac
3
+ size 537029125
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e028d427da0864de9c9bb6b990581599720abaaccea53dee180d2d7affd8eded
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01d01d4ece2a97b3018abb37144598fd8cdffd408b84edf962b287a9d8c9a7b4
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<|endoftext|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|endoftext|>",
6
+ "max_length": 2048,
7
+ "model_max_length": 2048,
8
+ "pad_to_multiple_of": null,
9
+ "pad_token": "<|endoftext|>",
10
+ "pad_token_type_id": 0,
11
+ "padding_side": "right",
12
+ "stride": 0,
13
+ "tokenizer_class": "GPTNeoXTokenizer",
14
+ "truncation_side": "right",
15
+ "truncation_strategy": "longest_first",
16
+ "unk_token": "<|endoftext|>"
17
+ }
trainer_state.json ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1600,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.2,
13
+ "learning_rate": 9.896907216494846e-05,
14
+ "loss": 1.6184,
15
+ "step": 64
16
+ },
17
+ {
18
+ "epoch": 0.4,
19
+ "learning_rate": 9.484536082474227e-05,
20
+ "loss": 1.0842,
21
+ "step": 128
22
+ },
23
+ {
24
+ "epoch": 0.6,
25
+ "learning_rate": 9.072164948453609e-05,
26
+ "loss": 1.0218,
27
+ "step": 192
28
+ },
29
+ {
30
+ "epoch": 0.8,
31
+ "learning_rate": 8.65979381443299e-05,
32
+ "loss": 1.0143,
33
+ "step": 256
34
+ },
35
+ {
36
+ "epoch": 1.0,
37
+ "learning_rate": 8.247422680412371e-05,
38
+ "loss": 0.9605,
39
+ "step": 320
40
+ },
41
+ {
42
+ "epoch": 1.2,
43
+ "learning_rate": 7.835051546391753e-05,
44
+ "loss": 0.9059,
45
+ "step": 384
46
+ },
47
+ {
48
+ "epoch": 1.4,
49
+ "learning_rate": 7.422680412371135e-05,
50
+ "loss": 0.8852,
51
+ "step": 448
52
+ },
53
+ {
54
+ "epoch": 1.6,
55
+ "learning_rate": 7.010309278350515e-05,
56
+ "loss": 0.8474,
57
+ "step": 512
58
+ },
59
+ {
60
+ "epoch": 1.8,
61
+ "learning_rate": 6.597938144329897e-05,
62
+ "loss": 0.8894,
63
+ "step": 576
64
+ },
65
+ {
66
+ "epoch": 2.0,
67
+ "learning_rate": 6.185567010309279e-05,
68
+ "loss": 0.861,
69
+ "step": 640
70
+ },
71
+ {
72
+ "epoch": 2.2,
73
+ "learning_rate": 5.7731958762886594e-05,
74
+ "loss": 0.8004,
75
+ "step": 704
76
+ },
77
+ {
78
+ "epoch": 2.4,
79
+ "learning_rate": 5.360824742268041e-05,
80
+ "loss": 0.7686,
81
+ "step": 768
82
+ },
83
+ {
84
+ "epoch": 2.6,
85
+ "learning_rate": 4.948453608247423e-05,
86
+ "loss": 0.7503,
87
+ "step": 832
88
+ },
89
+ {
90
+ "epoch": 2.8,
91
+ "learning_rate": 4.536082474226804e-05,
92
+ "loss": 0.7802,
93
+ "step": 896
94
+ },
95
+ {
96
+ "epoch": 3.0,
97
+ "learning_rate": 4.1237113402061855e-05,
98
+ "loss": 0.7627,
99
+ "step": 960
100
+ },
101
+ {
102
+ "epoch": 3.2,
103
+ "learning_rate": 3.7113402061855674e-05,
104
+ "loss": 0.6986,
105
+ "step": 1024
106
+ },
107
+ {
108
+ "epoch": 3.4,
109
+ "learning_rate": 3.2989690721649485e-05,
110
+ "loss": 0.6726,
111
+ "step": 1088
112
+ },
113
+ {
114
+ "epoch": 3.6,
115
+ "learning_rate": 2.8865979381443297e-05,
116
+ "loss": 0.6931,
117
+ "step": 1152
118
+ },
119
+ {
120
+ "epoch": 3.8,
121
+ "learning_rate": 2.4742268041237116e-05,
122
+ "loss": 0.7089,
123
+ "step": 1216
124
+ },
125
+ {
126
+ "epoch": 4.0,
127
+ "learning_rate": 2.0618556701030927e-05,
128
+ "loss": 0.6627,
129
+ "step": 1280
130
+ },
131
+ {
132
+ "epoch": 4.2,
133
+ "learning_rate": 1.6494845360824743e-05,
134
+ "loss": 0.6304,
135
+ "step": 1344
136
+ },
137
+ {
138
+ "epoch": 4.4,
139
+ "learning_rate": 1.2371134020618558e-05,
140
+ "loss": 0.6246,
141
+ "step": 1408
142
+ },
143
+ {
144
+ "epoch": 4.6,
145
+ "learning_rate": 8.247422680412371e-06,
146
+ "loss": 0.6272,
147
+ "step": 1472
148
+ },
149
+ {
150
+ "epoch": 4.8,
151
+ "learning_rate": 4.123711340206186e-06,
152
+ "loss": 0.628,
153
+ "step": 1536
154
+ },
155
+ {
156
+ "epoch": 5.0,
157
+ "learning_rate": 0.0,
158
+ "loss": 0.6033,
159
+ "step": 1600
160
+ }
161
+ ],
162
+ "logging_steps": 64,
163
+ "max_steps": 1600,
164
+ "num_train_epochs": 5,
165
+ "save_steps": 64,
166
+ "total_flos": 1.0172436457734144e+17,
167
+ "trial_name": null,
168
+ "trial_params": null
169
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff167a52ec01ecc304782bbe68032336f2138182867621872aafe714497cc79b
3
+ size 4027