arnocandel commited on
Commit
6b732b5
1 Parent(s): d1aea17

commit files to HF hub

Browse files
Files changed (1) hide show
  1. README.md +10 -17
README.md CHANGED
@@ -11,8 +11,7 @@ tags:
11
  - large language model
12
  - open-source
13
  datasets:
14
- - h2oai/openassistant_oasst1
15
- - h2oai/openassistant_oasst1_h2ogpt
16
  ---
17
  # h2oGPT Model Card
18
  ## Summary
@@ -20,9 +19,9 @@ datasets:
20
  H2O.ai's `h2ogpt-oasst1-512-12b` is a 12 billion parameter instruction-following large language model licensed for commercial use.
21
 
22
  - Base model: [EleutherAI/pythia-12b](https://huggingface.co/EleutherAI/pythia-12b)
23
- - Fine-tuning dataset: [h2oai/openassistant_oasst1](https://huggingface.co/datasets/h2oai/openassistant_oasst1) and [h2oai/openassistant_oasst1_h2ogpt](https://huggingface.co/datasets/h2oai/openassistant_oasst1_h2ogpt)
24
  - Data-prep and fine-tuning code: [H2O.ai GitHub](https://github.com/h2oai/h2ogpt)
25
- - Training logs: [zip](https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/pythia-12b.openassistant_oasst1.json.1_epochs.d45a9d34d34534e076cc6797614b322bd0efb11c.15.zip) and [zip](https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/h2ogpt-oasst1-512-12b.h2oaiopenassistant_oasst1_h2ogpt.2_epochs.fcaae7ef70600de8c97c9b38cb3f0075467cdad1.2.zip)
26
 
27
  ## Chatbot
28
 
@@ -107,12 +106,6 @@ GPTNeoXConfig {
107
  "pt": "AutoModelForCausalLM"
108
  }
109
  },
110
- "custom_pipelines": {
111
- "text-generation": {
112
- "impl": "h2oai_pipeline.H2OTextGenerationPipeline",
113
- "pt": "AutoModelForCausalLM"
114
- }
115
- },
116
  "eos_token_id": 0,
117
  "hidden_act": "gelu",
118
  "hidden_size": 5120,
@@ -140,22 +133,22 @@ GPTNeoXConfig {
140
  Model validation results using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
141
 
142
 
143
- [eval source code](https://github.com/h2oai/h2ogpt/issues/35#issuecomment-1521119301)
144
 
145
  | Task |Version| Metric |Value | |Stderr|
146
  |-------------|------:|--------|-----:|---|-----:|
 
 
147
  |arc_easy | 0|acc |0.6932|± |0.0095|
148
  | | |acc_norm|0.6225|± |0.0099|
149
- |openbookqa | 0|acc |0.2900|± |0.0203|
150
- | | |acc_norm|0.3740|± |0.0217|
151
- |winogrande | 0|acc |0.6369|± |0.0135|
152
  |hellaswag | 0|acc |0.5140|± |0.0050|
153
  | | |acc_norm|0.6803|± |0.0047|
 
 
154
  |piqa | 0|acc |0.7682|± |0.0098|
155
  | | |acc_norm|0.7661|± |0.0099|
156
- |boolq | 1|acc |0.6685|± |0.0082|
157
- |arc_challenge| 0|acc |0.3157|± |0.0136|
158
- | | |acc_norm|0.3507|± |0.0139|
159
 
160
 
161
  ## Disclaimer
 
11
  - large language model
12
  - open-source
13
  datasets:
14
+ - h2oai/openassistant_oasst1_h2ogpt_graded
 
15
  ---
16
  # h2oGPT Model Card
17
  ## Summary
 
19
  H2O.ai's `h2ogpt-oasst1-512-12b` is a 12 billion parameter instruction-following large language model licensed for commercial use.
20
 
21
  - Base model: [EleutherAI/pythia-12b](https://huggingface.co/EleutherAI/pythia-12b)
22
+ - Fine-tuning dataset: [h2oai/openassistant_oasst1_h2ogpt_graded](https://huggingface.co/datasets/h2oai/openassistant_oasst1_h2ogpt_graded)
23
  - Data-prep and fine-tuning code: [H2O.ai GitHub](https://github.com/h2oai/h2ogpt)
24
+ - Training logs: [zip](https://huggingface.co/h2oai/h2ogpt-oasst1-512-12b/blob/main/pythia-12b-deduped.h2oaiopenassistant_oasst1_h2ogpt_graded.3_epochs.2ccf687ea3f3f3775a501838e81c1a0066430455.4.zip)
25
 
26
  ## Chatbot
27
 
 
106
  "pt": "AutoModelForCausalLM"
107
  }
108
  },
 
 
 
 
 
 
109
  "eos_token_id": 0,
110
  "hidden_act": "gelu",
111
  "hidden_size": 5120,
 
133
  Model validation results using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness).
134
 
135
 
136
+ [eval source code](https://github.com/h2oai/h2ogpt/issues/125#issuecomment-1548239108)
137
 
138
  | Task |Version| Metric |Value | |Stderr|
139
  |-------------|------:|--------|-----:|---|-----:|
140
+ |arc_challenge| 0|acc |0.3157|± |0.0136|
141
+ | | |acc_norm|0.3507|± |0.0139|
142
  |arc_easy | 0|acc |0.6932|± |0.0095|
143
  | | |acc_norm|0.6225|± |0.0099|
144
+ |boolq | 1|acc |0.6685|± |0.0082|
 
 
145
  |hellaswag | 0|acc |0.5140|± |0.0050|
146
  | | |acc_norm|0.6803|± |0.0047|
147
+ |openbookqa | 0|acc |0.2900|± |0.0203|
148
+ | | |acc_norm|0.3740|± |0.0217|
149
  |piqa | 0|acc |0.7682|± |0.0098|
150
  | | |acc_norm|0.7661|± |0.0099|
151
+ |winogrande | 0|acc |0.6369|± |0.0135|
 
 
152
 
153
 
154
  ## Disclaimer