Update inference parameters
Browse files
README.md
CHANGED
@@ -2,8 +2,7 @@
|
|
2 |
license: apache-2.0
|
3 |
base_model: princeton-nlp/Sheared-Pythia-160m
|
4 |
datasets:
|
5 |
-
- THUDM/webglm-qa
|
6 |
-
- Felladrin/formatted-csv-webglm-qa
|
7 |
widget:
|
8 |
- text: |-
|
9 |
<|references|>
|
@@ -17,8 +16,12 @@ widget:
|
|
17 |
<|answer|>
|
18 |
inference:
|
19 |
parameters:
|
20 |
-
max_new_tokens:
|
21 |
-
repetition_penalty: 1.
|
|
|
|
|
|
|
|
|
22 |
---
|
23 |
|
24 |
# Princeton NLP's Sheared-Pythia-160m trained on the WebGLM-QA dataset
|
@@ -34,4 +37,4 @@ inference:
|
|
34 |
<|question|>
|
35 |
{question}</s>
|
36 |
<|answer|>
|
37 |
-
```
|
|
|
2 |
license: apache-2.0
|
3 |
base_model: princeton-nlp/Sheared-Pythia-160m
|
4 |
datasets:
|
5 |
+
- THUDM/webglm-qa
|
|
|
6 |
widget:
|
7 |
- text: |-
|
8 |
<|references|>
|
|
|
16 |
<|answer|>
|
17 |
inference:
|
18 |
parameters:
|
19 |
+
max_new_tokens: 250
|
20 |
+
repetition_penalty: 1.025
|
21 |
+
do_sample: true
|
22 |
+
temperature: 0.4
|
23 |
+
top_p: 0.25
|
24 |
+
top_k: 7
|
25 |
---
|
26 |
|
27 |
# Princeton NLP's Sheared-Pythia-160m trained on the WebGLM-QA dataset
|
|
|
37 |
<|question|>
|
38 |
{question}</s>
|
39 |
<|answer|>
|
40 |
+
```
|