nicholascao
commited on
Commit
•
6d0b3bc
1
Parent(s):
c4b7b21
update model
Browse files- README.md +5 -5
- generation_config.json +1 -1
- pytorch_model.bin +2 -2
README.md
CHANGED
@@ -12,13 +12,13 @@ pipeline_tag: text-generation
|
|
12 |
tags:
|
13 |
- chat
|
14 |
widget:
|
15 |
-
- text: "<Human>: Hello <eoh> <Assistant>:
|
16 |
example_title: "Hello"
|
17 |
-
- text: "<Human>: 你好 <eoh> <Assistant>:
|
18 |
example_title: "你好"
|
19 |
-
- text: "<Human>: What should I do if I can't sleep at night? <eoh> <Assistant>:
|
20 |
example_title: "insomnia"
|
21 |
-
- text: "<Human>: 晚上睡不着应该怎么办? <eoh> <Assistant>:
|
22 |
example_title: "失眠"
|
23 |
inference:
|
24 |
parameters:
|
@@ -42,7 +42,7 @@ tokenizer = AutoTokenizer.from_pretrained('nicholascao/chatbloom-1b7-sft')
|
|
42 |
model = AutoModelForCausalLM.from_pretrained('nicholascao/chatbloom-1b7-sft').half()
|
43 |
generation_config = GenerationConfig.from_pretrained('nicholascao/chatbloom-1b7-sft')
|
44 |
|
45 |
-
inputs = tokenizer('<Human>: Hello <eoh> <Assistant>:
|
46 |
model.to(torch.cuda.current_device())
|
47 |
|
48 |
output = model.generate(**inputs, generation_config=generation_config)
|
|
|
12 |
tags:
|
13 |
- chat
|
14 |
widget:
|
15 |
+
- text: "<Human>: Hello <eoh> <Assistant>:"
|
16 |
example_title: "Hello"
|
17 |
+
- text: "<Human>: 你好 <eoh> <Assistant>:"
|
18 |
example_title: "你好"
|
19 |
+
- text: "<Human>: What should I do if I can't sleep at night? <eoh> <Assistant>:"
|
20 |
example_title: "insomnia"
|
21 |
+
- text: "<Human>: 晚上睡不着应该怎么办? <eoh> <Assistant>:"
|
22 |
example_title: "失眠"
|
23 |
inference:
|
24 |
parameters:
|
|
|
42 |
model = AutoModelForCausalLM.from_pretrained('nicholascao/chatbloom-1b7-sft').half()
|
43 |
generation_config = GenerationConfig.from_pretrained('nicholascao/chatbloom-1b7-sft')
|
44 |
|
45 |
+
inputs = tokenizer('<Human>: Hello <eoh> <Assistant>:', return_tensors='pt').to(torch.cuda.current_device())
|
46 |
model.to(torch.cuda.current_device())
|
47 |
|
48 |
output = model.generate(**inputs, generation_config=generation_config)
|
generation_config.json
CHANGED
@@ -4,7 +4,7 @@
|
|
4 |
"do_sample": true,
|
5 |
"early_stopping": true,
|
6 |
"eos_token_id": 2,
|
7 |
-
"max_new_tokens":
|
8 |
"pad_token_id": 2,
|
9 |
"repetition_penalty": 1.1,
|
10 |
"temperature": 0.8,
|
|
|
4 |
"do_sample": true,
|
5 |
"early_stopping": true,
|
6 |
"eos_token_id": 2,
|
7 |
+
"max_new_tokens": 512,
|
8 |
"pad_token_id": 2,
|
9 |
"repetition_penalty": 1.1,
|
10 |
"temperature": 0.8,
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:076ce6313d3e4895c711440cf34ee5205223b9bd26286fa2fa41e9ce2f92579d
|
3 |
+
size 3444055875
|