nicholascao commited on
Commit
6d0b3bc
1 Parent(s): c4b7b21

update model

Browse files
Files changed (3) hide show
  1. README.md +5 -5
  2. generation_config.json +1 -1
  3. pytorch_model.bin +2 -2
README.md CHANGED
@@ -12,13 +12,13 @@ pipeline_tag: text-generation
12
  tags:
13
  - chat
14
  widget:
15
- - text: "<Human>: Hello <eoh> <Assistant>: "
16
  example_title: "Hello"
17
- - text: "<Human>: 你好 <eoh> <Assistant>: "
18
  example_title: "你好"
19
- - text: "<Human>: What should I do if I can't sleep at night? <eoh> <Assistant>: "
20
  example_title: "insomnia"
21
- - text: "<Human>: 晚上睡不着应该怎么办? <eoh> <Assistant>: "
22
  example_title: "失眠"
23
  inference:
24
  parameters:
@@ -42,7 +42,7 @@ tokenizer = AutoTokenizer.from_pretrained('nicholascao/chatbloom-1b7-sft')
42
  model = AutoModelForCausalLM.from_pretrained('nicholascao/chatbloom-1b7-sft').half()
43
  generation_config = GenerationConfig.from_pretrained('nicholascao/chatbloom-1b7-sft')
44
 
45
- inputs = tokenizer('<Human>: Hello <eoh> <Assistant>: ', return_tensors='pt').to(torch.cuda.current_device())
46
  model.to(torch.cuda.current_device())
47
 
48
  output = model.generate(**inputs, generation_config=generation_config)
 
12
  tags:
13
  - chat
14
  widget:
15
+ - text: "<Human>: Hello <eoh> <Assistant>:"
16
  example_title: "Hello"
17
+ - text: "<Human>: 你好 <eoh> <Assistant>:"
18
  example_title: "你好"
19
+ - text: "<Human>: What should I do if I can't sleep at night? <eoh> <Assistant>:"
20
  example_title: "insomnia"
21
+ - text: "<Human>: 晚上睡不着应该怎么办? <eoh> <Assistant>:"
22
  example_title: "失眠"
23
  inference:
24
  parameters:
 
42
  model = AutoModelForCausalLM.from_pretrained('nicholascao/chatbloom-1b7-sft').half()
43
  generation_config = GenerationConfig.from_pretrained('nicholascao/chatbloom-1b7-sft')
44
 
45
+ inputs = tokenizer('<Human>: Hello <eoh> <Assistant>:', return_tensors='pt').to(torch.cuda.current_device())
46
  model.to(torch.cuda.current_device())
47
 
48
  output = model.generate(**inputs, generation_config=generation_config)
generation_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "do_sample": true,
5
  "early_stopping": true,
6
  "eos_token_id": 2,
7
- "max_new_tokens": 384,
8
  "pad_token_id": 2,
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.8,
 
4
  "do_sample": true,
5
  "early_stopping": true,
6
  "eos_token_id": 2,
7
+ "max_new_tokens": 512,
8
  "pad_token_id": 2,
9
  "repetition_penalty": 1.1,
10
  "temperature": 0.8,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0f55ca3cd52f4220e9046d88293c8b8cf65ea17e17c1c35a60c58941651a7fb
3
- size 3444055647
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076ce6313d3e4895c711440cf34ee5205223b9bd26286fa2fa41e9ce2f92579d
3
+ size 3444055875