sanagnos commited on
Commit
618f20f
·
1 Parent(s): 3277d74

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Galactica-6.7b finetuned on webgpt and prompt_dialogue (version v2)
2
+
3
+ Demo use:
4
+
5
+ ```
6
+ import torch
7
+ from torch import nn
8
+ from torch.nn import functional as F
9
+ import transformers
10
+
11
+ base_path = 'sanagnos/galactica-6.7b-finetuned'
12
+
13
+ model = transformers.OPTForCausalLM.from_pretrained(
14
+ base_path, load_in_8bit=True, device_map='auto', low_cpu_mem_usage=True,
15
+ torch_dtype=torch.float16, offload_state_dict=True
16
+ )
17
+
18
+ model.gradient_checkpointing_enable() # reduce number of stored activations
19
+ model.model.decoder.project_in = lambda x: x.requires_grad_(True)
20
+
21
+ class CastOutputToFloat(nn.Sequential):
22
+ def forward(self, x): return super().forward(x).to(torch.float32)
23
+
24
+ model.lm_head = CastOutputToFloat(model.lm_head)
25
+
26
+ tokenizer = transformers.AutoTokenizer.from_pretrained(base_path)
27
+
28
+ batch = tokenizer.encode("<question>What are the symptoms of Alzheimer's disease?<answer>", return_tensors="pt")
29
+
30
+ with torch.cuda.amp.autocast():
31
+ out = model.generate(
32
+ input_ids=batch.to(model.device),
33
+ max_length=300,
34
+ do_sample=True,
35
+ top_k=40,
36
+ num_beams=1,
37
+ num_return_sequences=1,
38
+ eos_token_id=tokenizer.additional_special_tokens_ids[tokenizer.additional_special_tokens.index('<question>')]
39
+ )
40
+
41
+ print(tokenizer.decode(out[0, :-1]).replace('<question>', "User:\n").replace('<answer>', '\nAssistant:\n'))
42
+ ```