inkyant commited on
Commit
0e46a0b
β€’
1 Parent(s): 991f437

add easy inference

Browse files
README.md CHANGED
@@ -13,9 +13,6 @@ model-index:
13
  results: []
14
  ---
15
 
16
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
- should probably proofread and complete it, then remove this comment. -->
18
-
19
  # gemma-2b-dolly-qa
20
 
21
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the generator dataset.
 
13
  results: []
14
  ---
15
 
 
 
 
16
  # gemma-2b-dolly-qa
17
 
18
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on the generator dataset.
adapter_config.json β†’ adapter_model/adapter_config.json RENAMED
File without changes
adapter_model.safetensors β†’ adapter_model/adapter_model.safetensors RENAMED
File without changes
run.py ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch, sys
2
+ from peft import PeftModel
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
+
5
+ import warnings
6
+ warnings.filterwarnings("ignore")
7
+
8
+ access_token = sys.argv[2]
9
+ device = "xpu:0" if sys.argv[1] == "gpu" else "cpu:0"
10
+
11
+ tokenizer = AutoTokenizer.from_pretrained("./tokenizer/")
12
+ tokenizer.pad_token = tokenizer.eos_token
13
+ tokenizer.padding_side = "right"
14
+
15
+ base_model = AutoModelForCausalLM.from_pretrained(
16
+ "google/gemma-2b",
17
+ token=access_token,
18
+ low_cpu_mem_usage=True,
19
+ return_dict=True,
20
+ torch_dtype=torch.bfloat16,
21
+ )
22
+
23
+ model = PeftModel.from_pretrained(base_model, "adapter_model")
24
+ model = model.to(device)
25
+
26
+ print("Prompt:", " ".join(sys.argv[3:]))
27
+
28
+ inputs = tokenizer(" ".join(sys.argv[3:]), return_tensors="pt").to(device)
29
+ outputs = model.generate(**inputs, max_new_tokens=200,
30
+ do_sample=False, top_k=100,temperature=0.1,
31
+ eos_token_id=tokenizer.eos_token_id)
32
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
special_tokens_map.json β†’ tokenizer/special_tokens_map.json RENAMED
File without changes
tokenizer.json β†’ tokenizer/tokenizer.json RENAMED
File without changes
tokenizer.model β†’ tokenizer/tokenizer.model RENAMED
File without changes
tokenizer_config.json β†’ tokenizer/tokenizer_config.json RENAMED
File without changes