pvduy commited on
Commit
010f2b8
·
1 Parent(s): 02d631f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Inference code:
2
+ ```python
3
+ from datasets import load_dataset
4
+
5
+ from transformers import AutoTokenizer, AutoModelForCausalLM
6
+
7
+ dataset = load_dataset("CarperAI/openai_summarize_tldr")
8
+
9
+ val_prompts = [sample["prompt"] for sample in dataset["valid"]]
10
+
11
+ kwargs = {
12
+ "max_new_tokens": 50,
13
+ "do_sample": True,
14
+ "top_k": 0,
15
+ "top_p": 1,
16
+ }
17
+
18
+ model = AutoModelForCausalLM.from_pretrained("pvduy/ppo_pythia6B_sample")
19
+ model.eval()
20
+ tokenizer = AutoTokenizer.from_pretrained("pvduy/ppo_pythia6B_sample")
21
+ tokenizer.pad_token_id = tokenizer.eos_token_id
22
+
23
+ count = 0
24
+
25
+ for prompt in val_prompts:
26
+ output_tk = tokenizer(prompt, return_tensors="pt")
27
+ outputs = model.generate(output_tk.input_ids, attention_mask=output_tk.attention_mask, **kwargs)
28
+ print("Prompt:", prompt)
29
+ print("Output:", tokenizer.decode(outputs[0], skip_special_tokens=True).split("TL;DR:")[1].strip())
30
+ print("=================================")
31
+ count += 1
32
+ if count == 10:
33
+ break
34
+ ```