dmayhem93 commited on
Commit
77ef3fd
1 Parent(s): b335bf4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -23,16 +23,16 @@ Start chatting with `FreeWilly2` using the following code snippet:
23
  import torch
24
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
25
 
26
- tokenizer = AutoTokenizer.from_pretrained("stabilityai/FreeWilly2, use_fast=False)
27
- model = AutoModelForCausalLM.from_pretrained("stabilityai/FreeWilly2", torch_dtype=torch.float16, low_cpu_mem_usage=True, use_accelerate=True)
28
- generator = pipeline(model=model, tokenizer=tokenizer)
29
  system_prompt = "### System:\nYou are Free Willy, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
30
 
31
  message = "Write me a poem please"
32
  prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
 
 
33
 
34
- output = generator(prompt, num_return_sequences=1, do_sample=True, top_p=0.95, top_k=0)
35
- print(output)
36
  ```
37
 
38
  FreeWilly should be used with this prompt format:
@@ -102,4 +102,3 @@ Although the aforementioned dataset helps to steer the base language models into
102
  primaryClass={cs.CL}
103
  }
104
  ```
105
-
 
23
  import torch
24
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
25
 
26
+ tokenizer = AutoTokenizer.from_pretrained("stabilityai/FreeWilly2", use_fast=False)
27
+ model = AutoModelForCausalLM.from_pretrained("stabilityai/FreeWilly2", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
 
28
  system_prompt = "### System:\nYou are Free Willy, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
29
 
30
  message = "Write me a poem please"
31
  prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
32
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
33
+ output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_tokens=256)
34
 
35
+ print(tokenizer.decode(output[0], skip_special_tokens=True))
 
36
  ```
37
 
38
  FreeWilly should be used with this prompt format:
 
102
  primaryClass={cs.CL}
103
  }
104
  ```