ctranslate2-4you commited on
Commit
ab7c9b6
·
verified ·
1 Parent(s): dc5c3a3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -58,19 +58,19 @@ import ctranslate2
58
  from transformers import AutoTokenizer
59
  import sys
60
 
61
- def generate_response(prompt, system_message, model_path):
62
  generator = ctranslate2.Generator(model_path, device="cuda", compute_type="int8")
63
  tokenizer = AutoTokenizer.from_pretrained(model_path)
64
  formatted_prompt = f"""<|im_start|>system<|im_sep|>{system_message}<|im_end|>
65
  <|im_start|>user<|im_sep|>{prompt}<|im_end|>
66
  <|im_start|>assistant<|im_sep|>"""
67
- tokens = tokenizer.tokenize(formatted_prompt)
68
  for step in generator.generate_tokens([tokens], max_length=1024, sampling_temperature=0.7):
69
- token = step.token
70
- decoded_token = tokenizer.decode([step.token_id])
71
- print(decoded_token, end="", flush=True)
72
  if token in tokenizer.eos_token or token in tokenizer.all_special_tokens:
73
  break
 
 
74
 
75
  if __name__ == "__main__":
76
  model_path = "path/to/your/phi-4-ct2-model"
@@ -78,6 +78,5 @@ if __name__ == "__main__":
78
  user_prompt = "Write a short poem about a cat."
79
  print("\nGenerating response:")
80
  generate_response(user_prompt, system_message, model_path)
81
-
82
  ```
83
  </details>
 
58
  from transformers import AutoTokenizer
59
  import sys
60
 
61
+ def generate_response(prompt: str, system_message: str, model_path: str) -> None:
62
  generator = ctranslate2.Generator(model_path, device="cuda", compute_type="int8")
63
  tokenizer = AutoTokenizer.from_pretrained(model_path)
64
  formatted_prompt = f"""<|im_start|>system<|im_sep|>{system_message}<|im_end|>
65
  <|im_start|>user<|im_sep|>{prompt}<|im_end|>
66
  <|im_start|>assistant<|im_sep|>"""
67
+ tokens: List[str] = tokenizer.tokenize(formatted_prompt)
68
  for step in generator.generate_tokens([tokens], max_length=1024, sampling_temperature=0.7):
69
+ token: str = step.token
 
 
70
  if token in tokenizer.eos_token or token in tokenizer.all_special_tokens:
71
  break
72
+ decoded_token: str = tokenizer.decode([step.token_id])
73
+ print(decoded_token, end="", flush=True)
74
 
75
  if __name__ == "__main__":
76
  model_path = "path/to/your/phi-4-ct2-model"
 
78
  user_prompt = "Write a short poem about a cat."
79
  print("\nGenerating response:")
80
  generate_response(user_prompt, system_message, model_path)
 
81
  ```
82
  </details>