michaelfeil commited on
Commit
368b912
1 Parent(s): 71be13b

Upload bigcode/gpt_bigcode-santacoder ctranslate fp16 weights

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -173,9 +173,9 @@ Speedup inference while reducing memory by 2x-4x using int8 inference in C++ on
173
 
174
  quantized version of [bigcode/gpt_bigcode-santacoder](https://huggingface.co/bigcode/gpt_bigcode-santacoder)
175
  ```bash
176
- pip install hf-hub-ctranslate2>=2.0.8
177
  ```
178
- Converted on 2023-05-30 using
179
  ```
180
  ct2-transformers-converter --model bigcode/gpt_bigcode-santacoder --output_dir /home/michael/tmp-ct2fast-gpt_bigcode-santacoder --force --copy_files tokenizer.json README.md tokenizer_config.json special_tokens_map.json .gitattributes --quantization float16 --trust_remote_code
181
  ```
@@ -199,7 +199,8 @@ model = GeneratorCT2fromHfHub(
199
  )
200
  outputs = model.generate(
201
  text=["How do you call a fast Flan-ingo?", "User: How are you doing? Bot:"],
202
- max_length=64
 
203
  )
204
  print(outputs)
205
  ```
 
173
 
174
  quantized version of [bigcode/gpt_bigcode-santacoder](https://huggingface.co/bigcode/gpt_bigcode-santacoder)
175
  ```bash
176
+ pip install hf-hub-ctranslate2>=2.0.8 ctranslate2>=3.14.0
177
  ```
178
+ Converted on 2023-05-31 using
179
  ```
180
  ct2-transformers-converter --model bigcode/gpt_bigcode-santacoder --output_dir /home/michael/tmp-ct2fast-gpt_bigcode-santacoder --force --copy_files tokenizer.json README.md tokenizer_config.json special_tokens_map.json .gitattributes --quantization float16 --trust_remote_code
181
  ```
 
199
  )
200
  outputs = model.generate(
201
  text=["How do you call a fast Flan-ingo?", "User: How are you doing? Bot:"],
202
+ max_length=64,
203
+ include_prompt_in_result=False
204
  )
205
  print(outputs)
206
  ```