Update README.md
Browse files
README.md
CHANGED
@@ -46,8 +46,7 @@ Inference using:
|
|
46 |
<span style="color: #0000FF;">with</span> torch.cuda.amp.autocast():
|
47 |
output_tokens = model.generate(**batch, max_new_tokens=<span style="color: #098658;">50</span>)
|
48 |
|
49 |
-
<span style="color: #0000FF;">print</span>(
|
50 |
-
', tokenizer.decode(output_tokens[<span style="color: #098658;">0</span>], skip_special_tokens=<span style="color: #0000FF;">True</span>))
|
51 |
</code>
|
52 |
</pre>
|
53 |
|
|
|
46 |
<span style="color: #0000FF;">with</span> torch.cuda.amp.autocast():
|
47 |
output_tokens = model.generate(**batch, max_new_tokens=<span style="color: #098658;">50</span>)
|
48 |
|
49 |
+
<span style="color: #0000FF;">print</span>(<span style="color: #A31515;">"\n\n"</span>, tokenizer.decode(output_tokens[<span style="color: #098658;">0</span>], skip_special_tokens=<span style="color: #0000FF;">True</span>))
|
|
|
50 |
</code>
|
51 |
</pre>
|
52 |
|