macedonizer commited on
Commit
6ceeafd
1 Parent(s): 2953b64

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -5
README.md CHANGED
@@ -30,10 +30,32 @@ from transformers import AutoTokenizer, AutoModelWithLMHead
30
 
31
  tokenizer = AutoTokenizer.from_pretrained('macedonizer/blaze-koneski') \\nmodel = AutoModelWithLMHead.from_pretrained('macedonizer/blaze-koneski')
32
 
33
- input_text = 'Моска '
34
-
35
- if len(input_text) == 0: \\n encoded_input = tokenizer(input_text, return_tensors="pt") \\n output = model.generate( \\n bos_token_id=random.randint(1, 50000), \\n do_sample=True, \\n top_k=50, \\n max_length=1024, \\n top_p=0.95, \\n num_return_sequences=1, \\n ) \\nelse: \\n encoded_input = tokenizer(input_text, return_tensors="pt") \\n output = model.generate( \\n **encoded_input, \\n bos_token_id=random.randint(1, 50000), \\n do_sample=True, \\n top_k=50, \\n max_length=1024, \\n top_p=0.95, \\n num_return_sequences=1, \\n )
36
-
37
- decoded_output = [] \\nfor sample in output: \\n decoded_output.append(tokenizer.decode(sample, skip_special_tokens=True))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
 
39
  print(decoded_output)
30
 
31
  tokenizer = AutoTokenizer.from_pretrained('macedonizer/blaze-koneski') \\nmodel = AutoModelWithLMHead.from_pretrained('macedonizer/blaze-koneski')
32
 
33
+ input_text = 'Москва '
34
+
35
+ if len(input_text) == 0: \
36
+ encoded_input = tokenizer(input_text, return_tensors="pt") \
37
+ output = model.generate( \
38
+ bos_token_id=random.randint(1, 50000), \
39
+ do_sample=True, \
40
+ top_k=50, \
41
+ max_length=1024, \
42
+ top_p=0.95, \
43
+ num_return_sequences=1, \
44
+ ) \
45
+ else: \
46
+ encoded_input = tokenizer(input_text, return_tensors="pt") \
47
+ output = model.generate( \
48
+ **encoded_input, \
49
+ bos_token_id=random.randint(1, 50000), \
50
+ do_sample=True, \
51
+ top_k=50, \
52
+ max_length=1024, \
53
+ top_p=0.95, \
54
+ num_return_sequences=1, \
55
+ )
56
+
57
+ decoded_output = [] \
58
+ for sample in output: \
59
+ decoded_output.append(tokenizer.decode(sample, skip_special_tokens=True))
60
 
61
  print(decoded_output)