qhduan commited on
Commit
0bee447
1 Parent(s): f37adf3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -0
README.md CHANGED
@@ -45,4 +45,79 @@ generated_ids = model.generate(
45
  ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False)
46
  print(ret)
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  ```
 
45
  ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False)
46
  print(ret)
47
 
48
+ ```
49
+
50
+ ## Prompt
51
+
52
+ You could give model a style or a specific language, for example:
53
+
54
+ ```python
55
+ inputs = tokenizer('''<|endoftext|>
56
+ function add(a, b){
57
+ return a + b;
58
+ }
59
+ // docstring
60
+ /**
61
+ * Calculate number add.
62
+ * @param a {number} the first number to add
63
+ * @param b {number} the second number to add
64
+ * @return the result of a + b
65
+ */
66
+ <|endoftext|>
67
+ function getDateAfterNDay(n){
68
+ return moment().add(n, 'day')
69
+ }
70
+ // docstring
71
+ /**''', return_tensors='pt')
72
+
73
+ doc_max_length = 128
74
+
75
+ generated_ids = model.generate(
76
+ **inputs,
77
+ max_length=inputs.input_ids.shape[1] + doc_max_length,
78
+ do_sample=False,
79
+ return_dict_in_generate=True,
80
+ num_return_sequences=1,
81
+ output_scores=True,
82
+ pad_token_id=50256,
83
+ eos_token_id=50256 # <|endoftext|>
84
+ )
85
+
86
+ ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False)
87
+ print(ret)
88
+
89
+ inputs = tokenizer('''<|endoftext|>
90
+ function add(a, b){
91
+ return a + b;
92
+ }
93
+ // docstring
94
+ /**
95
+ * 计算数字相加
96
+ * @param a {number} 第一个加数
97
+ * @param b {number} 第二个加数
98
+ * @return 返回 a + b 的结果
99
+ */
100
+ <|endoftext|>
101
+ function getDateAfterNDay(n){
102
+ return moment().add(n, 'day')
103
+ }
104
+ // docstring
105
+ /**''', return_tensors='pt')
106
+
107
+ doc_max_length = 128
108
+
109
+ generated_ids = model.generate(
110
+ **inputs,
111
+ max_length=inputs.input_ids.shape[1] + doc_max_length,
112
+ do_sample=False,
113
+ return_dict_in_generate=True,
114
+ num_return_sequences=1,
115
+ output_scores=True,
116
+ pad_token_id=50256,
117
+ eos_token_id=50256 # <|endoftext|>
118
+ )
119
+
120
+ ret = tokenizer.decode(generated_ids.sequences[0], skip_special_tokens=False)
121
+ print(ret)
122
+
123
  ```