Update app.py
Browse files
app.py
CHANGED
@@ -136,28 +136,43 @@ def main():
|
|
136 |
|
137 |
text=remove_html_markup(df_answer.loc[i, "resumen"])
|
138 |
text=remove_URL(text)
|
|
|
139 |
inputs = tokenizer(query, text[0:512], return_tensors='tf')
|
140 |
-
|
141 |
-
|
142 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
143 |
#answer_start_scores = tf.nn.softmax(outputs.start_logits)
|
144 |
#answer_end_scores = tf.nn.softmax(outputs.end_logits)
|
145 |
#######################
|
146 |
-
start_probabilities = tf.nn.softmax(outputs.start_logits, axis=-1)[0]
|
147 |
-
end_probabilities = tf.nn.softmax(outputs.end_logits, axis=-1)[0]
|
148 |
-
scores = start_probabilities[:, None] * end_probabilities[None, :]
|
149 |
-
scores = tf.linalg.band_part(scores, 0, -1)
|
150 |
-
scores = tf.reshape(scores, [-1])
|
151 |
-
st.write(scores)
|
152 |
-
max_index = np.argmax(scores)
|
153 |
-
st.write(max_index)
|
154 |
-
start_index = max_index // scores.shape[1]
|
155 |
-
end_index = max_index % scores.shape[1]
|
156 |
#st.write(start_index)
|
157 |
#st.write(scores[start_index:end_index])
|
158 |
#######################
|
159 |
-
predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1]
|
160 |
-
answer=tokenizer.decode(predict_answer_tokens)
|
161 |
|
162 |
|
163 |
if (len(answer)>0):
|
|
|
136 |
|
137 |
text=remove_html_markup(df_answer.loc[i, "resumen"])
|
138 |
text=remove_URL(text)
|
139 |
+
|
140 |
inputs = tokenizer(query, text[0:512], return_tensors='tf')
|
141 |
+
input_ids = inputs["input_ids"].numpy()[0]
|
142 |
+
|
143 |
+
text_tokens = tokenizer.convert_ids_to_tokens(input_ids)
|
144 |
+
answer_start_scores, answer_end_scores = qa_model(inputs)
|
145 |
+
|
146 |
+
answer_start = tf.argmax(answer_start_scores, axis=1).numpy()[0]
|
147 |
+
answer_end = (tf.argmax(answer_end_scores, axis=1) + 1).numpy()[0]
|
148 |
+
|
149 |
+
answer = tokenizer.convert_tokens_to_string(tokenizer.convert_ids_to_tokens(input_ids[answer_start:answer_end]))
|
150 |
+
|
151 |
+
df_answer.loc[x] = answer, max(answer_start_scores.numpy()[0]), 0, 0
|
152 |
+
st.write(df_answer.sort_values(by=['score']).tail(10))
|
153 |
+
|
154 |
+
#outputs = qa_model(input_ids=inputs['input_ids'], attention_mask=inputs['attention_mask'])
|
155 |
+
|
156 |
+
#answer_start_index = int(tf.math.argmax(outputs.start_logits, axis=-1)[0])
|
157 |
+
#answer_end_index = int(tf.math.argmax(outputs.end_logits, axis=-1)[0])
|
158 |
#answer_start_scores = tf.nn.softmax(outputs.start_logits)
|
159 |
#answer_end_scores = tf.nn.softmax(outputs.end_logits)
|
160 |
#######################
|
161 |
+
#start_probabilities = tf.nn.softmax(outputs.start_logits, axis=-1)[0]
|
162 |
+
#end_probabilities = tf.nn.softmax(outputs.end_logits, axis=-1)[0]
|
163 |
+
#scores = start_probabilities[:, None] * end_probabilities[None, :]
|
164 |
+
#scores = tf.linalg.band_part(scores, 0, -1)
|
165 |
+
#scores = tf.reshape(scores, [-1])
|
166 |
+
#st.write(scores)
|
167 |
+
#max_index = np.argmax(scores)
|
168 |
+
#st.write(max_index)
|
169 |
+
#start_index = max_index // scores.shape[1]
|
170 |
+
#end_index = max_index % scores.shape[1]
|
171 |
#st.write(start_index)
|
172 |
#st.write(scores[start_index:end_index])
|
173 |
#######################
|
174 |
+
#predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1]
|
175 |
+
#answer=tokenizer.decode(predict_answer_tokens)
|
176 |
|
177 |
|
178 |
if (len(answer)>0):
|