Lisibonny commited on
Commit
bc75b5c
1 Parent(s): 64dce50

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +30 -15
app.py CHANGED
@@ -136,28 +136,43 @@ def main():
136
 
137
  text=remove_html_markup(df_answer.loc[i, "resumen"])
138
  text=remove_URL(text)
 
139
  inputs = tokenizer(query, text[0:512], return_tensors='tf')
140
- outputs = qa_model(input_ids=inputs['input_ids'], attention_mask=inputs['attention_mask'])
141
- answer_start_index = int(tf.math.argmax(outputs.start_logits, axis=-1)[0])
142
- answer_end_index = int(tf.math.argmax(outputs.end_logits, axis=-1)[0])
 
 
 
 
 
 
 
 
 
 
 
 
 
 
143
  #answer_start_scores = tf.nn.softmax(outputs.start_logits)
144
  #answer_end_scores = tf.nn.softmax(outputs.end_logits)
145
  #######################
146
- start_probabilities = tf.nn.softmax(outputs.start_logits, axis=-1)[0]
147
- end_probabilities = tf.nn.softmax(outputs.end_logits, axis=-1)[0]
148
- scores = start_probabilities[:, None] * end_probabilities[None, :]
149
- scores = tf.linalg.band_part(scores, 0, -1)
150
- scores = tf.reshape(scores, [-1])
151
- st.write(scores)
152
- max_index = np.argmax(scores)
153
- st.write(max_index)
154
- start_index = max_index // scores.shape[1]
155
- end_index = max_index % scores.shape[1]
156
  #st.write(start_index)
157
  #st.write(scores[start_index:end_index])
158
  #######################
159
- predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1]
160
- answer=tokenizer.decode(predict_answer_tokens)
161
 
162
 
163
  if (len(answer)>0):
 
136
 
137
  text=remove_html_markup(df_answer.loc[i, "resumen"])
138
  text=remove_URL(text)
139
+
140
  inputs = tokenizer(query, text[0:512], return_tensors='tf')
141
+ input_ids = inputs["input_ids"].numpy()[0]
142
+
143
+ text_tokens = tokenizer.convert_ids_to_tokens(input_ids)
144
+ answer_start_scores, answer_end_scores = qa_model(inputs)
145
+
146
+ answer_start = tf.argmax(answer_start_scores, axis=1).numpy()[0]
147
+ answer_end = (tf.argmax(answer_end_scores, axis=1) + 1).numpy()[0]
148
+
149
+ answer = tokenizer.convert_tokens_to_string(tokenizer.convert_ids_to_tokens(input_ids[answer_start:answer_end]))
150
+
151
+ df_answer.loc[x] = answer, max(answer_start_scores.numpy()[0]), 0, 0
152
+ st.write(df_answer.sort_values(by=['score']).tail(10))
153
+
154
+ #outputs = qa_model(input_ids=inputs['input_ids'], attention_mask=inputs['attention_mask'])
155
+
156
+ #answer_start_index = int(tf.math.argmax(outputs.start_logits, axis=-1)[0])
157
+ #answer_end_index = int(tf.math.argmax(outputs.end_logits, axis=-1)[0])
158
  #answer_start_scores = tf.nn.softmax(outputs.start_logits)
159
  #answer_end_scores = tf.nn.softmax(outputs.end_logits)
160
  #######################
161
+ #start_probabilities = tf.nn.softmax(outputs.start_logits, axis=-1)[0]
162
+ #end_probabilities = tf.nn.softmax(outputs.end_logits, axis=-1)[0]
163
+ #scores = start_probabilities[:, None] * end_probabilities[None, :]
164
+ #scores = tf.linalg.band_part(scores, 0, -1)
165
+ #scores = tf.reshape(scores, [-1])
166
+ #st.write(scores)
167
+ #max_index = np.argmax(scores)
168
+ #st.write(max_index)
169
+ #start_index = max_index // scores.shape[1]
170
+ #end_index = max_index % scores.shape[1]
171
  #st.write(start_index)
172
  #st.write(scores[start_index:end_index])
173
  #######################
174
+ #predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1]
175
+ #answer=tokenizer.decode(predict_answer_tokens)
176
 
177
 
178
  if (len(answer)>0):