import gradio as gr from transformers import pipeline from tensorflow.keras.preprocessing.sequence import pad_sequences from tensorflow.keras.preprocessing.text import text_to_word_sequence import pickle import re from tensorflow.keras.models import load_model # Load long model # with open('lstm-qa-long-answers-model/tokenizer.pickle', 'rb') as handle: # tokenizer = pickle.load(handle) # long_answer_model = load_model('lstm-qa-long-answers-model/model.h5') def clean_text(text): text = re.sub(r'<.*?>', '', text) text = re.sub(r'\[\d+\]', '', text) text = re.sub(r'[^a-zA-Z0-9\s().,]', '', text) return text def remove_parentheses(text): pattern = r'\([^)]*\)' return re.sub(pattern, '', text) # def split_into_sentences(text): # sentences = re.split(r'\.\s*', text) # return [s.strip() for s in sentences if s] # def predict_answer(context, question): # sentences = split_into_sentences(context) # best_sentence = None # best_score = 0 # for sentence in sentences: # clean_sentence = clean_text(sentence) # question_seq = tokenizer.texts_to_sequences([question]) # sentence_seq = tokenizer.texts_to_sequences([clean_sentence]) # max_sentence_length = 300 # padded_question = pad_sequences(question_seq, padding='post') # padded_sentence = pad_sequences(sentence_seq, maxlen=max_sentence_length, padding='post', truncating='post') # score = long_answer_model.predict([padded_sentence, padded_question])[0] # if score > best_score: # best_score = score # best_sentence = clean_sentence # return best_score, best_sentence # Load short model distilbert_base_uncased = pipeline(model="Nighter/QA_wiki_data_short_answer", from_tf=True) bert_base_uncased = pipeline(model="Nighter/QA_bert_base_uncased_wiki_data_short_answer", from_tf=True) roberta_base = pipeline(model="Nighter/QA_wiki_data_roberta_base_short_answer", from_tf=True) longformer_base = pipeline(model="Nighter/Fine-tuned-longformer-base") # Function to answer on all models def answer_questions(context, question): # long_score, long_answer = predict_answer(context, question) distilbert_base_uncased_result = distilbert_base_uncased(question=question, context=remove_parentheses(context)) bert_base_uncased_result = bert_base_uncased(question=question, context=remove_parentheses(context)) roberta_base_result = roberta_base(question=question, context=remove_parentheses(context)) longformer_base_result = longformer_base(question=question, context=remove_parentheses(context)) return distilbert_base_uncased_result['answer'], distilbert_base_uncased_result['score'], bert_base_uncased_result['answer'], bert_base_uncased_result['score'], roberta_base_result['answer'], longformer_base_result['score'], longformer_base_result['answer'], roberta_base_result['score'] #, long_answer, long_score # App Interface with gr.Blocks() as app: gr.Markdown("