# import all packages
import requests
import streamlit as st
from sklearn.model_selection import StratifiedKFold
from sklearn.model_selection import train_test_split
from sklearn.model_selection import KFold
# tokenizer
from transformers import AutoTokenizer, DistilBertTokenizerFast
# sequence tagging model + training-related
from transformers import DistilBertForTokenClassification, Trainer, TrainingArguments
import numpy as np
import pandas as pd
import torch
import json
import sys
import os
#from datasets import load_metric
from sklearn.metrics import classification_report
from pandas import read_csv
from sklearn.linear_model import LogisticRegression
import sklearn.model_selection
from sklearn.feature_extraction.text import TfidfTransformer
from sklearn.feature_extraction.text import CountVectorizer
#from sklearn.naive_bayes import MultinomialNB
#from sklearn.model_selection import GridSearchCV
from sklearn.pipeline import Pipeline, FeatureUnion
import math
from sklearn.metrics import accuracy_score
from sklearn.metrics import precision_recall_fscore_support
from sklearn.model_selection import train_test_split
#from sklearn.metrics import Scorer
#from sklearn.metrics import SCORERS
import json
import re
import numpy as np
import pandas as pd
import re
import nltk
nltk.download("punkt")
#stemmer = nltk.SnowballStemmer("english")
#from nltk.corpus import stopwords
import string
from sklearn.model_selection import train_test_split
# import seaborn as sns
# from sklearn.metrics import confusion_matrix
# from sklearn.metrics import classification_report, ConfusionMatrixDisplay
from transformers import AutoTokenizer, Trainer, TrainingArguments, AutoModelForSequenceClassification, AutoConfig
import torch
from torch.utils.data import TensorDataset, DataLoader, RandomSampler, SequentialSampler
import itertools
import json
import glob
from transformers import TextClassificationPipeline, TFAutoModelForSequenceClassification, AutoTokenizer
from transformers import pipeline
import pickle
import urllib.request
# from sklearn.feature_extraction.text import TfidfTransformer
# from sklearn.feature_extraction.text import CountVectorizer
#from PyPDF2 import PdfReader
#from urllib.request import urlopen
#from tabulate import tabulate
import csv
#import gdown
import pdfplumber
import pathlib
import shutil
import webbrowser
from streamlit.components.v1 import html
import streamlit.components.v1 as components
from PyPDF2 import PdfReader
from huggingface_hub import HfApi
import io
from datasets import load_dataset
import dill
# from huggingface_hub import Repository
# from git import repo
# api = HfApi()
#device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
# from git import Repo
# Repo.clone_from('https://github.com/gseetha04/IMA-weights.git', branch='master')
def main():
st.title("Text to Causal Knowledge Graph")
st.sidebar.title("Please upload your text documents in one file here:")
k=2
seed = 1
k1= 5
text_list = []
causal_sents = []
uploaded_file = None
try:
uploaded_file = st.sidebar.file_uploader("Choose a file", type = "pdf")
except:
uploaded_file = PdfReader('sample_anno.pdf')
st.error("Please upload your own PDF to be analyzed")
# try:
# #if uploaded_file is not None:
# uploaded_file = st.sidebar.file_uploader("Choose a file", type = "pdf")
# #st.stop()
# except:
# #uploaded_file = PdfReader('sample_anno.pdf')
# st.error("Please upload your own PDF to be analyzed")
#except:
#if uploaded_file
#st.write("Upload a pdf file...")
#st.stop()
if uploaded_file is not None:
reader = PdfReader(uploaded_file)
for page in reader.pages:
text = page.extract_text()
text_list.append(text)
else:
st.error("Please upload your own PDF to be analyzed")
st.stop()
text_list_final = [x.replace('\n', '') for x in text_list]
text_list_final = re.sub('"', '', str(text_list_final))
sentences = nltk.sent_tokenize(text_list_final)
result =[]
for i in sentences:
result1 = i.lower()
result2 = re.sub(r'[^\w\s]','',result1)
result.append(result2)
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased") #bert-base-uncased
model_path = "checkpoint-2850"
model = AutoModelForSequenceClassification.from_pretrained(model_path,id2label={0:'non-causal',1:'causal'})
pipe1 = pipeline("text-classification", model=model,tokenizer=tokenizer)
for sent in result:
pred = pipe1(sent)
for lab in pred:
if lab['label'] == 'causal': #causal
causal_sents.append(sent)
model_name = "distilbert-base-cased"
tokenizer = DistilBertTokenizerFast.from_pretrained(model_name)
model_path1 = "DistilBertforTokenclassification"
model = DistilBertForTokenClassification.from_pretrained(model_path1) #len(unique_tags),, num_labels= 7, , id2label={0:'CT',1:'E',2:'C',3:'O'}
pipe = pipeline('ner', model=model, tokenizer=tokenizer,aggregation_strategy='simple') #grouped_entities=True
sentence_pred = []
class_list = []
entity_list = []
for k in causal_sents:
pred= pipe(k)
#st.write(pred)
for i in pred:
sentence_pred.append(k)
class_list.append(i['word'])
entity_list.append(i['entity_group'])
filename = 'Checkpoint-classification.sav'
count_vect = CountVectorizer(ngram_range=[1,3])
tfidf_transformer=TfidfTransformer()
#loaded_model = pickle.load(open(filename, 'rb'))
loaded_model = pickle.load(open(filename, 'rb'))
#loaded_vectorizer = dill.load(open('vectorizefile_classification.pickle', 'rb'))
#loaded_vectorizer = pickle.load(open('vectorizefile_classification.pickle', 'rb'))
from sklearn.pipeline import Pipeline
pipeline1 = Pipeline([('count_vect',count_vect),('tfidf_transformer',tfidf_transformer)])
pipeline_test_output = pipeline1.fit_transform(class_list)
#pipeline_test_output = loaded_vectorizer.transform(class_list)
predicted = loaded_model.predict(pipeline_test_output)
pred1 = predicted
level0 = []
count =0
for i in predicted:
if i == 3:
level0.append('Non-Performance')
count +=1
else:
level0.append('Performance')
count +=1
list_pred = {0: 'Customers',1:'Employees',2:'Investors',3:'Non-performance',4:'Society',5:'Unclassified'}
pred_val = [list_pred[i] for i in pred1]
#print('count',count)
for ind,(sent,preds) in enumerate(zip(class_list,pred_val)):
if 'customers' in sent or 'client' in sent or 'consumer' in sent or 'user' in sent:
pred_val[ind] = 'Customers'
elif 'investor' in sent or 'finance' in sent or 'shareholder' in sent or 'stockholder' in sent or 'owners' in sent:
pred_val[ind] = 'Investors'
elif 'employee' in sent or 'worker' in sent or 'staff' in sent:
pred_val[ind] = 'Employees'
elif 'society' in sent or 'societal' in sent or 'social responsib*' in sent or 'social performance' in sent or 'community' in sent:
pred_val[ind] = 'Society'
sent_id, unique = pd.factorize(sentence_pred)
final_list = pd.DataFrame(
{'Id': sent_id,
'Fullsentence': sentence_pred,
'Component': class_list,
'causeOrEffect': entity_list,
'Labellevel1': level0,
'Labellevel2': pred_val
})
s = final_list['Component'].shift(-1)
m = s.str.startswith('##', na=False)
final_list.loc[m, 'Component'] += (' ' + s[m])
final_list1 = final_list[~final_list['Component'].astype(str).str.startswith('##')]
li = []
uni = final_list1['Id'].unique()
for i in uni:
df_new = final_list1[final_list1['Id'] == i]
uni1 = df_new['Id'].unique()
if 'E' not in df_new.values:
li.append(uni1)
out = np.concatenate(li).ravel()
li_pan = pd.DataFrame(out,columns=['Id'])
df3 = pd.merge(final_list1, li_pan[['Id']], on='Id', how='left', indicator=True) \
.query("_merge == 'left_only'") \
.drop("_merge",axis=1)
#df = df3.groupby(['Id','Fullsentence','causeOrEffect', 'Labellevel1', 'Labellevel2'])['Component'].apply(', '.join).reset_index()
#st.write(df)
#df = df3
df3["causeOrEffect"].replace({"C": "cause", "E": "effect"}, inplace=True)
df_final = df3[df3['causeOrEffect'] != 'CT']
df3['New string'] = df_final['Component'].replace(r'[##]+', ' ', regex=True)
df_final = df_final.drop("Component",axis=1)
df_final.insert(2, "Component", df3['New string'], True)
df_final1 = df_final[df_final['Component'].str.split().str.len().gt(1)]
#st.write(df_final[df_final['Component'].str.len() != 1])
#df_final1.to_csv('predictions.csv')
# buffer = io.BytesIO()
# with pd.ExcelWriter(buffer, engine="xlsxwriter") as writer:
# df_final.to_excel(writer, sheet_name="Sheet1", index=False)
# writer.close()
count_NP_NP = 0
count_NP_investor = 0
count_NP_customer = 0
count_NP_employees = 0
count_NP_society = 0
count_inv_np = 0
count_inv_investor = 0
count_inv_customer = 0
count_inv_employee = 0
count_inv_society = 0
count_cus_np = 0
count_cus_investor = 0
count_cus_customer = 0
count_cus_employee = 0
count_cus_society = 0
count_emp_np = 0
count_emp_investor = 0
count_emp_customer = 0
count_emp_employee = 0
count_emp_society = 0
count_soc_np = 0
count_soc_investor = 0
count_soc_customer = 0
count_soc_employee = 0
count_soc_society = 0
for i in range(0,df_final['Id'].max()):
j = df_final.loc[df_final['Id'] == i]
cause_tab = j.loc[j['causeOrEffect'] == 'cause']
effect_tab = j.loc[j['causeOrEffect'] == 'effect']
cause_coun_NP = (cause_tab.Labellevel2 == 'Non-performance').sum()
effect_coun_NP = (effect_tab.Labellevel2 == 'Non-performance').sum()
if (cause_coun_NP > 0) and (effect_coun_NP > 0):
count_NP = cause_coun_NP if cause_coun_NP >= effect_coun_NP else effect_coun_NP
else:
count_NP = 0
effect_NP_inv = (effect_tab.Labellevel2 == 'Investors').sum()
if (cause_coun_NP > 0) and (effect_NP_inv > 0):
count_NP_inv = cause_coun_NP if cause_coun_NP >= effect_NP_inv else effect_NP_inv
else:
count_NP_inv = 0
effect_NP_cus = (effect_tab.Labellevel2 == 'Customers').sum()
if (cause_coun_NP > 0) and (effect_NP_cus > 0):
count_NP_cus = cause_coun_NP if cause_coun_NP >= effect_NP_cus else effect_NP_cus
else:
count_NP_cus = 0
effect_NP_emp = (effect_tab.Labellevel2 == 'Employees').sum()
if (cause_coun_NP > 0) and (effect_NP_emp > 0):
count_NP_emp = cause_coun_NP if cause_coun_NP >= effect_NP_emp else effect_NP_emp
else:
count_NP_emp = 0
effect_NP_soc = (effect_tab.Labellevel2 == 'Society').sum()
if (cause_coun_NP > 0) and (effect_NP_soc > 0):
count_NP_soc = cause_coun_NP if cause_coun_NP >= effect_NP_soc else effect_NP_soc
else:
count_NP_soc = 0
cause_coun_inv = (cause_tab.Labellevel2 == 'Investors').sum()
effect_coun_inv = (effect_tab.Labellevel2 == 'Non-performance').sum()
if (cause_coun_inv > 0) and (effect_coun_inv > 0):
count_NP_inv = cause_coun_inv if cause_coun_inv >= effect_coun_inv else effect_coun_inv
else:
count_NP_inv = 0
effect_inv_inv = (effect_tab.Labellevel2 == 'Investors').sum()
if (cause_coun_inv > 0) and (effect_inv_inv > 0):
count_inv_inv = cause_coun_inv if cause_coun_inv >= effect_inv_inv else effect_inv_inv
else:
count_inv_inv = 0
effect_inv_cus = (effect_tab.Labellevel2 == 'Customers').sum()
if (cause_coun_inv > 0) and (effect_inv_cus > 0):
count_inv_cus = cause_coun_inv if cause_coun_inv >= effect_inv_cus else effect_inv_cus
else:
count_inv_cus = 0
effect_inv_emp = (effect_tab.Labellevel2 == 'Employees').sum()
if (cause_coun_inv > 0) and (effect_inv_emp > 0):
count_inv_emp = cause_coun_inv if cause_coun_inv >= effect_inv_emp else effect_inv_emp
else:
count_inv_emp = 0
effect_inv_soc = (effect_tab.Labellevel2 == 'Society').sum()
if (cause_coun_inv > 0) and (effect_inv_soc > 0):
count_inv_soc = cause_coun_inv if cause_coun_inv >= effect_inv_soc else effect_inv_soc
else:
count_inv_soc = 0
cause_coun_cus = (cause_tab.Labellevel2 == 'Customers').sum()
effect_coun_cus = (effect_tab.Labellevel2 == 'Non-performance').sum()
if (cause_coun_cus > 0) and (effect_coun_cus > 0):
count_NP_cus = cause_coun_cus if cause_coun_cus >= effect_coun_cus else effect_coun_cus
else:
count_NP_cus = 0
effect_cus_inv = (effect_tab.Labellevel2 == 'Investors').sum()
if (cause_coun_cus > 0) and (effect_cus_inv > 0):
count_cus_inv = cause_coun_cus if cause_coun_cus >= effect_cus_inv else effect_cus_inv
else:
count_cus_inv = 0
effect_cus_cus = (effect_tab.Labellevel2 == 'Customers').sum()
if (cause_coun_cus > 0) and (effect_cus_cus > 0):
count_cus_cus = cause_coun_cus if cause_coun_cus >= effect_cus_cus else effect_cus_cus
else:
count_cus_cus = 0
effect_cus_emp = (effect_tab.Labellevel2 == 'Employees').sum()
if (cause_coun_cus > 0) and (effect_cus_emp > 0):
count_cus_emp = cause_coun_cus if cause_coun_cus >= effect_cus_emp else effect_cus_emp
else:
count_cus_emp = 0
effect_cus_soc = (effect_tab.Labellevel2 == 'Society').sum()
if (cause_coun_cus > 0) and (effect_cus_soc > 0):
count_cus_soc = cause_coun_cus if cause_coun_cus >= effect_cus_soc else effect_cus_soc
else:
count_cus_soc = 0
cause_coun_emp = (cause_tab.Labellevel2 == 'Employees').sum()
effect_coun_emp = (effect_tab.Labellevel2 == 'Non-performance').sum()
if (cause_coun_emp > 0) and (effect_coun_emp > 0):
count_NP_emp = cause_coun_emp if cause_coun_emp >= effect_coun_emp else effect_coun_emp
else:
count_NP_emp = 0
effect_emp_inv = (effect_tab.Labellevel2 == 'Investors').sum()
if (cause_coun_emp > 0) and (effect_emp_inv > 0):
count_emp_inv = cause_coun_emp if cause_coun_emp >= effect_emp_inv else effect_emp_inv
else:
count_emp_inv = 0
effect_emp_cus = (effect_tab.Labellevel2 == 'Customers').sum()
if (cause_coun_emp > 0) and (effect_emp_cus > 0):
count_emp_cus = cause_coun_emp if cause_coun_emp >= effect_emp_cus else effect_emp_cus
else:
count_emp_cus = 0
effect_emp_emp = (effect_tab.Labellevel2 == 'Employees').sum()
if (cause_coun_emp > 0) and (effect_emp_emp > 0):
count_emp_emp = cause_coun_emp if cause_coun_emp >= effect_emp_emp else effect_emp_emp
else:
count_emp_emp = 0
effect_emp_soc = (effect_tab.Labellevel2 == 'Society').sum()
if (cause_coun_emp > 0) and (effect_emp_soc > 0):
count_emp_soc = cause_coun_emp if cause_coun_emp >= effect_emp_soc else effect_emp_soc
else:
count_emp_soc = 0
cause_coun_soc = (cause_tab.Labellevel2 == 'Society').sum()
effect_coun_soc = (effect_tab.Labellevel2 == 'Non-performance').sum()
if (cause_coun_soc > 0) and (effect_coun_soc > 0):
count_NP_soc = cause_coun_soc if cause_coun_soc >= effect_coun_soc else effect_coun_soc
else:
count_NP_soc = 0
effect_soc_inv = (effect_tab.Labellevel2 == 'Investors').sum()
if (cause_coun_soc > 0) and (effect_soc_inv > 0):
count_soc_inv = cause_coun_soc if cause_coun_soc >= effect_soc_inv else effect_soc_inv
else:
count_soc_inv = 0
effect_soc_cus = (effect_tab.Labellevel2 == 'Customers').sum()
if (cause_coun_soc > 0) and (effect_soc_cus > 0):
count_soc_cus = cause_coun_soc if cause_coun_soc >= effect_soc_cus else effect_soc_cus
else:
count_soc_cus = 0
effect_soc_emp = (effect_tab.Labellevel2 == 'Employees').sum()
if (cause_coun_soc > 0) and (effect_soc_emp > 0):
count_soc_emp = cause_coun_soc if cause_coun_soc >= effect_soc_emp else effect_soc_emp
else:
count_soc_emp = 0
effect_soc_soc = (effect_tab.Labellevel2 == 'Society').sum()
if (cause_coun_soc > 0) and (effect_soc_soc > 0):
count_soc_soc = cause_coun_soc if cause_coun_soc >= effect_soc_soc else effect_soc_soc
else:
count_soc_soc = 0
count_NP_NP = count_NP_NP + count_NP
count_NP_investor = count_NP_investor + count_NP_inv
count_NP_customer = count_NP_customer + count_NP_cus
count_NP_employees = count_NP_employees + count_NP_emp
count_NP_society = count_NP_society + count_NP_soc
count_inv_np = count_inv_np + count_NP_inv
count_inv_investor = count_inv_investor + count_inv_inv
count_inv_customer = count_inv_customer + count_inv_cus
count_inv_employee = count_inv_employee + count_inv_emp
count_inv_society = count_inv_society + count_inv_soc
count_cus_np = count_cus_np + count_NP_cus
count_cus_investor = count_cus_investor + count_cus_inv
count_cus_customer = count_cus_customer + count_cus_cus
count_cus_employee = count_cus_employee + count_cus_emp
count_cus_society = count_cus_society + count_cus_soc
count_emp_np = count_emp_np + count_NP_emp
count_emp_investor = count_emp_investor + count_emp_inv
count_emp_customer = count_emp_customer + count_emp_cus
count_emp_employee = count_emp_employee + count_emp_emp
count_emp_society = count_emp_society + count_emp_soc
count_soc_np = count_soc_np + count_NP_soc
count_soc_investor = count_soc_investor + count_soc_inv
count_soc_customer = count_soc_customer + count_soc_cus
count_soc_employee = count_soc_employee + count_soc_emp
count_soc_society = count_soc_society + count_soc_soc
df_tab = pd.DataFrame(columns = ['Non-performance', 'Investors', 'Customers', 'Employees', 'Society'],index=['Non-performance', 'Investors', 'Customers', 'Employees', 'Society'], dtype=object)
df_tab.loc['Non-performance'] = [count_NP_NP, count_NP_investor, count_NP_customer, count_NP_employees, count_NP_society]
df_tab.loc['Investors'] = [count_inv_np, count_inv_investor, count_inv_customer, count_inv_employee, count_inv_society]
df_tab.loc['Customers'] = [count_cus_np, count_cus_investor, count_cus_customer, count_cus_employee, count_cus_society]
df_tab.loc['Employees'] = [count_emp_np, count_emp_investor, count_emp_customer, count_emp_employee, count_emp_society]
df_tab.loc['Society'] = [count_soc_np, count_soc_investor, count_soc_customer, count_soc_employee, count_soc_society]
# df_tab = pd.DataFrame({
# 'Non-performance': [count_NP_NP, count_NP_investor, count_NP_customer, count_NP_employees, count_NP_society],
# 'Investors': [count_inv_np, count_inv_investor, count_inv_customer, count_inv_employee, count_inv_society],
# 'Customers': [count_cus_np, count_cus_investor, count_cus_customer, count_cus_employee, count_cus_society],
# 'Employees': [count_emp_np, count_emp_investor, count_emp_customer, count_emp_employee, count_emp_society],
# 'Society': [count_soc_np, count_soc_investor, count_soc_customer, count_soc_employee, count_soc_society]},
# index=['Non-performance', 'Investors', 'Customers', 'Employees', 'Society'])
#df_tab.to_csv('final_data.csv')
buffer = io.BytesIO()
with pd.ExcelWriter(buffer, engine="xlsxwriter") as writer:
df_tab.to_excel(writer,sheet_name="count_result",index=False)
df_final1.to_excel(writer,sheet_name="Detailed_results",index=False)
writer.close()
df = pd.read_csv('final_data.csv', index_col=0)
#474-515
# # Convert to JSON format
json_data = []
for row in df_tab.index:
for col in df_tab.columns:
json_data.append({
'source': row,
'target': col,
'value': int(df.loc[row, col])
})
# Write JSON to file
with open('ch.json', 'w+') as f:
json.dump(json_data, f)
# # repo.git_pull()
# # repo.git_add("ch.json")
# # repo.git_commit(commit_message="add ch.json :)")
# # repo.push()
# csv_file = "predictions.csv"
# #json_file = "smalljson.json"
# # Open the CSV file and read the data
# with open(csv_file, "r") as f:
# csv_data = csv.DictReader(f)
# # Convert the CSV data to a list of dictionaries
# data_list = []
# for row in csv_data:
# data_list.append(dict(row))
# # Convert the list of dictionaries to JSON
# json_data = json.dumps(data_list)
# # Write the JSON data to a file
# #with open("smalljson.json", "r+") as fi:
# #data = fi.read()
# #fi.seek(0)
# with open('smalljson.json','w+') as fi:
# # data = json.load(fi)
# #st.write(data)
# # fi.seek(0)
# fi.write(json_data)
# fi.truncate()
# file_jso.truncate(0)
# file_jso.write(json_data)
# st.write('after',file_jso.read())
# file_jso.close()
# api.upload_file(
# path_or_fileobj=json_data,
# path_in_repo="ch.json",
# repo_id="Seetha/IMA-pipeline-streamlit",
# repo_type="dataset",
# token="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",
# )
#dataset = load_dataset('json', data_files='smalljson.json')
def convert_df(df):
#IMPORTANT: Cache the conversion to prevent computation on every rerun
return df.to_csv().encode('utf-8')
csv1 = convert_df(df_final1.astype(str))
csv2 = convert_df(df_tab.astype(str))
with st.container():
#st.download_button(label="Download the detailed result table_csv",data=csv1,file_name='results.csv',mime='text/csv')
# st.download_button(label="Download the result table_csv",data=csv2,file_name='final_data.csv',mime='text/csv')
#st.download_button(label="Download the detailed result table",data=buffer,file_name="df_final.xlsx",mime="application/vnd.ms-excel")
st.download_button(label="Download the result table",data=buffer,file_name="t2cg_outputs.xlsx",mime="application/vnd.ms-excel")
# repo_dir = 'IMA-pipeline-streamlit'
# repo = Repo(repo_dir)
# file_list = [
# '/app/ima-pipeline-streamlit/results.csv',
# '/app/ima-pipeline-streamlit/final_data.csv'
# ]
# commit_message = 'Add the generated files to Github'
# repo.index.add(file_list)
# repo.index.commit(commit_message)
# origin = repo.remote('origin')
# origin.push()
# # LINK TO THE CSS FILE
# def tree_css(file_name):
# with open('tree.css')as f:
# st.markdown(f"", unsafe_allow_html = True)
# def div_css(file_name):
# with open('div.css')as f:
# st.markdown(f"", unsafe_allow_html = True)
# def side_css(file_name):
# with open('side.css')as f:
# st.markdown(f"", unsafe_allow_html = True)
# tree_css('tree.css')
# div_css('div.css')
# side_css('side.css')
# STREAMLIT_STATIC_PATH = pathlib.Path(st.__path__[0]) / 'static'
# CSS_PATH = (STREAMLIT_STATIC_PATH / "css")
# if not CSS_PATH.is_dir():
# CSS_PATH.mkdir()
# # css_file = CSS_PATH / "tree.css"
# # css_file1 = CSS_PATH / "div.css"
# # css_file2 = CSS_PATH / "side.css"
# css_file2 = CSS_PATH / "ch.json"
# if not css_file2.exists():
# shutil.copy("css/ch.json", css_file2)
# shutil.copy("assets/css/tree.css", css_file)
# shutil.copy("assets/css/div.css", css_file1)
# shutil.copy("assets/css/side.css", css_file2)
# HtmlFile = open("index.html", 'r', encoding='utf-8')
# source_code = HtmlFile.read()
# #print(source_code)
# components.html(source_code)
# # # Define your javascript
# my_js = """
# alert("Hello World");
# """
# Wrapt the javascript as html code
#my_html = f""
# with st.container():
# # Execute your app
# st.title("Visualization example")
# # components.html(source_code)
# #html(my_html)
# #webbrowser.open('https://webpages.charlotte.edu/ltotapal/')
# # embed streamlit docs in a streamlit app
# #components.iframe("https://webpages.charlotte.edu/ltotapal/")
# st.markdown('Text to Knowledge graph link', unsafe_allow_html=True)
if __name__ == '__main__':
main()