Upload app.py
Browse files
app.py
CHANGED
@@ -10,14 +10,9 @@ import tensorflow as tf
|
|
10 |
from nltk.corpus import stopwords
|
11 |
from nltk.tokenize import word_tokenize
|
12 |
from nltk.stem import WordNetLemmatizer
|
13 |
-
from gensim.models import Word2Vec
|
14 |
nltk.download('punkt')
|
15 |
nltk.download('stopwords')
|
16 |
|
17 |
-
# # Load tokenizer
|
18 |
-
# with open("tokenizer.pkl", "rb") as tokenizer_file:
|
19 |
-
# tokenizer = pickle.load(tokenizer_file)
|
20 |
-
# Define the model
|
21 |
model_path= 'model'
|
22 |
|
23 |
# Load model
|
@@ -114,11 +109,8 @@ def run():
|
|
114 |
if submitted:
|
115 |
df_inf = {'preprocessing_review': text}
|
116 |
df_inf = pd.DataFrame([df_inf])
|
117 |
-
# Preprocess the text (apply the same preprocessing steps as used during training)
|
118 |
df_inf['preprocessing_review'] = df_inf['preprocessing_review'].apply(lambda x: review_preprocessing(x))
|
119 |
-
|
120 |
-
# df_inf = pad_sequences(df_inf, maxlen=700)
|
121 |
-
# Make the prediction using the loaded model
|
122 |
|
123 |
y_pred_inf = model.predict(df_inf['preprocessing_review'])
|
124 |
y_pred_inf = np.argmax(df_inf['preprocessing_review'], axis = -1)
|
|
|
10 |
from nltk.corpus import stopwords
|
11 |
from nltk.tokenize import word_tokenize
|
12 |
from nltk.stem import WordNetLemmatizer
|
|
|
13 |
nltk.download('punkt')
|
14 |
nltk.download('stopwords')
|
15 |
|
|
|
|
|
|
|
|
|
16 |
model_path= 'model'
|
17 |
|
18 |
# Load model
|
|
|
109 |
if submitted:
|
110 |
df_inf = {'preprocessing_review': text}
|
111 |
df_inf = pd.DataFrame([df_inf])
|
|
|
112 |
df_inf['preprocessing_review'] = df_inf['preprocessing_review'].apply(lambda x: review_preprocessing(x))
|
113 |
+
|
|
|
|
|
114 |
|
115 |
y_pred_inf = model.predict(df_inf['preprocessing_review'])
|
116 |
y_pred_inf = np.argmax(df_inf['preprocessing_review'], axis = -1)
|