feat: reduce load while model is first processing the data
Browse files- app.py +0 -4
- features/text_classifier/model_loader.py +6 -1
app.py
CHANGED
|
@@ -6,11 +6,7 @@ import nltk
|
|
| 6 |
|
| 7 |
@asynccontextmanager
|
| 8 |
async def lifespan(app: FastAPI):
|
| 9 |
-
# Ensure punkt is available
|
| 10 |
-
nltk.download("punkt")
|
| 11 |
|
| 12 |
-
nltk.download('punkt_tab')
|
| 13 |
-
|
| 14 |
# Your model warmup
|
| 15 |
warmup()
|
| 16 |
yield
|
|
|
|
| 6 |
|
| 7 |
@asynccontextmanager
|
| 8 |
async def lifespan(app: FastAPI):
|
|
|
|
|
|
|
| 9 |
|
|
|
|
|
|
|
| 10 |
# Your model warmup
|
| 11 |
warmup()
|
| 12 |
yield
|
features/text_classifier/model_loader.py
CHANGED
|
@@ -5,7 +5,7 @@ from transformers import GPT2LMHeadModel, GPT2TokenizerFast, GPT2Config
|
|
| 5 |
from huggingface_hub import snapshot_download
|
| 6 |
import torch
|
| 7 |
from dotenv import load_dotenv
|
| 8 |
-
|
| 9 |
load_dotenv()
|
| 10 |
REPO_ID = "Pujan-Dev/AI-Text-Detector"
|
| 11 |
MODEL_DIR = "./models"
|
|
@@ -18,6 +18,11 @@ _model, _tokenizer = None, None
|
|
| 18 |
|
| 19 |
def warmup():
|
| 20 |
global _model, _tokenizer
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
download_model_repo()
|
| 22 |
_model, _tokenizer = load_model()
|
| 23 |
logging.info("Its ready")
|
|
|
|
| 5 |
from huggingface_hub import snapshot_download
|
| 6 |
import torch
|
| 7 |
from dotenv import load_dotenv
|
| 8 |
+
import nltk
|
| 9 |
load_dotenv()
|
| 10 |
REPO_ID = "Pujan-Dev/AI-Text-Detector"
|
| 11 |
MODEL_DIR = "./models"
|
|
|
|
| 18 |
|
| 19 |
def warmup():
|
| 20 |
global _model, _tokenizer
|
| 21 |
+
# Ensure punkt is available
|
| 22 |
+
nltk.download("punkt")
|
| 23 |
+
|
| 24 |
+
nltk.download('punkt_tab')
|
| 25 |
+
|
| 26 |
download_model_repo()
|
| 27 |
_model, _tokenizer = load_model()
|
| 28 |
logging.info("Its ready")
|