Pujan-Dev commited on
Commit
942cb54
·
1 Parent(s): 6bc5c14

feat: reduce load while model is first processing the data

Browse files
app.py CHANGED
@@ -6,11 +6,7 @@ import nltk
6
 
7
  @asynccontextmanager
8
  async def lifespan(app: FastAPI):
9
- # Ensure punkt is available
10
- nltk.download("punkt")
11
 
12
- nltk.download('punkt_tab')
13
-
14
  # Your model warmup
15
  warmup()
16
  yield
 
6
 
7
  @asynccontextmanager
8
  async def lifespan(app: FastAPI):
 
 
9
 
 
 
10
  # Your model warmup
11
  warmup()
12
  yield
features/text_classifier/model_loader.py CHANGED
@@ -5,7 +5,7 @@ from transformers import GPT2LMHeadModel, GPT2TokenizerFast, GPT2Config
5
  from huggingface_hub import snapshot_download
6
  import torch
7
  from dotenv import load_dotenv
8
-
9
  load_dotenv()
10
  REPO_ID = "Pujan-Dev/AI-Text-Detector"
11
  MODEL_DIR = "./models"
@@ -18,6 +18,11 @@ _model, _tokenizer = None, None
18
 
19
  def warmup():
20
  global _model, _tokenizer
 
 
 
 
 
21
  download_model_repo()
22
  _model, _tokenizer = load_model()
23
  logging.info("Its ready")
 
5
  from huggingface_hub import snapshot_download
6
  import torch
7
  from dotenv import load_dotenv
8
+ import nltk
9
  load_dotenv()
10
  REPO_ID = "Pujan-Dev/AI-Text-Detector"
11
  MODEL_DIR = "./models"
 
18
 
19
  def warmup():
20
  global _model, _tokenizer
21
+ # Ensure punkt is available
22
+ nltk.download("punkt")
23
+
24
+ nltk.download('punkt_tab')
25
+
26
  download_model_repo()
27
  _model, _tokenizer = load_model()
28
  logging.info("Its ready")