from transformers import AutoModelForSeq2SeqLM, T5Tokenizer import torch MODEL_NAME = "PhunvVi/C3QG" # Hugging Face repo path def load_tokenizer(): return T5Tokenizer.from_pretrained(MODEL_NAME) def load_model(): model = AutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME) model.eval() if torch.cuda.is_available(): model.to("cuda") return model