Spaces:
Running
Running
File size: 1,303 Bytes
dc480fb 732d546 dc480fb 732d546 dc480fb 732d546 dc480fb 732d546 dc480fb 732d546 dc480fb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
import os
from dotenv import load_dotenv
from transformers import TFBertForSequenceClassification, BertTokenizerFast
import tensorflow as tf
# Load environment variables
load_dotenv()
def load_model(model_name):
try:
# Try loading the model as a TensorFlow model
model = TFBertForSequenceClassification.from_pretrained(model_name, use_auth_token=os.getenv('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'))
except OSError:
# If loading fails, assume it's a PyTorch model and use from_pt=True
model = TFBertForSequenceClassification.from_pretrained(model_name, use_auth_token=os.getenv('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'), from_pt=True)
return model
def load_tokenizer(model_name):
tokenizer = BertTokenizerFast.from_pretrained(model_name, use_auth_token=os.getenv('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'))
return tokenizer
def predict(text, model, tokenizer):
inputs = tokenizer(text, return_tensors="tf")
outputs = model(**inputs)
return outputs
def main():
model_name = os.getenv('Erfan11/Neuracraft')
model = load_model(model_name)
tokenizer = load_tokenizer(model_name)
# Example usage
text = "Sample input text"
result = predict(text, model, tokenizer)
print(result)
if __name__ == "__main__":
main() |