File size: 434 Bytes
d97f0ed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e404693
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)

model = AutoModelForSequenceClassification.from_pretrained('techthiyanes/Bert_Bahasa_Sentiment')

inputs = tokenizer("saya tidak", return_tensors="pt")

labels = torch.tensor([1]).unsqueeze(0) 

outputs = model(**inputs, labels=labels)

loss = outputs.loss

logits = outputs.logits

outputs
hello