Edit model card
def scaling(x, min_x, max_x, r1, r2):
    # Scale data x (n_samples x 1) to [r1, r2]
    x_s = x
    x_s = (x_s - min_x) * (r2 - r1) / (max_x - min_x)
    x_s = r1 + x_s
    return x_s

def descaling(x_s, min_x, max_x, r1, r2):
    # Re-scale data x (n_samples x 1) to [min_x, max_x]
    x = x_s
    x = (x - r1) * (max_x - min_x) / (r2 - r1) + min_x
    return x

# Inference example
with torch.no_grad():
    x = "They are equally important, absolutely, and just as real as each other."
    x = tokenizer([x], return_tensors="pt", add_special_tokens=True, padding=True)
    y_hat = model(**x.to(device)).logits
    y_hat = torch.tanh(y_hat).cpu()

l_hat = descaling(y_hat, 1, 7, -1, 1)[0].numpy() 
print(l_hat)
# [C, O, E, A, S]
# [6.0583944 4.4941516 1.6538751 5.5261126 4.725995 ]
Downloads last month
4
Safetensors
Model size
125M params
Tensor type
F32
·