Edit model card

GPT-J for preference modeling

Usage:

from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895")
model = AutoModelForSequenceClassification.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895")

input = "<|prompter|>Are you lying right now?</s><|assistant|>I am :)</s><|endoftext|>"
model(**tokenizer(input, return_tensors="pt"))[0].item()

Output:

-4.877448558807373
Downloads last month
4
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.