--- datasets: - EleutherAI/pile language: - en tags: - Text Generation - pytorch - causal-lm --- ```python from transformers import GPTNeoXForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained( "afterless/reverse-pythia-160m" ) model = GPTNeoXForCausalLM.from_pretrained( "afterless/reverse-pythia-160m" ) inputs = tokenizer( "but I told him, the cheese was the best", return_token_type_ids=False, return_tensors="pt" ) inputs['input_ids'] = t.flip(inputs.input_ids, (1,)) tokens = t.flip(model.generate(**inputs), (1,)) tokenizer.decode(tokens[0]) ```