```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM model_id = "/home/mwiti/mwitiderrick/output_oneshot" model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype=torch.float16) tokenizer = AutoTokenizer.from_pretrained(model_id) inputs = tokenizer("Hello my name is", return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=20) print(tokenizer.batch_decode(outputs)[0]) """ Hello my name is Katie and I am a student at the University of Gloucestershire. I am currently studying """ ```