from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline tokenizer = AutoTokenizer.from_pretrained("cv43/llmpot") model = AutoModelForSeq2SeqLM.from_pretrained("cv43/llmpot") pipe = pipeline("text2text-generation", model=model, tokenizer=tokenizer, framework="pt") request = "02b10000000b00100000000204ffffffff" result = pipe(request) print(f"Request: {request}, Response: {result[0]['generated_text']}")