LH-Tech-AI's picture
Create use.py
0ee2d12 verified
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
path = "spark_v4_fp16_final"
tokenizer = AutoTokenizer.from_pretrained(path)
model = AutoModelForCausalLM.from_pretrained(path).to("cuda")
prompts = [
"Artificial Intelligence is",
"The main concept of physics is",
"In the year 1969, "
]
for prompt in prompts:
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(
**inputs,
max_new_tokens=200,
do_sample=True,
top_k=25,
temperature=0.8,
pad_token_id=tokenizer.eos_token_id
)
print(f"PROMPT: {prompt}")
print(f"OUTPUT: {tokenizer.decode(outputs[0], skip_special_tokens=True)}\n{'-'*40}")