Text Generation
Transformers
PyTorch
Safetensors
gpt2
stable-diffusion
prompt-generator
arxiv:2210.14140
Inference Endpoints
text-generation-inference
FredZhang7 commited on
Commit
25733c4
1 Parent(s): b61e4ef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -44,7 +44,7 @@ tokenizer.add_special_tokens({'pad_token': '[PAD]'})
44
  model = GPT2LMHeadModel.from_pretrained('FredZhang7/distilgpt2-stable-diffusion-v2')
45
 
46
  prompt = r'a cat sitting' # the beginning of the prompt
47
- temperature = 0.9 # a higher temperature will produce more diverse results, but with a higher risk of less coherent text.
48
  top_k = 8 # the number of tokens to sample from at each step
49
  max_length = 80 # the maximum number of tokens for the output of the model
50
  repitition_penalty = 1.2 # the penalty value for each repetition of a token
 
44
  model = GPT2LMHeadModel.from_pretrained('FredZhang7/distilgpt2-stable-diffusion-v2')
45
 
46
  prompt = r'a cat sitting' # the beginning of the prompt
47
+ temperature = 0.9 # a higher temperature will produce more diverse results, but with a higher risk of less coherent text
48
  top_k = 8 # the number of tokens to sample from at each step
49
  max_length = 80 # the maximum number of tokens for the output of the model
50
  repitition_penalty = 1.2 # the penalty value for each repetition of a token