Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -11,10 +11,10 @@ model_id = "CompVis/stable-diffusion-v1-4"
|
|
11 |
device = "cuda"
|
12 |
|
13 |
#If you are running this code locally, you need to either do a 'huggingface-cli login` or paste your User Access Token from here https://huggingface.co/settings/tokens into the use_auth_token field below.
|
14 |
-
pipe = StableDiffusionPipeline.from_pretrained(model_id, use_auth_token="
|
15 |
pipe = pipe.to(device)
|
16 |
#When running locally, you won`t have access to this, so you can remove this part
|
17 |
-
word_list_dataset = load_dataset("stabilityai/word-list", data_files="list.txt", use_auth_token="
|
18 |
word_list = word_list_dataset["train"]['text']
|
19 |
|
20 |
def infer(prompt, samples, steps, scale, seed):
|
|
|
11 |
device = "cuda"
|
12 |
|
13 |
#If you are running this code locally, you need to either do a 'huggingface-cli login` or paste your User Access Token from here https://huggingface.co/settings/tokens into the use_auth_token field below.
|
14 |
+
pipe = StableDiffusionPipeline.from_pretrained(model_id, use_auth_token="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", revision="fp16", torch_dtype=torch.float16)
|
15 |
pipe = pipe.to(device)
|
16 |
#When running locally, you won`t have access to this, so you can remove this part
|
17 |
+
word_list_dataset = load_dataset("stabilityai/word-list", data_files="list.txt", use_auth_token="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")
|
18 |
word_list = word_list_dataset["train"]['text']
|
19 |
|
20 |
def infer(prompt, samples, steps, scale, seed):
|