runtime error

οΏ½οΏ½β–ˆβ–ˆβ–‰ | 8/9 [00:07<00:00, 1.12it/s] Loading pipeline components...: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 9/9 [00:07<00:00, 1.21it/s] Loading pipeline components...: 0%| | 0/9 [00:00<?, ?it/s] Loading pipeline components...: 22%|β–ˆβ–ˆβ– | 2/9 [00:01<00:06, 1.16it/s] Loading pipeline components...: 44%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 4/9 [00:03<00:04, 1.15it/s] Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] Loading checkpoint shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:01<00:00, 1.06it/s] Loading checkpoint shards: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:01<00:00, 1.06it/s] Loading pipeline components...: 89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 8/9 [00:06<00:00, 1.35it/s] Loading pipeline components...: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 9/9 [00:06<00:00, 1.43it/s] You are using the default legacy behaviour of the <class 'transformers.models.t5.tokenization_t5.T5Tokenizer'>. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. Traceback (most recent call last): File "/home/user/app/app.py", line 44, in <module> def text2img(prompt, enhance_prompt, negative_prompt, num_inference_steps, height, width, guidance_scale, seed, num_images_per_prompt, progress=gr.Progress(track_tqdm=True)): File "/usr/local/lib/python3.10/site-packages/spaces/zero/decorator.py", line 113, in _GPU client.startup_report() File "/usr/local/lib/python3.10/site-packages/spaces/zero/client.py", line 45, in startup_report raise RuntimeError("Error while initializing ZeroGPU: Unknown") RuntimeError: Error while initializing ZeroGPU: Unknown

Container logs:

Fetching error logs...