runtime error

The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`. 0it [00:00, ?it/s] 0it [00:00, ?it/s] Traceback (most recent call last): File "/home/user/app/app.py", line 23, in <module> pipe = DiffusionPipeline.from_pretrained("runwayml/stable-diffusion-inpainting", dtype=torch.float16, revision="fp16", use_auth_token=auth_token).to(device) File "/home/user/.local/lib/python3.10/site-packages/diffusers/pipelines/pipeline_utils.py", line 1041, in from_pretrained cached_folder = cls.download( File "/home/user/.local/lib/python3.10/site-packages/diffusers/pipelines/pipeline_utils.py", line 1584, in download info = model_info( File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 118, in _inner_fn return fn(*args, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 1686, in model_info headers = self._build_hf_headers(token=token) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 5719, in _build_hf_headers return build_hf_headers( File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 118, in _inner_fn return fn(*args, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_headers.py", line 121, in build_hf_headers token_to_send = get_token_to_send(token) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_headers.py", line 153, in get_token_to_send raise LocalTokenNotFoundError( huggingface_hub.utils._headers.LocalTokenNotFoundError: Token is required (`token=True`), but no token found. You need to provide a token or be logged in to Hugging Face with `huggingface-cli login` or `huggingface_hub.login`. See https://huggingface.co/settings/tokens.

Container logs:

Fetching error logs...