runtime error

Space failed. Exit code: 1. Reason: mpl return forward_call(*args, **kwargs) File "/home/user/.cache/huggingface/modules/transformers_modules/OpenAssistant/falcon-7b-sft-top1-696/9f95b9e269c1d99c29f70be8d5ef3d0fb1bd1c30/modelling_RW.py", line 648, in forward outputs = block( File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl return forward_call(*args, **kwargs) File "/home/user/.cache/huggingface/modules/transformers_modules/OpenAssistant/falcon-7b-sft-top1-696/9f95b9e269c1d99c29f70be8d5ef3d0fb1bd1c30/modelling_RW.py", line 381, in forward layernorm_output = self.input_layernorm(hidden_states) File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl return forward_call(*args, **kwargs) File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 190, in forward return F.layer_norm( File "/opt/conda/lib/python3.9/site-packages/torch/nn/functional.py", line 2515, in layer_norm return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled) RuntimeError: "LayerNormKernelImpl" not implemented for 'Half' 2023-10-22T08:40:47.202144Z ERROR warmup{max_input_length=1024 max_prefill_tokens=4096 max_total_tokens=2048}:warmup: text_generation_client: router/client/src/lib.rs:33: Server error: "LayerNormKernelImpl" not implemented for 'Half' Error: Warmup(Generation("\"LayerNormKernelImpl\" not implemented for 'Half'")) 2023-10-22T08:40:47.248685Z ERROR text_generation_launcher: Webserver Crashed 2023-10-22T08:40:47.248756Z  INFO text_generation_launcher: Shutting down shards 2023-10-22T08:40:47.994827Z  INFO shard-manager: text_generation_launcher: Shard terminated rank=0 Error: WebserverFailed

Container logs:

Fetching error logs...