runtime error

Running on device: cpu CPU threads: 8 gguf_init_from_file: GGUFv1 is no longer supported. please use a more up-to-date version error loading model: llama_model_loader: failed to load model from Llama-2-ko-7B-chat-gguf-q4_0.bin llama_load_model_from_file: failed to load model AVX = 1 | AVX2 = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | Traceback (most recent call last): File "/home/user/app/app.py", line 12, in <module> llm = Llama(model_path = 'Llama-2-ko-7B-chat-gguf-q4_0.bin', File "/home/user/.local/lib/python3.10/site-packages/llama_cpp/llama.py", line 962, in __init__ self._n_vocab = self.n_vocab() File "/home/user/.local/lib/python3.10/site-packages/llama_cpp/llama.py", line 2274, in n_vocab return self._model.n_vocab() File "/home/user/.local/lib/python3.10/site-packages/llama_cpp/llama.py", line 251, in n_vocab assert self.model is not None AssertionError

Container logs:

Fetching error logs...