runtime error

Exit code: 1. Reason: .......... llama_new_context_with_model: n_seq_max = 1 llama_new_context_with_model: n_ctx = 8000 llama_new_context_with_model: n_ctx_per_seq = 8000 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_new_context_with_model: n_ctx_per_seq (8000) < n_ctx_train (32768) -- the full capacity of the model will not be utilized llama_kv_cache_init: CPU KV buffer size = 1000.00 MiB llama_new_context_with_model: KV self size = 1000.00 MiB, K (f16): 500.00 MiB, V (f16): 500.00 MiB llama_new_context_with_model: CPU output buffer size = 0.12 MiB llama_new_context_with_model: CPU compute buffer size = 547.63 MiB llama_new_context_with_model: graph nodes = 1030 llama_new_context_with_model: graph splits = 1 CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | LLAMAFILE = 1 | OPENMP = 1 | AARCH64_REPACK = 1 | Model metadata: {'tokenizer.ggml.unknown_token_id': '0', 'tokenizer.ggml.eos_token_id': '2', 'general.architecture': 'llama', 'llama.rope.freq_base': '10000.000000', 'llama.context_length': '32768', 'general.name': '.', 'llama.embedding_length': '4096', 'llama.feed_forward_length': '14336', 'llama.attention.layer_norm_rms_epsilon': '0.000010', 'llama.rope.dimension_count': '128', 'tokenizer.ggml.bos_token_id': '1', 'llama.attention.head_count': '32', 'llama.block_count': '32', 'llama.attention.head_count_kv': '8', 'general.quantization_version': '2', 'tokenizer.ggml.model': 'llama', 'general.file_type': '17'} Traceback (most recent call last): File "/app/app.py", line 20, in <module> mychatbot = gr.Chatbot( File "/usr/local/lib/python3.10/site-packages/gradio/component_meta.py", line 179, in wrapper return fn(self, **kwargs) TypeError: Chatbot.__init__() got an unexpected keyword argument 'likeable'

Container logs:

Fetching error logs...