runtime error
Exit code: 1. Reason: ............................................................. llama_new_context_with_model: n_seq_max = 1 llama_new_context_with_model: n_ctx = 8000 llama_new_context_with_model: n_ctx_per_seq = 8000 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 10000.0 llama_new_context_with_model: freq_scale = 1 llama_new_context_with_model: n_ctx_pre_seq (8000) > n_ctx_train (4096) -- possible training context overflow llama_kv_cache_init: CPU KV buffer size = 6250.00 MiB llama_new_context_with_model: KV self size = 6250.00 MiB, K (f16): 3125.00 MiB, V (f16): 3125.00 MiB llama_new_context_with_model: CPU output buffer size = 0.12 MiB llama_new_context_with_model: CPU compute buffer size = 680.63 MiB llama_new_context_with_model: graph nodes = 1286 llama_new_context_with_model: graph splits = 1 CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | LLAMAFILE = 1 | OPENMP = 1 | AARCH64_REPACK = 1 | Model metadata: {'tokenizer.ggml.unknown_token_id': '0', 'tokenizer.ggml.eos_token_id': '2', 'general.architecture': 'llama', 'llama.context_length': '4096', 'general.name': 'LLaMA v2', 'llama.embedding_length': '5120', 'llama.feed_forward_length': '13824', 'llama.attention.layer_norm_rms_epsilon': '0.000010', 'llama.rope.dimension_count': '128', 'llama.attention.head_count': '40', 'tokenizer.ggml.bos_token_id': '1', 'llama.block_count': '40', 'llama.attention.head_count_kv': '40', 'general.quantization_version': '2', 'tokenizer.ggml.model': 'llama', 'general.file_type': '15'} Traceback (most recent call last): File "/app/app.py", line 26, in <module> mychatbot = gr.Chatbot( File "/usr/local/lib/python3.10/site-packages/gradio/component_meta.py", line 179, in wrapper return fn(self, **kwargs) TypeError: Chatbot.__init__() got an unexpected keyword argument 'likeable'
Container logs:
Fetching error logs...