--- language: - en - zh license: llama3 library_name: transformers base_model: unsloth/llama-3-8b-bnb-4bit datasets: - erhwenkuo/alpaca-data-gpt4-chinese-zhtw pipeline_tag: text-generation tags: - llama-3 prompt_template: >- {{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|> {{ .Response }}<|eot_id|> --- # LLAMA 3 8B with capable to output Traditional Chinese ## ✨ Recommend using LMStudio for this model I tried using Ollama to run it, but it became quite delulu. So for now, I'm sticking with LMStudio :)The performance isn't actually that great, but it's capable of answering some basic questions. Sometimes it just acts really dumb though :( > LLAMA 3.1 can actually output pretty well Chinese, so this repo can be ignored.