language:
- en
- ko
pipeline_tag: text-generation
tags:
- facebook
- meta
- pytorch
- llama
- llama-3
- llama-3-ko
license: llama3
license_name: llama3
license_link: https://llama.meta.com/llama3/license
- Original model is beomi/Llama-3-Open-Ko-8B
- quantized using llama.cpp
Ollama
Modelfile
FROM Llama-3-Open-Ko-8B-Q8_0.gguf
TEMPLATE """{{- if .System }}
<s>{{ .System }}</s>
{{- end }}
<s>Human:
{{ .Prompt }}</s>
<s>Assistant:
"""
SYSTEM """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions."""
PARAMETER temperature 0
PARAMETER num_predict 3000
PARAMETER num_ctx 4096
PARAMETER stop <s>
PARAMETER stop </s>
Update @ 2024.04.24: Release Llama-3-Open-Ko-8B model & Llama-3-Open-Ko-8B-Instruct-preview
Model Details
Llama-3-Open-Ko-8B
The Llama-3-Open-Ko-8B model is a continued pretrained language model based on the Llama-3-8B framework. This model is trained with over 60GB of deduplicated texts sourced from publicly available resources. With the new Llama-3 tokenizer, the model has been pretrained with more than 17.7B tokens, which is slightly more than that processed by the Korean tokenizer of Llama-2. Training was conducted on a TPUv5e-256, supported by Google's TRC program.
Llama-3-Open-Ko-8B-Instruct-preview
The Instruction model, named Llama-3-Open-Ko-8B-Instruct-preview, incorporates concepts from the Chat Vector paper. This model is a preview and has not been fine-tuned with any Korean instruction set, making it a strong starting point for developing new chat and instruct models.
Meta Llama-3
Developed and released by Meta, the Meta Llama 3 family of large language models (LLMs) are optimized for dialogue use cases and excel across common industry benchmarks, emphasizing helpfulness and safety.
Model Developers: Junbum Lee (Beomi)
Variations: Llama-3-Open-Ko is available in one configuration — 8B.
Input/Output: Models accept text input and generate text and code.
Model Architecture: Llama 3 utilizes an optimized transformer architecture.
Training Data | Params | Context length | GQA | Token count | Knowledge cutoff | |
Llama-3-Open-Ko | Same as Open-Solar-Ko Dataset | 8B | 8k | Yes | 17.7B+ | Jun, 2023 |
*Dataset list available here
Intended Use
Commercial and Research Applications: Llama 3 is designed for use in English, tailored for assistant-like chat in its instruction-tuned models, while the pretrained models are versatile across various natural language generation tasks.
Out-of-scope: Any use violating applicable laws, regulations, or the Acceptable Use Policy and Llama 3 Community License is prohibited.
Responsibility & Safety
Meta's commitment to Responsible AI includes steps to limit misuse and harm while supporting the open source community. Developers are encouraged to implement safety best practices and use resources like Meta Llama Guard 2 and Code Shield to tailor safety needs specifically to their use cases.
Responsible Release
Following a rigorous process against misuse, we ensure all safety and ethical guidelines are adhered to, as detailed in our Responsible Use Guide.
Ethical Considerations and Limitations
Llama 3 is built on the principles of openness, inclusivity, and helpfulness, designed to be accessible and valuable across diverse backgrounds and use cases. Developers should undertake thorough safety testing and tuning for specific applications before deployment.
Citation Instructions
Llama-3-Open-Ko @article{llama3openko, title={Llama-3-Open-Ko}, author={L, Junbum}, year={2024}, url={https://huggingface.co/beomi/Llama-3-Open-Ko-8B} }
Original Llama-3 @article{llama3modelcard, title={Llama 3 Model Card}, author={AI@Meta}, year={2024}, url={https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md} }