--- library_name: transformers language: - en - ko pipeline_tag: translation tags: - llama-3-ko license: mit datasets: - recipes --- ### Model Card for Model ID ### Model Details Model Card: llama3-pre1-ds-lora3 with Fine-Tuning Model Overview Model Name: llama3-pre1-ds-lora3 Model Type: Transformer-based Language Model Model Size: 8 billion parameters by: 4yo1 Languages: English and Korean ### Model Description llama3-pre1-ds-lora3 is a language model pre-trained on a diverse corpus of English and Korean texts. This fine-tuning approach allows the model to adapt to specific tasks or datasets with a minimal number of additional parameters, making it efficient and effective for specialized applications. ### how to use - sample code ```python from transformers import AutoConfig, AutoModel, AutoTokenizer config = AutoConfig.from_pretrained("4yo1/llama3-pre1-ds-lora3") model = AutoModel.from_pretrained("4yo1/llama3-pre1-ds-lora3") tokenizer = AutoTokenizer.from_pretrained("4yo1/llama3-pre1-ds-lora3") ``` datasets: - recipes license: mit