--- license: mit datasets: - kyujinpy/KOR-OpenOrca-Platypus-v3 --- ### Developed by chPark ### Training Strategy We fine-tuned this model based on [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1-deprecated) ### Run the model ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "realPCH/ko_solra_merge" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id) text = "[INST] Put instruction here. [/INST]" inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=20) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ```