Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

K2S3-SOLAR-11b-v4.0 - GGUF

Original model description:

license: cc-by-nc-4.0 language: - ko


Developed by :

  • K2S3

Model Number:

  • K2S3-SOLAR-11b-v4.0

Base Model :

Training Data

  • The training data for this model includes the Standard Korean Dictionary, training data from KULLM at Korea University, abstracts of master's and doctoral theses, Korean language samples from AI Hub, alpaca-gpt4-data, and samples from The OpenOrca Dataset.
  • ์ด ๋ชจ๋ธ์˜ ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ์—๋Š” ํ‘œ์ค€๊ตญ์–ด๋Œ€์‚ฌ์ „, ๊ณ ๋ ค๋Œ€ํ•™๊ต KULLM์—์„œ ์ œ๊ณตํ•œ ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ, ์„์‚ฌ ๋ฐ ๋ฐ•์‚ฌํ•™์œ„ ๋…ผ๋ฌธ์˜ ์ดˆ๋ก, AI Hub์—์„œ ์ œ๊ณตํ•œ ํ•œ๊ตญ์–ด ๋ฐ์ดํ„ฐ ์ƒ˜ํ”Œ, alpaca-gpt4-data, ๊ทธ๋ฆฌ๊ณ  OpenOrca Dataset์—์„œ ์ œ๊ณตํ•œ ์ƒ˜ํ”Œ๋“ค์ด ํฌํ•จ๋ฉ๋‹ˆ๋‹ค.

Training Method

  • This model was fine-tuned on the "upstage/SOLAR-10.7B-v1.0" base model using a full parameter tuning method with SFT (Supervised Fine-Tuning).
  • ์ด ๋ชจ๋ธ์€ "upstage/SOLAR-10.7B-v1.0" ๊ธฐ๋ฐ˜ ๋ชจ๋ธ์„ SFT๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ์ „์ฒด ํŒŒ๋ผ๋ฏธํ„ฐ ์กฐ์ • ๋ฐฉ๋ฒ•์œผ๋กœ ๋ฏธ์„ธ์กฐ์ •๋˜์—ˆ์Šต๋‹ˆ๋‹ค.

Hardware

  • Hardware: Utilized two A100 (80G*2EA) GPUs for training.
  • Training Factors: This model was fine-tuned with SFT, using the HuggingFace SFTtrainer and applied fsdp.
  • ์ด ๋ชจ๋ธ์€ SFT๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ HuggingFace SFTtrainer์™€ fsdp๋ฅผ ์ ์šฉํ•˜์—ฌ ๋ฏธ์„ธ์กฐ์ •๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
Downloads last month
260
GGUF
Model size
10.7B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .