Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

M-DIE-M-10.7B - GGUF

Original model description:

license: cc-by-nc-sa-4.0 language: - en - ko

Data Is Everything.

To try other models(involving commercial-available model), please check out our Demo Page(🔨constructing)

This model is made by Ados based on upstage/SOLAR-10.7B-Instruct-v1.0.

Train Dataset

Dataset used for training is collected primarily from huggingface and utilized using our own translation model.

  • Language
    • KR 73%
    • EN 24%
    • Others 3%
  • Type
    • single turn QA (alpaca style) 29%
    • multi turn QA (vicuna style) 21%
    • instructed QA 26%
    • summary 12%
    • translation 12%

After collecting data, we removed low quality rows. We chose 30% high quality from raw data manually and using deduplication methods.

We also refined problematic data such as code blocks, listing, repetition and other common issues we found.

Prompt template

### System:
You are an AI assistant, please behave and help the user. Your name is OLLM(오름) by Ados(주식회사아도스), OLLM stands for On-premise LLM.

### User: On-premise LLM이 뭔가요?

### Assistant:

For more informations, please contact us.

To try other models(involving commercial-available model), please check out our Demo Page(🔨constructing)

License

@misc{kim2023solar,
      title={SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling}, 
      author={Dahyun Kim and Chanjun Park and Sanghoon Kim and Wonsung Lee and Wonho Song and Yunsu Kim and Hyeonwoo Kim and Yungi Kim and Hyeonju Lee and Jihoo Kim and Changbae Ahn and Seonghoon Yang and Sukyung Lee and Hyunbyung Park and Gyoungjin Gim and Mikyoung Cha and Hwalsuk Lee and Sunghun Kim},
      year={2023},
      eprint={2312.15166},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
Downloads last month
189
GGUF
Model size
10.7B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .