YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

llama3-8B-DarkIdol-2.1-Uncensored-32K - GGUF

Original model description:

license: llama3 language: - en - ja - zh tags: - roleplay - llama3 - sillytavern - idol

Special Thanks:

fast quantizations

Model Description:

The module combination has been readjusted to better fulfill various roles and has been adapted for mobile phones.

image/png

Chang Log

2024-06-26

  • 32k

2024-06-26

  • ไน‹ๅ‰็‰ˆๆœฌ็š„่ฟญไปฃๅคชๅคšไบ†,ๅทฒ็ปๅผ€ๅง‹ๅ‡บ็Žฐ่ฟ‡ๆ‹Ÿๅˆ็Žฐ่ฑก.้‡ๆ–ฐไฝฟ็”จไบ†ๆ–ฐ็š„ๅทฅ่‰บ้‡ๆ–ฐๅˆถไฝœๆจกๅž‹,่™ฝ็„ถๅˆถไฝœๅคๆ‚ไบ†,็ป“ๆžœๅพˆๅฅฝ,ๆ–ฐ็š„่ฟญไปฃๅทฅ่‰บๅฆ‚ๅ›พ
  • The previous version had undergone excessive iterations, resulting in overfitting. We have recreated the model using a new process, which, although more complex to produce, has yielded excellent results. The new iterative process is depicted in the figure.

image/png

Questions

  • The model's response results are for reference only, please do not fully trust them.
  • I am unable to test Japanese and Korean parts very well. Based on my testing, Korean performs excellently, but sometimes Japanese may have furigana (if anyone knows a good Japanese language module, - I need to replace the module for integration).
  • With the new manufacturing process, overfitting and crashes have been reduced, but there may be new issues, so please leave a message if you encounter any.
  • testing with other tools is not comprehensive.but there may be new issues, so please leave a message if you encounter any.
  • The range between 32K and 64K was not tested, and the approach was somewhat casual. I didn't expect the results to be exceptionally good.

้—ฎ้ข˜

  • ๆจกๅž‹ๅ›žๅค็ป“ๆžœไป…ไพ›ๅ‚่€ƒ,่ฏทๅ‹ฟๅฎŒๅ…จ็›ธไฟก
  • ๆ—ฅ่ฏญ,้Ÿฉ่ฏญ้ƒจๅˆ†ๆˆ‘ๆฒกๅŠžๆณ•่ฟ›่กŒๅพˆๅฅฝ็š„ๆต‹่ฏ•,ๆ นๆฎๆˆ‘ๆต‹่ฏ•ๆƒ…ๅ†ต,้Ÿฉ่ฏญ่กจ็Žฐ็š„ๅพˆๅฅฝ,ๆ—ฅ่ฏญๆœ‰ๆ—ถๅ€™ไผšๅ‡บ็Žฐๆณจ้Ÿณ(่ฐ็Ÿฅ้“ๅฅฝ็š„ๆ—ฅๆ–‡่ฏญ่จ€ๆจกๅ—,ๆˆ‘้œ€่ฆๆขๆจกๅ—้›†ๆˆ)
  • ๆ–ฐๅทฅ่‰บๅˆถไฝœ,่ฟ‡ๆ‹Ÿๅˆ็Žฐ่ฑกๅ’Œๅดฉๆบƒๅ‡ๅฐ‘ไบ†,ๅฏ่ƒฝไผšๆœ‰ๆ–ฐ็š„้—ฎ้ข˜,็ขฐๅˆฐไบ†่ฏท็ป™ๆˆ‘็•™่จ€
  • 32K-64kๅŒบ้—ดๆฒกๆœ‰ๆต‹่ฏ•,ๅš็š„ๆœ‰็‚น้šๆ„,ๆฒกๆƒณๅˆฐ็ป“ๆžœ็‰นๅˆซ็š„ๅฅฝ
  • ๅ…ถไป–ๅทฅๅ…ท็š„ๆต‹่ฏ•ไธๅฎŒๅ–„

Stop Strings

    stop = [
      "## Instruction:",
      "### Instruction:",
      "<|end_of_text|>",
      "  //:",
      "</s>",
      "<3```",
      "### Note:",
      "### Input:",
      "### Response:",
      "### Emoticons:"
    ],

Model Use

character

If you want to use vision functionality:

  • You must use the latest versions of Koboldcpp.

To use the multimodal capabilities of this model and use vision you need to load the specified mmproj file, this can be found inside this model repo. Llava MMProj

  • You can load the mmproj by using the corresponding section in the interface: image/png

Thank you:

To the authors for their hard work, which has given me more options to easily create what I want. Thank you for your efforts. - Hastagaras - Gryphe - cgato - ChaoticNeutrals - mergekit - merge - transformers - llama - Nitral-AI - MLP-KTLim - rinna - hfl - Rupesh2 - stephenlzc - theprint - Sao10K - turboderp - TheBossLevel123 - winglian - .........

base_model:

  • Nitral-AI/Hathor_Fractionate-L3-8B-v.05
  • Hastagaras/Jamet-8B-L3-MK.V-Blackroot
  • turboderp/llama3-turbcat-instruct-8b
  • aifeifei798/Meta-Llama-3-8B-Instruct
  • Sao10K/L3-8B-Stheno-v3.3-32K
  • TheBossLevel123/Llama3-Toxic-8B-Float16
  • cgato/L3-TheSpice-8b-v0.8.3 library_name: transformers tags:
  • mergekit
  • merge

llama3-8B-DarkIdol-1.3.1

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using aifeifei798/Meta-Llama-3-8B-Instruct as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: Sao10K/L3-8B-Stheno-v3.3-32K
  - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
  - model: cgato/L3-TheSpice-8b-v0.8.3
  - model: Nitral-AI/Hathor_Fractionate-L3-8B-v.05
  - model: TheBossLevel123/Llama3-Toxic-8B-Float16
  - model: turboderp/llama3-turbcat-instruct-8b
  - model: aifeifei798/Meta-Llama-3-8B-Instruct
merge_method: model_stock
base_model: aifeifei798/Meta-Llama-3-8B-Instruct
dtype: bfloat16

base_model:

  • hfl/llama-3-chinese-8b-instruct-v3
  • rinna/llama-3-youko-8b
  • MLP-KTLim/llama-3-Korean-Bllossom-8B library_name: transformers tags:
  • mergekit
  • merge

llama3-8B-DarkIdol-1.3.2

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using ./llama3-8B-DarkIdol-1.3.1 as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: hfl/llama-3-chinese-8b-instruct-v3
  - model: rinna/llama-3-youko-8b
  - model: MLP-KTLim/llama-3-Korean-Bllossom-8B
  - model: ./llama3-8B-DarkIdol-1.3.1
merge_method: model_stock
base_model: ./llama3-8B-DarkIdol-1.3.1
dtype: bfloat16

base_model:

  • theprint/Llama-3-8B-Lexi-Smaug-Uncensored
  • Rupesh2/OrpoLlama-3-8B-instruct-uncensored
  • stephenlzc/dolphin-llama3-zh-cn-uncensored library_name: transformers tags:
  • mergekit
  • merge

llama3-8B-DarkIdol-2.0-Uncensored

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using ./llama3-8B-DarkIdol-1.3.2 as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: Rupesh2/OrpoLlama-3-8B-instruct-uncensored
  - model: stephenlzc/dolphin-llama3-zh-cn-uncensored
  - model: theprint/Llama-3-8B-Lexi-Smaug-Uncensored
  - model: ./llama3-8B-DarkIdol-1.3.2
merge_method: model_stock
base_model: ./llama3-8B-DarkIdol-2.0-Uncensored
dtype: bfloat16

base_model:

  • winglian/Llama-3-8b-64k-PoSE library_name: transformers tags:
  • mergekit
  • merge

llama3-8B-DarkIdol-2.1-Uncensored-32K

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the Model Stock merge method using winglian/Llama-3-8b-64k-PoSE as a base.

Models Merged

The following models were included in the merge:

  • ./llama3-8B-DarkIdol-1.3.2
  • ./llama3-8B-DarkIdol-2.0
  • ./llama3-8B-DarkIdol-1.3.1

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: ./llama3-8B-DarkIdol-1.3.1
  - model: ./llama3-8B-DarkIdol-1.3.2
  - model: ./llama3-8B-DarkIdol-2.0
  - model: winglian/Llama-3-8b-64k-PoSE
merge_method: model_stock
base_model: winglian/Llama-3-8b-64k-PoSE
dtype: bfloat16
Downloads last month
137
GGUF
Model size
8.03B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.