PEFT
Not-For-All-Audiences
Edit model card

Pippa-13b-qlora

This is a repository of my Llama-2-13b Qlora checkpoints of the PIPPA-13b-ShareGPT dataset.

You can read more about the dataset on its relevant page. It's a ShareGPT reformat of the PIPPA dataset by PygmalionAI. The reformat was done to allow for axolotl compatability.

Architecture

  • Model Architecture: Llama-2-13b
  • Training Algorithm: QLora
  • Dataset Used: PIPPA-ShareGPT (pippa_sharegpt_trimmed.jsonl)

Training Details

  • Dataset: PIPPA-ShareGPT
  • Datset type: ShareGPT
  • Training Parameters: See Here
  • Training Environment: Axolotl
  • sequence_len: 4096

Instruct Format

ShareGPT gets converted to vicuna format. The dataset uses modified roles of USER and CHARACTER instead of USER and ASSISTANT.

SYSTEM: Enter roleplay mode...
USER: {prompt}
CHARACTER:

Notes

This Qlora was produced as an experiment to see how the public version of PIPPA can affect a model. As a result, I have no idea if this lora is of great quality or absolute garbage.

Acknowledgments

Thanks to:

  • PygmalionAI: The creators of the PIPPA dataset
  • Axolotl: Finetuning suite

Donate?

All my infrastructure and cloud expenses are paid out of pocket. If you'd like to donate, you can do so here: https://ko-fi.com/kingbri

You should not feel obligated to donate, but if you do, I'd appreciate it.

Axolotl stuff

Training procedure

The following bitsandbytes quantization config was used during training:

  • quant_method: bitsandbytes
  • load_in_8bit: False
  • load_in_4bit: True
  • llm_int8_threshold: 6.0
  • llm_int8_skip_modules: None
  • llm_int8_enable_fp32_cpu_offload: False
  • llm_int8_has_fp16_weight: False
  • bnb_4bit_quant_type: nf4
  • bnb_4bit_use_double_quant: True
  • bnb_4bit_compute_dtype: bfloat16

Framework versions

  • PEFT 0.6.0.dev0
Downloads last month
1
Unable to determine this model’s pipeline type. Check the docs .