daha-kot's picture
Update README.md
08a4a14
|
raw
history blame
2.49 kB
metadata
library_name: peft
base_model: core42/jais-13b
license: mit
datasets:
  - arbml/Ashaar_dataset
language:
  - ar
metrics:
  - perplexity
  - bertscore

Model Card for Model ID

Fine-tuned using QLoRA for poem generation task.

Model Description

We utilize Ashaar dataset and fine-tune the model to generate poems.

The input to the model is structred as follows:

'''

### Instruction: Generate a poem based on the following title, and the given era:

### Input: {Title of a poem + poet era}

### Response: {Poem verses}

'''

  • Developed by: Abdelrahman ’Boda’ Sadallah, Anastasiia Demidova, Daria Kotova
  • Model type: Causal LM
  • Language(s) (NLP): Arabic
  • Finetuned from model [optional]: core42/jais-13b

Model Sources

Uses

The model is the result of our AI project. If you intend to use it, please, refer to the repo.

Recommendations

For improving stories generation, you can play parameters: temeperature, top_p/top_k, repetition_penalty, etc.

Training Details

Training Data

Link to the dataset on huggungface: https://huggingface.co/datasets/arbml/ashaar.

Evaluation

Testing Data, Factors & Metrics

Test split of the same dataset.

Metrics

We are using perplexity and BERTScore.

Results

Perplexity: 48.3125

BERTScore: 59.33

Training procedure

The following bitsandbytes quantization config was used during training:

  • quant_method: bitsandbytes
  • load_in_8bit: False
  • load_in_4bit: True
  • llm_int8_threshold: 6.0
  • llm_int8_skip_modules: None
  • llm_int8_enable_fp32_cpu_offload: False
  • llm_int8_has_fp16_weight: False
  • bnb_4bit_quant_type: fp4
  • bnb_4bit_use_double_quant: False
  • bnb_4bit_compute_dtype: float32

Framework versions

  • PEFT 0.6.0.dev0