SGaleshchuk's picture
Update README.md
8af07d6 verified
metadata
license: apache-2.0
datasets:
  - csebuetnlp/xlsum
  - SGaleshchuk/XL_SUM_ukr_synthetic_hallucinations
language:
  - uk
  - en
tags:
  - text-generation-inference
pipeline_tag: text-generation

The following training arguments used for Llama-2 finetuning with Ukrainian corpora pf XL-SUM:

  • learning-rate=2e-4,
  • maximum number of tokens=512,
  • 15 epochs. Lora perf arguments:
  • rank = 32,
  • lora-alpha=16,
  • dropout = 0.1.