metadata
license: apache-2.0
datasets:
- csebuetnlp/xlsum
- SGaleshchuk/XL_SUM_ukr_synthetic_hallucinations
language:
- uk
- en
tags:
- text-generation-inference
pipeline_tag: text-generation
The following training arguments used for Llama-2 finetuning with Ukrainian corpora pf XL-SUM:
- learning-rate=2e-4,
- maximum number of tokens=512,
- 15 epochs. Lora perf arguments:
- rank = 32,
- lora-alpha=16,
- dropout = 0.1.