--- license: cc-by-nc-4.0 language: - en pipeline_tag: text-generation datasets: - Skylion007/openwebtext - Locutusque/TM-DATA inference: parameters: do_sample: True temperature: 0.7 top_p: 0.2 top_k: 7 max_new_tokens: 250 repetition_penalty: 1.15 --- # Training This model was trained on two datasets, shown in this model page. - Skylion007/openwebtext: 1,000,000 examples at a batch size of 32-4096 (1 epoch) - Locutusque/TM-DATA: All examples at a batch size of 12288 (3 epochs) Training took approximately 500 GPU hours on a single Titan V. # Metrics You can look at the training metrics here: https://wandb.ai/locutusque/TinyMistral-V2/runs/g0rvw6wc # License This model is released under the cc-by-nc-4.0 license. This is because the data used to train this model is under this same license.