Edit model card

llama3-dnapretrain-kaniwa

This is a LoRA adapter.

The base model is the longer-context LLaMA-3-8b-Instruct developed by Gradient and Crusoe: gradientai/Llama-3-8B-Instruct-262k

The dataset was part of BYU's 2019 kaniwa (Chenopodium pallidicaule) genome, from https://genomevolution.org/coge/GenomeInfo.pl?gid=53872

The adapter was finetuned for 3 hours on an A100. The data was split into ~20k nucleotide snippets with an Alpaca like message format.

Training Notebook: https://colab.research.google.com/drive/1XZcCYGFQGtz3_AKSR4F67WYXl6DIwP4R

Sample message:

Write information about the nucleotide sequence.

### Sequence:
GCCTATAGTGTGTAGCTAATGAGCCTAGGTTATCGACCCTAATCT...

### Annotation:
Information about location in the kaniwa chromosome: >lcl|Cp5

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Genome Citation

Mangelson H, et al. The genome of Chenopodium pallidicaule: an emerging Andean super grain. Appl. Plant Sci. 2019;7:e11300. doi: 10.1002/aps3.11300

Downloads last month
0
Unable to determine this model's library. Check the docs .

Finetuned from

Collection including monsoon-nlp/llama3-dnapretrain-kaniwa