Update README.md
Browse files
README.md
CHANGED
|
@@ -10,7 +10,7 @@ tags:
|
|
| 10 |
|
| 11 |
The Jamba-DNA-v1-134M-hg38 Large Language Model (LLM) is a pretrained generative DNA sequence model with 134M parameters.
|
| 12 |
It is derived from Jamba model, which was simplified for DNA: the number of layers and the hidden size were reduced.
|
| 13 |
-
The model was pretrained using 100kb DNA sequences from the hg38 human genome assembly.
|
| 14 |
In comparison, Mistral DNA models were trained using 10 kb DNA sequences, so there are able to deal with a small genomic sequence context.
|
| 15 |
|
| 16 |
## Model Architecture
|
|
|
|
| 10 |
|
| 11 |
The Jamba-DNA-v1-134M-hg38 Large Language Model (LLM) is a pretrained generative DNA sequence model with 134M parameters.
|
| 12 |
It is derived from Jamba model, which was simplified for DNA: the number of layers and the hidden size were reduced.
|
| 13 |
+
The model was pretrained using **100kb** DNA sequences from the hg38 human genome assembly.
|
| 14 |
In comparison, Mistral DNA models were trained using 10 kb DNA sequences, so there are able to deal with a small genomic sequence context.
|
| 15 |
|
| 16 |
## Model Architecture
|