Ostixe360 commited on
Commit
26f31da
1 Parent(s): 2311dd1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ This Model is a test to combine [Jamba](https://huggingface.co/ai21labs/Jamba-v0
25
 
26
  The goal is to developpe and test if this kind of architectures have not too much quality loss for a fast inference.
27
 
28
- Only 17.8M parameter over 1000 is in bf16 precision
29
 
30
 
31
  - **Model type:** Mixture of attention head mixture of depth and mixture of expert 1.58bit linear layers **excepted for attention layer**
 
25
 
26
  The goal is to developpe and test if this kind of architectures have not too much quality loss for a fast inference.
27
 
28
+ Only 17.8M parameter over 1025 is in bf16 precision wich is ~ 1.7% of the total number of parameters
29
 
30
 
31
  - **Model type:** Mixture of attention head mixture of depth and mixture of expert 1.58bit linear layers **excepted for attention layer**