alvp commited on
Commit
a04fbcb
1 Parent(s): a43ecc5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -3
README.md CHANGED
@@ -1,3 +1,46 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - bertin-project/bonanza-hf
5
+ - bertin-project/zenobia-instruct-hf
6
+ language:
7
+ - es
8
+ - ca
9
+ pipeline_tag: text-generation
10
+ ---
11
+ # Gromenauer-7B-Instruct
12
+
13
+ <div align=center>
14
+ <img alt="gromenauer-7B logo" src="https://huggingface.co/bertin-project/Gromenauer-7B/resolve/main/images/gromenauer.png" width="200px">
15
+ </div>
16
+
17
+ ## Overview
18
+ Gromenauer-7B-Instruct is an instruct fine-tuned version of the [bertin-project/Gromenauer-7B](https://huggingface.co/bertin-project/Gromenauer-7B) model using the [bertin-project/bonanza-hf](https://huggingface.co/datasets/bertin-project/bonanza-hf) and [bertin-project/zenobia-instruct-hf](https://huggingface.co/datasets/bertin-project/zenobia-instruct-hf) datasets.
19
+
20
+ ## Model Details
21
+
22
+ - **Model Type**: Mistral
23
+ - **Sequence Length**: 8192
24
+ - **Hidden Dimension**: 4096
25
+ - **Intermediate Dimension**: 14336
26
+ - **Number of Layers**: 32
27
+ - **Number of Attention Heads**: 32
28
+ - **Number of Key-Value Heads**: 8
29
+ - **Activation Function**: SiLU
30
+ - **Initializer Range**: 0.02
31
+ - **Layer Norm Epsilon**: 1.0e-05
32
+ - **Use Flash Attention**: Yes
33
+ - **Gradient Checkpointing**: Enabled (Block Size: 5)
34
+ - **Sliding Window Attention**: 4096
35
+ - **Use Bias**: No
36
+
37
+ ## Training Details
38
+
39
+ - **Tokenizer**: [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta)
40
+ - **Batch Size**: 512
41
+ - **Learning Rate**: 1e-5
42
+ - **Optimizer**: Adam with beta1=0.9, beta2=0.95, epsilon=1e-8
43
+ - **Weight Decay**: 0.1
44
+ - **Warmup Steps**: 200
45
+ - **Learning Rate Schedule**: Cosine
46
+ - **Number of Training Epochs**: 5