fix `_name_or_path` in config.json (#3)
Browse files- update config.json (0a86922a8af6074096d8b9439fd6458c27b9e71c)
- add requiements (eee24f04b87735f3deaf4df649f1dd39fe176585)
Co-authored-by: LAin <not-lain@users.noreply.huggingface.co>
- README.md +6 -0
- config.json +2 -2
- requirements.txt +2 -0
README.md
CHANGED
@@ -36,6 +36,12 @@ The training setup was `4xA100's 80GB` and took ~6 hours to pretrain and ~13 hou
|
|
36 |
---
|
37 |
|
38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
## Training:
|
40 |
Training code Released !!!
|
41 |
https://github.com/tensoic/Cerule
|
|
|
36 |
---
|
37 |
|
38 |
|
39 |
+
## Installing requirements
|
40 |
+
```
|
41 |
+
pip install -qr https://huggingface.co/Tensoic/Cerule-v0.1/resolve/main/requirements.txt
|
42 |
+
```
|
43 |
+
|
44 |
+
|
45 |
## Training:
|
46 |
Training code Released !!!
|
47 |
https://github.com/tensoic/Cerule
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "Tensoic/Cerule",
|
3 |
"architectures": [
|
4 |
"CeruleGemmaForCausalLM"
|
5 |
],
|
@@ -34,7 +34,7 @@
|
|
34 |
"tokenizer_model_max_length": 2048,
|
35 |
"tokenizer_padding_side": "right",
|
36 |
"torch_dtype": "bfloat16",
|
37 |
-
"transformers_version": "4.39.
|
38 |
"tune_mm_mlp_adapter": false,
|
39 |
"use_cache": true,
|
40 |
"use_mm_proj": true,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "Tensoic/Cerule-v0.1",
|
3 |
"architectures": [
|
4 |
"CeruleGemmaForCausalLM"
|
5 |
],
|
|
|
34 |
"tokenizer_model_max_length": 2048,
|
35 |
"tokenizer_padding_side": "right",
|
36 |
"torch_dtype": "bfloat16",
|
37 |
+
"transformers_version": "4.39.1",
|
38 |
"tune_mm_mlp_adapter": false,
|
39 |
"use_cache": true,
|
40 |
"use_mm_proj": true,
|
requirements.txt
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
transformers>=4.39.1
|
2 |
+
flash_attn
|