adarshxs not-lain commited on
Commit
74117a8
1 Parent(s): 3bc57fb

fix `_name_or_path` in config.json (#3)

Browse files

- update config.json (0a86922a8af6074096d8b9439fd6458c27b9e71c)
- add requiements (eee24f04b87735f3deaf4df649f1dd39fe176585)


Co-authored-by: LAin <not-lain@users.noreply.huggingface.co>

Files changed (3) hide show
  1. README.md +6 -0
  2. config.json +2 -2
  3. requirements.txt +2 -0
README.md CHANGED
@@ -36,6 +36,12 @@ The training setup was `4xA100's 80GB` and took ~6 hours to pretrain and ~13 hou
36
  ---
37
 
38
 
 
 
 
 
 
 
39
  ## Training:
40
  Training code Released !!!
41
  https://github.com/tensoic/Cerule
 
36
  ---
37
 
38
 
39
+ ## Installing requirements
40
+ ```
41
+ pip install -qr https://huggingface.co/Tensoic/Cerule-v0.1/resolve/main/requirements.txt
42
+ ```
43
+
44
+
45
  ## Training:
46
  Training code Released !!!
47
  https://github.com/tensoic/Cerule
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Tensoic/Cerule",
3
  "architectures": [
4
  "CeruleGemmaForCausalLM"
5
  ],
@@ -34,7 +34,7 @@
34
  "tokenizer_model_max_length": 2048,
35
  "tokenizer_padding_side": "right",
36
  "torch_dtype": "bfloat16",
37
- "transformers_version": "4.39.0.dev0",
38
  "tune_mm_mlp_adapter": false,
39
  "use_cache": true,
40
  "use_mm_proj": true,
 
1
  {
2
+ "_name_or_path": "Tensoic/Cerule-v0.1",
3
  "architectures": [
4
  "CeruleGemmaForCausalLM"
5
  ],
 
34
  "tokenizer_model_max_length": 2048,
35
  "tokenizer_padding_side": "right",
36
  "torch_dtype": "bfloat16",
37
+ "transformers_version": "4.39.1",
38
  "tune_mm_mlp_adapter": false,
39
  "use_cache": true,
40
  "use_mm_proj": true,
requirements.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ transformers>=4.39.1
2
+ flash_attn