automerger commited on
Commit
bd74026
1 Parent(s): 8bdbbbf

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -18,19 +18,19 @@ Experiment26Inex12-7B is an automated merge created by [Maxime Labonne](https://
18
 
19
  ```yaml
20
  models:
21
- - model: yam-peleg/Experiment26-7B
22
  # No parameters necessary for base model
23
  - model: MSL7/INEX12-7b
24
  parameters:
25
  density: 0.53
26
  weight: 0.6
27
  merge_method: dare_ties
28
- base_model: yam-peleg/Experiment26-7B
29
  parameters:
30
  int8_mask: true
31
  dtype: bfloat16
32
  random_seed: 0
33
- ```
34
 
35
  ## 💻 Usage
36
 
 
18
 
19
  ```yaml
20
  models:
21
+ - model: rwitz/experiment26-truthy-iter-1
22
  # No parameters necessary for base model
23
  - model: MSL7/INEX12-7b
24
  parameters:
25
  density: 0.53
26
  weight: 0.6
27
  merge_method: dare_ties
28
+ base_model: rwitz/experiment26-truthy-iter-1
29
  parameters:
30
  int8_mask: true
31
  dtype: bfloat16
32
  random_seed: 0
33
+ ```
34
 
35
  ## 💻 Usage
36
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "yam-peleg/Experiment26-7B",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -20,7 +20,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.38.2",
24
  "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "rwitz/experiment26-truthy-iter-1",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.0",
24
  "use_cache": true,
25
  "vocab_size": 32000
26
  }
mergekit_config.yml CHANGED
@@ -1,15 +1,14 @@
1
 
2
  models:
3
- - model: yam-peleg/Experiment26-7B
4
  # No parameters necessary for base model
5
  - model: MSL7/INEX12-7b
6
  parameters:
7
  density: 0.53
8
  weight: 0.6
9
  merge_method: dare_ties
10
- base_model: yam-peleg/Experiment26-7B
11
  parameters:
12
  int8_mask: true
13
  dtype: bfloat16
14
  random_seed: 0
15
-
 
1
 
2
  models:
3
+ - model: rwitz/experiment26-truthy-iter-1
4
  # No parameters necessary for base model
5
  - model: MSL7/INEX12-7b
6
  parameters:
7
  density: 0.53
8
  weight: 0.6
9
  merge_method: dare_ties
10
+ base_model: rwitz/experiment26-truthy-iter-1
11
  parameters:
12
  int8_mask: true
13
  dtype: bfloat16
14
  random_seed: 0
 
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55c262ff1dbbfbd141a3574fcc51e2c03c060d3d0d990c12752ce721a4bd9af6
3
  size 9825524456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f68dc3060c7cabca4644a7ac795fa9fddf95f3a877b14846923544ca2ae8bb3c
3
  size 9825524456
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7efe26a40ef07a2c6828a208f328c910ab2713e28974502d77198509828512c8
3
  size 4657973592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b236e3118868003af70ac9a19ed4420c974899f5a31f455cd1d21b47bb1a62eb
3
  size 4657973592