Upload folder using huggingface_hub

#3
Files changed (4) hide show
  1. README.md +2 -2
  2. config.json +15 -6
  3. model.safetensors +3 -0
  4. model.safetensors.index.json +0 -0
README.md CHANGED
@@ -7,15 +7,15 @@ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review
7
  agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
8
  Face and click below. Requests are processed immediately.
9
  extra_gated_button_content: Acknowledge license
10
- base_model: google/gemma-2-2b
11
  tags:
12
  - mlx
 
13
  ---
14
 
15
  # sigtakahashi/google-de-gozari
16
 
17
  The Model [sigtakahashi/google-de-gozari](https://huggingface.co/sigtakahashi/google-de-gozari) was
18
- converted to MLX format from [google/gemma-2-2b](https://huggingface.co/google/gemma-2-2b)
19
  using mlx-lm version **0.20.0**.
20
 
21
  ## Use with mlx
 
7
  agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
8
  Face and click below. Requests are processed immediately.
9
  extra_gated_button_content: Acknowledge license
 
10
  tags:
11
  - mlx
12
+ base_model: google/gemma-2-9b
13
  ---
14
 
15
  # sigtakahashi/google-de-gozari
16
 
17
  The Model [sigtakahashi/google-de-gozari](https://huggingface.co/sigtakahashi/google-de-gozari) was
18
+ converted to MLX format from [google/gemma-2-9b](https://huggingface.co/google/gemma-2-9b)
19
  using mlx-lm version **0.20.0**.
20
 
21
  ## Use with mlx
config.json CHANGED
@@ -12,21 +12,30 @@
12
  "head_dim": 256,
13
  "hidden_act": "gelu_pytorch_tanh",
14
  "hidden_activation": "gelu_pytorch_tanh",
15
- "hidden_size": 2304,
16
  "initializer_range": 0.02,
17
- "intermediate_size": 9216,
18
  "max_position_embeddings": 8192,
19
  "model_type": "gemma2",
20
- "num_attention_heads": 8,
21
- "num_hidden_layers": 26,
22
- "num_key_value_heads": 4,
23
  "pad_token_id": 0,
 
 
 
 
 
 
 
 
24
  "query_pre_attn_scalar": 256,
25
  "rms_norm_eps": 1e-06,
26
  "rope_theta": 10000.0,
27
  "sliding_window": 4096,
 
28
  "torch_dtype": "float32",
29
- "transformers_version": "4.42.4",
30
  "use_cache": true,
31
  "vocab_size": 256000
32
  }
 
12
  "head_dim": 256,
13
  "hidden_act": "gelu_pytorch_tanh",
14
  "hidden_activation": "gelu_pytorch_tanh",
15
+ "hidden_size": 3584,
16
  "initializer_range": 0.02,
17
+ "intermediate_size": 14336,
18
  "max_position_embeddings": 8192,
19
  "model_type": "gemma2",
20
+ "num_attention_heads": 16,
21
+ "num_hidden_layers": 42,
22
+ "num_key_value_heads": 8,
23
  "pad_token_id": 0,
24
+ "quantization": {
25
+ "group_size": 64,
26
+ "bits": 4
27
+ },
28
+ "quantization_config": {
29
+ "group_size": 64,
30
+ "bits": 4
31
+ },
32
  "query_pre_attn_scalar": 256,
33
  "rms_norm_eps": 1e-06,
34
  "rope_theta": 10000.0,
35
  "sliding_window": 4096,
36
+ "sliding_window_size": 4096,
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.42.0.dev0",
39
  "use_cache": true,
40
  "vocab_size": 256000
41
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40a7e4142b05d9a2fa92b3f10ed4e2004b246da5fc0749a35f8dfbe5e6e44a29
3
+ size 5199450578
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff