Save model using Keras.
Browse files- README.md +2 -8
- config.json +1 -1
- metadata.json +1 -1
- model.weights.h5 +1 -1
README.md
CHANGED
@@ -1,10 +1,6 @@
|
|
|
|
1 |
---
|
2 |
library_name: keras
|
3 |
-
datasets:
|
4 |
-
- arbml/CIDAR
|
5 |
-
language:
|
6 |
-
- ar
|
7 |
-
- en
|
8 |
---
|
9 |
|
10 |
This model has been uploaded using the Keras library and can be used with JAX,
|
@@ -16,6 +12,4 @@ See [Model Cards documentation](https://huggingface.co/docs/hub/model-cards) for
|
|
16 |
more information.
|
17 |
|
18 |
For more details about the model architecture, check out
|
19 |
-
[config.json](./config.json).
|
20 |
-
|
21 |
-
[Notebook](https://www.kaggle.com/code/ruqiyas/arabic-fine-tune-gemma-models-in-keras-using-lora)
|
|
|
1 |
+
|
2 |
---
|
3 |
library_name: keras
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
This model has been uploaded using the Keras library and can be used with JAX,
|
|
|
12 |
more information.
|
13 |
|
14 |
For more details about the model architecture, check out
|
15 |
+
[config.json](./config.json).
|
|
|
|
config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"module": "keras_nlp.src.models.gemma.gemma_causal_lm", "class_name": "GemmaCausalLM", "config": {"backbone": {"module": "keras_nlp.src.models.gemma.gemma_backbone", "class_name": "GemmaBackbone", "config": {"name": "gemma_backbone", "trainable": true, "vocabulary_size": 256000, "num_layers": 42, "num_query_heads": 16, "num_key_value_heads": 8, "hidden_dim": 3584, "intermediate_dim": 28672, "head_dim": 256, "layer_norm_epsilon": 1e-06, "dropout": 0, "query_head_dim_normalize": true, "use_post_ffw_norm": true, "use_post_attention_norm": true, "final_logit_soft_cap": 30, "attention_logit_soft_cap": 50, "sliding_window_size": 4096, "use_sliding_window_attention": true}, "registered_name": "keras_nlp>GemmaBackbone", "build_config": {"input_shape": null}}, "preprocessor": {"module": "keras_nlp.src.models.gemma.gemma_causal_lm_preprocessor", "class_name": "GemmaCausalLMPreprocessor", "config": {"name": "gemma_causal_lm_preprocessor", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "float32"}, "registered_name": null}, "tokenizer": {"module": "keras_nlp.src.models.gemma.gemma_tokenizer", "class_name": "GemmaTokenizer", "config": {"name": "gemma_tokenizer", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "int32"}, "registered_name": null}, "proto": null, "sequence_length": null}, "registered_name": "keras_nlp>GemmaTokenizer"}, "sequence_length": 512, "add_start_token": true, "add_end_token": true}, "registered_name": "keras_nlp>GemmaCausalLMPreprocessor"}, "name": "gemma_causal_lm"}, "registered_name": "keras_nlp>GemmaCausalLM", "build_config": {"input_shape": null}, "compile_config": {"optimizer": {"module": "keras.optimizers", "class_name": "Adam", "config": {"name": "adam", "learning_rate": 4.999999873689376e-05, "weight_decay": null, "clipnorm": null, "global_clipnorm": null, "clipvalue": null, "use_ema": false, "ema_momentum": 0.99, "ema_overwrite_frequency": null, "loss_scale_factor": null, "gradient_accumulation_steps": null, "beta_1": 0.9, "beta_2": 0.999, "epsilon": 1e-07, "amsgrad": false}, "registered_name": null}, "loss": {"module": "keras.losses", "class_name": "SparseCategoricalCrossentropy", "config": {"name": "sparse_categorical_crossentropy", "reduction": "sum_over_batch_size", "from_logits": true, "ignore_class": null}, "registered_name": null}, "loss_weights": null, "metrics": null, "weighted_metrics": [{"module": "keras.metrics", "class_name": "SparseCategoricalAccuracy", "config": {"name": "sparse_categorical_accuracy", "dtype": "float32"}, "registered_name": null}], "run_eagerly": false, "steps_per_execution": 1, "jit_compile": true}}
|
|
|
1 |
+
{"module": "keras_nlp.src.models.gemma.gemma_causal_lm", "class_name": "GemmaCausalLM", "config": {"backbone": {"module": "keras_nlp.src.models.gemma.gemma_backbone", "class_name": "GemmaBackbone", "config": {"name": "gemma_backbone", "trainable": true, "vocabulary_size": 256000, "num_layers": 42, "num_query_heads": 16, "num_key_value_heads": 8, "hidden_dim": 3584, "intermediate_dim": 28672, "head_dim": 256, "layer_norm_epsilon": 1e-06, "dropout": 0, "query_head_dim_normalize": true, "use_post_ffw_norm": true, "use_post_attention_norm": true, "final_logit_soft_cap": 30, "attention_logit_soft_cap": 50, "sliding_window_size": 4096, "use_sliding_window_attention": true}, "registered_name": "keras_nlp>GemmaBackbone", "build_config": {"input_shape": null}}, "preprocessor": {"module": "keras_nlp.src.models.gemma.gemma_causal_lm_preprocessor", "class_name": "GemmaCausalLMPreprocessor", "config": {"name": "gemma_causal_lm_preprocessor", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "float32"}, "registered_name": null}, "tokenizer": {"module": "keras_nlp.src.models.gemma.gemma_tokenizer", "class_name": "GemmaTokenizer", "config": {"name": "gemma_tokenizer", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "int32"}, "registered_name": null}, "proto": null, "sequence_length": null, "add_bos": false, "add_eos": false}, "registered_name": "keras_nlp>GemmaTokenizer"}, "sequence_length": 512, "add_start_token": true, "add_end_token": true}, "registered_name": "keras_nlp>GemmaCausalLMPreprocessor"}, "name": "gemma_causal_lm"}, "registered_name": "keras_nlp>GemmaCausalLM", "build_config": {"input_shape": null}, "compile_config": {"optimizer": {"module": "keras.optimizers", "class_name": "Adam", "config": {"name": "adam", "learning_rate": 4.999999873689376e-05, "weight_decay": null, "clipnorm": null, "global_clipnorm": null, "clipvalue": null, "use_ema": false, "ema_momentum": 0.99, "ema_overwrite_frequency": null, "loss_scale_factor": null, "gradient_accumulation_steps": null, "beta_1": 0.9, "beta_2": 0.999, "epsilon": 1e-07, "amsgrad": false}, "registered_name": null}, "loss": {"module": "keras.losses", "class_name": "SparseCategoricalCrossentropy", "config": {"name": "sparse_categorical_crossentropy", "reduction": "sum_over_batch_size", "from_logits": true, "ignore_class": null}, "registered_name": null}, "loss_weights": null, "metrics": null, "weighted_metrics": [{"module": "keras.metrics", "class_name": "SparseCategoricalAccuracy", "config": {"name": "sparse_categorical_accuracy", "dtype": "float32"}, "registered_name": null}], "run_eagerly": false, "steps_per_execution": 1, "jit_compile": true}}
|
metadata.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"keras_version": "3.5.0", "date_saved": "2024-
|
|
|
1 |
+
{"keras_version": "3.5.0", "date_saved": "2024-09-14@07:15:55"}
|
model.weights.h5
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 37201030616
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8de96d94565dee377dae0e487719bde8858475bd1fe8a927f67716a2e08dc2cd
|
3 |
size 37201030616
|