Ruqiya commited on
Commit
db2961e
·
verified ·
1 Parent(s): 7b5c3bb

Save model using Keras.

Browse files
Files changed (4) hide show
  1. README.md +2 -8
  2. config.json +1 -1
  3. metadata.json +1 -1
  4. model.weights.h5 +1 -1
README.md CHANGED
@@ -1,10 +1,6 @@
 
1
  ---
2
  library_name: keras
3
- datasets:
4
- - arbml/CIDAR
5
- language:
6
- - ar
7
- - en
8
  ---
9
 
10
  This model has been uploaded using the Keras library and can be used with JAX,
@@ -16,6 +12,4 @@ See [Model Cards documentation](https://huggingface.co/docs/hub/model-cards) for
16
  more information.
17
 
18
  For more details about the model architecture, check out
19
- [config.json](./config.json).
20
-
21
- [Notebook](https://www.kaggle.com/code/ruqiyas/arabic-fine-tune-gemma-models-in-keras-using-lora)
 
1
+
2
  ---
3
  library_name: keras
 
 
 
 
 
4
  ---
5
 
6
  This model has been uploaded using the Keras library and can be used with JAX,
 
12
  more information.
13
 
14
  For more details about the model architecture, check out
15
+ [config.json](./config.json).
 
 
config.json CHANGED
@@ -1 +1 @@
1
- {"module": "keras_nlp.src.models.gemma.gemma_causal_lm", "class_name": "GemmaCausalLM", "config": {"backbone": {"module": "keras_nlp.src.models.gemma.gemma_backbone", "class_name": "GemmaBackbone", "config": {"name": "gemma_backbone", "trainable": true, "vocabulary_size": 256000, "num_layers": 42, "num_query_heads": 16, "num_key_value_heads": 8, "hidden_dim": 3584, "intermediate_dim": 28672, "head_dim": 256, "layer_norm_epsilon": 1e-06, "dropout": 0, "query_head_dim_normalize": true, "use_post_ffw_norm": true, "use_post_attention_norm": true, "final_logit_soft_cap": 30, "attention_logit_soft_cap": 50, "sliding_window_size": 4096, "use_sliding_window_attention": true}, "registered_name": "keras_nlp>GemmaBackbone", "build_config": {"input_shape": null}}, "preprocessor": {"module": "keras_nlp.src.models.gemma.gemma_causal_lm_preprocessor", "class_name": "GemmaCausalLMPreprocessor", "config": {"name": "gemma_causal_lm_preprocessor", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "float32"}, "registered_name": null}, "tokenizer": {"module": "keras_nlp.src.models.gemma.gemma_tokenizer", "class_name": "GemmaTokenizer", "config": {"name": "gemma_tokenizer", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "int32"}, "registered_name": null}, "proto": null, "sequence_length": null}, "registered_name": "keras_nlp>GemmaTokenizer"}, "sequence_length": 512, "add_start_token": true, "add_end_token": true}, "registered_name": "keras_nlp>GemmaCausalLMPreprocessor"}, "name": "gemma_causal_lm"}, "registered_name": "keras_nlp>GemmaCausalLM", "build_config": {"input_shape": null}, "compile_config": {"optimizer": {"module": "keras.optimizers", "class_name": "Adam", "config": {"name": "adam", "learning_rate": 4.999999873689376e-05, "weight_decay": null, "clipnorm": null, "global_clipnorm": null, "clipvalue": null, "use_ema": false, "ema_momentum": 0.99, "ema_overwrite_frequency": null, "loss_scale_factor": null, "gradient_accumulation_steps": null, "beta_1": 0.9, "beta_2": 0.999, "epsilon": 1e-07, "amsgrad": false}, "registered_name": null}, "loss": {"module": "keras.losses", "class_name": "SparseCategoricalCrossentropy", "config": {"name": "sparse_categorical_crossentropy", "reduction": "sum_over_batch_size", "from_logits": true, "ignore_class": null}, "registered_name": null}, "loss_weights": null, "metrics": null, "weighted_metrics": [{"module": "keras.metrics", "class_name": "SparseCategoricalAccuracy", "config": {"name": "sparse_categorical_accuracy", "dtype": "float32"}, "registered_name": null}], "run_eagerly": false, "steps_per_execution": 1, "jit_compile": true}}
 
1
+ {"module": "keras_nlp.src.models.gemma.gemma_causal_lm", "class_name": "GemmaCausalLM", "config": {"backbone": {"module": "keras_nlp.src.models.gemma.gemma_backbone", "class_name": "GemmaBackbone", "config": {"name": "gemma_backbone", "trainable": true, "vocabulary_size": 256000, "num_layers": 42, "num_query_heads": 16, "num_key_value_heads": 8, "hidden_dim": 3584, "intermediate_dim": 28672, "head_dim": 256, "layer_norm_epsilon": 1e-06, "dropout": 0, "query_head_dim_normalize": true, "use_post_ffw_norm": true, "use_post_attention_norm": true, "final_logit_soft_cap": 30, "attention_logit_soft_cap": 50, "sliding_window_size": 4096, "use_sliding_window_attention": true}, "registered_name": "keras_nlp>GemmaBackbone", "build_config": {"input_shape": null}}, "preprocessor": {"module": "keras_nlp.src.models.gemma.gemma_causal_lm_preprocessor", "class_name": "GemmaCausalLMPreprocessor", "config": {"name": "gemma_causal_lm_preprocessor", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "float32"}, "registered_name": null}, "tokenizer": {"module": "keras_nlp.src.models.gemma.gemma_tokenizer", "class_name": "GemmaTokenizer", "config": {"name": "gemma_tokenizer", "trainable": true, "dtype": {"module": "keras", "class_name": "DTypePolicy", "config": {"name": "int32"}, "registered_name": null}, "proto": null, "sequence_length": null, "add_bos": false, "add_eos": false}, "registered_name": "keras_nlp>GemmaTokenizer"}, "sequence_length": 512, "add_start_token": true, "add_end_token": true}, "registered_name": "keras_nlp>GemmaCausalLMPreprocessor"}, "name": "gemma_causal_lm"}, "registered_name": "keras_nlp>GemmaCausalLM", "build_config": {"input_shape": null}, "compile_config": {"optimizer": {"module": "keras.optimizers", "class_name": "Adam", "config": {"name": "adam", "learning_rate": 4.999999873689376e-05, "weight_decay": null, "clipnorm": null, "global_clipnorm": null, "clipvalue": null, "use_ema": false, "ema_momentum": 0.99, "ema_overwrite_frequency": null, "loss_scale_factor": null, "gradient_accumulation_steps": null, "beta_1": 0.9, "beta_2": 0.999, "epsilon": 1e-07, "amsgrad": false}, "registered_name": null}, "loss": {"module": "keras.losses", "class_name": "SparseCategoricalCrossentropy", "config": {"name": "sparse_categorical_crossentropy", "reduction": "sum_over_batch_size", "from_logits": true, "ignore_class": null}, "registered_name": null}, "loss_weights": null, "metrics": null, "weighted_metrics": [{"module": "keras.metrics", "class_name": "SparseCategoricalAccuracy", "config": {"name": "sparse_categorical_accuracy", "dtype": "float32"}, "registered_name": null}], "run_eagerly": false, "steps_per_execution": 1, "jit_compile": true}}
metadata.json CHANGED
@@ -1 +1 @@
1
- {"keras_version": "3.5.0", "date_saved": "2024-08-28@07:47:23"}
 
1
+ {"keras_version": "3.5.0", "date_saved": "2024-09-14@07:15:55"}
model.weights.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24bbfe29f3a8026ad1b5b9783f28b6c43c7433f99941b6051dfafd31fd5b87a6
3
  size 37201030616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8de96d94565dee377dae0e487719bde8858475bd1fe8a927f67716a2e08dc2cd
3
  size 37201030616