bastasie commited on
Commit
06da962
1 Parent(s): e3081e3

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +11 -11
config.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
  "model_type": "CustomModel",
3
- "architecture": "CustomModel",
4
- "input_size": 512,
5
  "hidden_size": 128,
6
- "output_size": 768,
7
- "vocab_size": 30522, // Example vocab size, adjust according to your tokenizer's vocabulary
8
- "layer_norm_epsilon": 1e-12,
9
- "hidden_dropout_prob": 0.1,
10
- "num_attention_heads": 12, // Adjust if your model uses attention mechanisms
11
- "attention_probs_dropout_prob": 0.1,
12
- "intermediate_size": 3072, // Example size, adjust based on your model's architecture
13
- "num_hidden_layers": 12, // Adjust based on your model's depth
14
- "initializer_range": 0.02
 
15
  }
 
1
  {
2
  "model_type": "CustomModel",
3
+ "input_size": 768,
 
4
  "hidden_size": 128,
5
+ "output_size": 2,
6
+ "a": 0.5,
7
+ "epsilon": 0.1,
8
+ "init_weights_range": "-1 to 1",
9
+ "architecture": "Single layer linear transformation followed by Ba-inspired activation and another linear transformation.",
10
+ "activation_function": "Ba-inspired custom activation",
11
+ "num_labels": 2,
12
+ "problem_type": "Classification",
13
+ "vocab_size": 30522, // Assuming you're using a tokenizer similar to BERT's; adjust accordingly.
14
+ "_comment": "This is a custom model configuration. Some fields are specific to this custom implementation and may require custom handling code."
15
  }