File size: 823 Bytes
f8bf680
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
{
  "backbone_checkpoint_name": "t5-large",
  "backbone_class": "T5ForConditionalGeneration",
  "backbone_hash": "6297bd1acc36524547c8a76cc03fef5c",
  "bottleneck_dim": null,
  "common_structure": null,
  "delta_type": "compacter",
  "factorized_phm": true,
  "factorized_phm_rule": false,
  "hypercomplex_division": 4,
  "hypercomplex_nonlinearity": "glorot-uniform",
  "kronecker_prod": null,
  "learn_phm": true,
  "modified_modules": [
    "SelfAttention",
    "DenseReluDense"
  ],
  "non_linearity": "gelu_new",
  "opendelta_version": "0.0.1",
  "phm_c_init": "normal",
  "phm_init_range": 0.0001,
  "phm_rank": 1,
  "reduction_factor": 16,
  "sequential": null,
  "shared_W_phm": false,
  "shared_phm_rule": false,
  "transformers_version": "4.17.0",
  "use_bias_down_sampler": true,
  "use_bias_up_sampler": true
}