dev-planningo commited on
Commit
50c5929
1 Parent(s): f5ad318

Upload text_encoder/config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. text_encoder/config.json +53 -0
text_encoder/config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--SG161222--RealVisXL_V4.0_Lightning/snapshots/2acc78212a1f215f839bda46103df087fa634576/text_encoder",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "neuron": {
19
+ "auto_cast": "all",
20
+ "auto_cast_type": "bf16",
21
+ "compiler_type": "neuronx-cc",
22
+ "compiler_version": "2.12.68.0+4480452af",
23
+ "disable_fallback": false,
24
+ "disable_fast_relayout": false,
25
+ "dynamic_batch_size": false,
26
+ "inline_weights_to_neff": true,
27
+ "input_names": [
28
+ "input_ids"
29
+ ],
30
+ "model_type": "clip-text-model",
31
+ "optlevel": "2",
32
+ "output_attentions": false,
33
+ "output_hidden_states": false,
34
+ "output_names": [
35
+ "last_hidden_state",
36
+ "pooler_output",
37
+ "hidden_states"
38
+ ],
39
+ "static_batch_size": 1,
40
+ "static_num_beams": 1,
41
+ "static_sequence_length": 77
42
+ },
43
+ "num_attention_heads": 12,
44
+ "num_hidden_layers": 12,
45
+ "output_hidden_states": true,
46
+ "pad_token_id": 1,
47
+ "projection_dim": 768,
48
+ "task": "feature-extraction",
49
+ "torch_dtype": "float32",
50
+ "torchscript": true,
51
+ "transformers_version": "4.36.2",
52
+ "vocab_size": 49408
53
+ }