iblfe commited on
Commit
e60d8d2
1 Parent(s): 5d73f5b

Upload 12 files

Browse files
assets.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"class_name": "MultiModalPredictor", "column_types": {"image_path": "categorical", "xmax": "numerical", "xmin": "numerical", "ymax": "numerical", "ymin": "numerical", "image": "image_path", "name": "categorical"}, "label_column": "name", "problem_type": "multiclass", "presets": null, "eval_metric_name": "accuracy", "validation_metric_name": "accuracy", "output_shape": 10, "classes": null, "save_path": "C:\\Users\\iblfe\\Downloads\\dataset\\dataset\\AutogluonModels\\ag-20231031_101638", "pretrained": true, "pretrained_path": null, "fit_called": true, "best_score": null, "total_train_time": null, "version": "0.8.2"}
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/electra-base-discriminator",
3
+ "architectures": [
4
+ "ElectraForPreTraining"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "embedding_size": 768,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "electra",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "summary_activation": "gelu",
22
+ "summary_last_dropout": 0.1,
23
+ "summary_type": "first",
24
+ "summary_use_proj": true,
25
+ "transformers_version": "4.26.1",
26
+ "type_vocab_size": 2,
27
+ "use_cache": true,
28
+ "vocab_size": 30522
29
+ }
config.yaml ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ names:
3
+ - fusion_mlp
4
+ - hf_text
5
+ - numerical_mlp
6
+ - timm_image
7
+ numerical_mlp:
8
+ hidden_size: 128
9
+ activation: leaky_relu
10
+ num_layers: 1
11
+ drop_rate: 0.1
12
+ normalization: layer_norm
13
+ d_token: 8
14
+ embedding_arch: null
15
+ data_types:
16
+ - numerical
17
+ merge: concat
18
+ hf_text:
19
+ checkpoint_name: local://hf_text
20
+ gradient_checkpointing: false
21
+ pooling_mode: cls
22
+ data_types:
23
+ - text
24
+ tokenizer_name: hf_auto
25
+ max_text_len: 512
26
+ insert_sep: true
27
+ low_cpu_mem_usage: false
28
+ text_segment_num: 2
29
+ stochastic_chunk: false
30
+ text_aug_detect_length: 10
31
+ text_trivial_aug_maxscale: 0.0
32
+ text_train_augment_types: null
33
+ timm_image:
34
+ checkpoint_name: swin_base_patch4_window7_224
35
+ mix_choice: all_logits
36
+ data_types:
37
+ - image
38
+ train_transforms:
39
+ - resize_shorter_side
40
+ - center_crop
41
+ - trivial_augment
42
+ val_transforms:
43
+ - resize_shorter_side
44
+ - center_crop
45
+ image_norm: imagenet
46
+ image_size: null
47
+ max_img_num_per_col: 2
48
+ fusion_mlp:
49
+ weight: 0.1
50
+ adapt_in_features: max
51
+ hidden_sizes:
52
+ - 128
53
+ activation: leaky_relu
54
+ drop_rate: 0.1
55
+ normalization: layer_norm
56
+ data_types: null
57
+ data:
58
+ image:
59
+ missing_value_strategy: zero
60
+ text:
61
+ normalize_text: false
62
+ categorical:
63
+ minimum_cat_count: 100
64
+ maximum_num_cat: 20
65
+ convert_to_text: true
66
+ numerical:
67
+ convert_to_text: false
68
+ scaler_with_mean: true
69
+ scaler_with_std: true
70
+ document:
71
+ missing_value_strategy: zero
72
+ label:
73
+ numerical_label_preprocessing: standardscaler
74
+ pos_label: null
75
+ mixup:
76
+ turn_on: false
77
+ mixup_alpha: 0.8
78
+ cutmix_alpha: 1.0
79
+ cutmix_minmax: null
80
+ prob: 1.0
81
+ switch_prob: 0.5
82
+ mode: batch
83
+ turn_off_epoch: 5
84
+ label_smoothing: 0.1
85
+ templates:
86
+ turn_on: false
87
+ num_templates: 30
88
+ template_length: 2048
89
+ preset_templates:
90
+ - super_glue
91
+ - rte
92
+ custom_templates: null
93
+ optimization:
94
+ optim_type: adamw
95
+ learning_rate: 0.0001
96
+ weight_decay: 0.001
97
+ lr_choice: layerwise_decay
98
+ lr_decay: 0.9
99
+ lr_schedule: cosine_decay
100
+ max_epochs: 10
101
+ max_steps: -1
102
+ warmup_steps: 0.1
103
+ end_lr: 0
104
+ lr_mult: 1
105
+ patience: 10
106
+ val_check_interval: 0.5
107
+ check_val_every_n_epoch: 1
108
+ skip_final_val: false
109
+ gradient_clip_val: 1
110
+ gradient_clip_algorithm: norm
111
+ track_grad_norm: -1
112
+ log_every_n_steps: 10
113
+ top_k: 3
114
+ top_k_average_method: greedy_soup
115
+ efficient_finetune: null
116
+ lora:
117
+ module_filter: null
118
+ filter:
119
+ - query
120
+ - value
121
+ - ^q$
122
+ - ^v$
123
+ - ^k$
124
+ - ^o$
125
+ r: 8
126
+ alpha: 8
127
+ loss_function: auto
128
+ focal_loss:
129
+ alpha: null
130
+ gamma: 2.0
131
+ reduction: mean
132
+ env:
133
+ num_gpus: 1
134
+ num_nodes: 1
135
+ batch_size: 128
136
+ per_gpu_batch_size: 8
137
+ eval_batch_size_ratio: 4
138
+ per_gpu_batch_size_evaluation: null
139
+ precision: 16
140
+ num_workers: 2
141
+ num_workers_evaluation: 2
142
+ fast_dev_run: false
143
+ deterministic: false
144
+ auto_select_gpus: true
145
+ strategy: null
146
+ deepspeed_allgather_size: 1000000000.0
147
+ deepspeed_allreduce_size: 1000000000.0
data_processors.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d176f6bba693fa5a50fde123e050dd2fa91e57296f738695bf3f30aa1ec4d9a2
3
+ size 15164
df_preprocessor.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec63606a1d92a466866a4aa34fa8ea72b5f6e8a5fd12ab6616e5160ef749d957
3
+ size 17162
events.out.tfevents.1698747405.影のリーダー.35680.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f881a620c6840d8ecb273825dd5455a905936e80126a6f40f9957922415346
3
+ size 3096003
hparams.yaml ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ optim_type: adamw
2
+ lr_choice: layerwise_decay
3
+ lr_schedule: cosine_decay
4
+ lr: 0.0001
5
+ lr_decay: 0.9
6
+ end_lr: 0
7
+ lr_mult: 1
8
+ weight_decay: 0.001
9
+ warmup_steps: 0.1
10
+ validation_metric_name: accuracy
11
+ custom_metric_func: null
12
+ efficient_finetune: null
13
+ trainable_param_names: []
14
+ mixup_fn: null
15
+ mixup_off_epoch: 5
16
+ skip_final_val: false
model.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa2b639b1d4d59750993081047b6f8d2802f35ac91a29e428694b63ec15dbe81
3
+ size 792247793
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "google/electra-base-discriminator",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "ElectraTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff