Yestin-Wang commited on
Commit
f5288d3
·
1 Parent(s): 50a85f3

Upload folder using huggingface_hub

Browse files
assets.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"class_name": "BaseLearner", "column_types": {"image_path": "image_path", "text": "text", "label": "categorical"}, "label_column": "label", "problem_type": "binary", "presets": null, "eval_metric_name": "accuracy", "validation_metric_name": "accuracy", "minmax_mode": "max", "output_shape": 2, "save_path": "/content/AutogluonModels/ag-20240108_103835", "pretrained": true, "pretrained_path": null, "fit_called": true, "best_score": null, "total_train_time": null, "version": "1.0.0"}
config.yaml ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ names:
3
+ - fusion_mlp
4
+ - hf_text
5
+ - timm_image
6
+ hf_text:
7
+ checkpoint_name: local://hf_text
8
+ gradient_checkpointing: false
9
+ pooling_mode: cls
10
+ data_types:
11
+ - text
12
+ tokenizer_name: hf_auto
13
+ use_fast: true
14
+ max_text_len: 512
15
+ insert_sep: true
16
+ low_cpu_mem_usage: false
17
+ text_segment_num: 2
18
+ stochastic_chunk: false
19
+ text_aug_detect_length: 10
20
+ text_trivial_aug_maxscale: 0.0
21
+ text_train_augment_types: null
22
+ timm_image:
23
+ checkpoint_name: timm/swinv2_base_window8_256.ms_in1k
24
+ mix_choice: all_logits
25
+ data_types:
26
+ - image
27
+ train_transforms:
28
+ - resize_shorter_side
29
+ - center_crop
30
+ - trivial_augment
31
+ val_transforms:
32
+ - resize_shorter_side
33
+ - center_crop
34
+ image_norm: imagenet
35
+ image_size: null
36
+ max_img_num_per_col: 2
37
+ fusion_mlp:
38
+ weight: 0.1
39
+ adapt_in_features: max
40
+ hidden_sizes:
41
+ - 128
42
+ activation: leaky_relu
43
+ drop_rate: 0.1
44
+ normalization: layer_norm
45
+ data_types: null
46
+ data:
47
+ image:
48
+ missing_value_strategy: zero
49
+ text:
50
+ normalize_text: false
51
+ categorical:
52
+ minimum_cat_count: 100
53
+ maximum_num_cat: 20
54
+ convert_to_text: true
55
+ numerical:
56
+ convert_to_text: false
57
+ scaler_with_mean: true
58
+ scaler_with_std: true
59
+ document:
60
+ missing_value_strategy: zero
61
+ label:
62
+ numerical_label_preprocessing: standardscaler
63
+ pos_label: null
64
+ column_features_pooling_mode: concat
65
+ mixup:
66
+ turn_on: false
67
+ mixup_alpha: 0.8
68
+ cutmix_alpha: 1.0
69
+ cutmix_minmax: null
70
+ prob: 1.0
71
+ switch_prob: 0.5
72
+ mode: batch
73
+ turn_off_epoch: 5
74
+ label_smoothing: 0.1
75
+ templates:
76
+ turn_on: false
77
+ num_templates: 30
78
+ template_length: 2048
79
+ preset_templates:
80
+ - super_glue
81
+ - rte
82
+ custom_templates: null
83
+ optimization:
84
+ optim_type: adamw
85
+ learning_rate: 0.0001
86
+ weight_decay: 0.001
87
+ lr_choice: layerwise_decay
88
+ lr_decay: 0.9
89
+ lr_schedule: cosine_decay
90
+ max_epochs: 10
91
+ max_steps: -1
92
+ warmup_steps: 0.1
93
+ end_lr: 0
94
+ lr_mult: 1
95
+ patience: 10
96
+ val_check_interval: 0.5
97
+ check_val_every_n_epoch: 1
98
+ skip_final_val: false
99
+ gradient_clip_val: 1
100
+ gradient_clip_algorithm: norm
101
+ track_grad_norm: -1
102
+ log_every_n_steps: 10
103
+ top_k: 3
104
+ top_k_average_method: greedy_soup
105
+ efficient_finetune: null
106
+ lora:
107
+ module_filter: null
108
+ filter:
109
+ - query
110
+ - value
111
+ - ^q$
112
+ - ^v$
113
+ - ^k$
114
+ - ^o$
115
+ r: 8
116
+ alpha: 8
117
+ loss_function: auto
118
+ focal_loss:
119
+ alpha: null
120
+ gamma: 2.0
121
+ reduction: mean
122
+ mask2former_loss:
123
+ loss_cross_entropy_weight: 10.0
124
+ loss_mask_weight: 5.0
125
+ loss_dice_weight: 5.0
126
+ extra_trainable_params: []
127
+ env:
128
+ num_gpus: 1
129
+ num_nodes: 1
130
+ batch_size: 128
131
+ per_gpu_batch_size: 8
132
+ eval_batch_size_ratio: 4
133
+ per_gpu_batch_size_evaluation: null
134
+ precision: 16-mixed
135
+ num_workers: 2
136
+ num_workers_evaluation: 2
137
+ accelerator: auto
138
+ fast_dev_run: false
139
+ deterministic: false
140
+ auto_select_gpus: true
141
+ strategy: auto
142
+ deepspeed_allgather_size: 1000000000.0
143
+ deepspeed_allreduce_size: 1000000000.0
144
+ compile:
145
+ turn_on: false
146
+ mode: default
147
+ dynamic: true
148
+ backend: inductor
data_processors.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:043bcc7958ac9e1fe6eb535f2c0fdeda746066f01a84d14b8f7e81f95e2c42fc
3
+ size 15421
df_preprocessor.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9c5ef7dacf348ac78db5ae5a939c7993375c9806559a9d490b32b6a5ba8a96
3
+ size 14093
events.out.tfevents.1704710458.acc5ce4afb75.1271.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:291ef20f12da6eb0e11098b1c6046e4c4c33e8f7a50453e0b5d927c3cf2d3759
3
+ size 857275
hf_text/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cardiffnlp/twitter-roberta-large-hate-latest",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "id2label": {
14
+ "0": "hate_gender",
15
+ "1": "hate_race",
16
+ "2": "hate_sexuality",
17
+ "3": "hate_religion",
18
+ "4": "hate_origin",
19
+ "5": "hate_disability",
20
+ "6": "hate_age",
21
+ "7": "not_hate"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 4096,
25
+ "label2id": {
26
+ "hate_age": 6,
27
+ "hate_disability": 5,
28
+ "hate_gender": 0,
29
+ "hate_origin": 4,
30
+ "hate_race": 1,
31
+ "hate_religion": 3,
32
+ "hate_sexuality": 2,
33
+ "not_hate": 7
34
+ },
35
+ "layer_norm_eps": 1e-05,
36
+ "max_position_embeddings": 514,
37
+ "model_type": "roberta",
38
+ "num_attention_heads": 16,
39
+ "num_hidden_layers": 24,
40
+ "pad_token_id": 1,
41
+ "position_embedding_type": "absolute",
42
+ "problem_type": "single_label_classification",
43
+ "torch_dtype": "float32",
44
+ "transformers_version": "4.31.0",
45
+ "type_vocab_size": 1,
46
+ "use_cache": true,
47
+ "vocab_size": 50265
48
+ }
hf_text/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
hf_text/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
hf_text/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
hf_text/tokenizer_config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "mask_token": "<mask>",
51
+ "max_length": 64,
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "stride": 0,
56
+ "tokenizer_class": "RobertaTokenizer",
57
+ "trim_offsets": true,
58
+ "truncation_side": "right",
59
+ "truncation_strategy": "longest_first",
60
+ "unk_token": "<unk>"
61
+ }
hf_text/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
hparams.yaml ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ optim_type: adamw
2
+ lr_choice: layerwise_decay
3
+ lr_schedule: cosine_decay
4
+ lr: 0.0001
5
+ lr_decay: 0.9
6
+ end_lr: 0
7
+ lr_mult: 1
8
+ weight_decay: 0.001
9
+ warmup_steps: 0.1
10
+ validation_metric_name: accuracy
11
+ custom_metric_func: null
12
+ efficient_finetune: null
13
+ mixup_off_epoch: 5
14
+ skip_final_val: false
15
+ track_grad_norm: -1
model.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a69917c07f5bfccdbdc920beb15c3ffc70861b233897eb7b7282b6ec7e2929b
3
+ size 1778714829