Narkantak commited on
Commit
94238d0
1 Parent(s): 49c4948

/kaggle/working/Narkantak/Intent-classify-12K

Browse files
Narkantak/Intent-classify-12K/config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "Install",
14
+ "1": "Compensation",
15
+ "2": "Change",
16
+ "3": "Summarize",
17
+ "4": "Terminate",
18
+ "5": "SBC"
19
+ },
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "Change": 2,
24
+ "Compensation": 1,
25
+ "Install": 0,
26
+ "SBC": 5,
27
+ "Summarize": 3,
28
+ "Terminate": 4
29
+ },
30
+ "layer_norm_eps": 1e-12,
31
+ "max_position_embeddings": 512,
32
+ "model_type": "bert",
33
+ "num_attention_heads": 12,
34
+ "num_hidden_layers": 12,
35
+ "pad_token_id": 0,
36
+ "position_embedding_type": "absolute",
37
+ "problem_type": "single_label_classification",
38
+ "torch_dtype": "float32",
39
+ "transformers_version": "4.38.2",
40
+ "type_vocab_size": 2,
41
+ "use_cache": true,
42
+ "vocab_size": 30522
43
+ }
Narkantak/Intent-classify-12K/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a62c4d9f70c329a6afe6e67264509fbef080878b83aa4afd472515ea7e9dc32
3
+ size 437970952
Narkantak/Intent-classify-12K/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
Narkantak/Intent-classify-12K/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
Narkantak/Intent-classify-12K/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "max_length": 512,
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
Narkantak/Intent-classify-12K/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6938a7f9f012635cb9368fe55a61a7d36c6dc8845a5f1454e394f1f564fb5b4
3
+ size 4856
Narkantak/Intent-classify-12K/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
README.md CHANGED
@@ -1,8 +1,13 @@
1
  ---
2
- library_name: peft
 
3
  tags:
4
  - generated_from_trainer
5
- base_model: Narkantak/Mistral-2x7b-Instruct-1x2
 
 
 
 
6
  model-index:
7
  - name: working
8
  results: []
@@ -13,9 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # working
15
 
16
- This model is a fine-tuned version of [Narkantak/Mistral-2x7b-Instruct-1x2](https://huggingface.co/Narkantak/Mistral-2x7b-Instruct-1x2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.3777
 
 
 
 
19
 
20
  ## Model description
21
 
@@ -34,57 +43,213 @@ More information needed
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
- - learning_rate: 0.0002
38
- - train_batch_size: 6
39
- - eval_batch_size: 6
40
  - seed: 42
41
- - gradient_accumulation_steps: 4
42
- - total_train_batch_size: 24
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - lr_scheduler_warmup_steps: 2
46
- - num_epochs: 30
47
- - mixed_precision_training: Native AMP
48
 
49
  ### Training results
50
 
51
- | Training Loss | Epoch | Step | Validation Loss |
52
- |:-------------:|:-----:|:----:|:---------------:|
53
- | 3.0 | 0.96 | 12 | 0.8713 |
54
- | 0.5738 | 2.0 | 25 | 0.4515 |
55
- | 0.3676 | 2.96 | 37 | 0.3039 |
56
- | 0.2127 | 4.0 | 50 | 0.2680 |
57
- | 0.1891 | 4.96 | 62 | 0.2556 |
58
- | 0.1427 | 6.0 | 75 | 0.2721 |
59
- | 0.132 | 6.96 | 87 | 0.2762 |
60
- | 0.1037 | 8.0 | 100 | 0.2844 |
61
- | 0.1006 | 8.96 | 112 | 0.2951 |
62
- | 0.0832 | 10.0 | 125 | 0.3044 |
63
- | 0.0829 | 10.96 | 137 | 0.3194 |
64
- | 0.073 | 12.0 | 150 | 0.3282 |
65
- | 0.0757 | 12.96 | 162 | 0.3415 |
66
- | 0.0692 | 14.0 | 175 | 0.3230 |
67
- | 0.0738 | 14.96 | 187 | 0.3427 |
68
- | 0.0666 | 16.0 | 200 | 0.3449 |
69
- | 0.0712 | 16.96 | 212 | 0.3450 |
70
- | 0.0652 | 18.0 | 225 | 0.3511 |
71
- | 0.0698 | 18.96 | 237 | 0.3570 |
72
- | 0.0641 | 20.0 | 250 | 0.3604 |
73
- | 0.0691 | 20.96 | 262 | 0.3662 |
74
- | 0.0638 | 22.0 | 275 | 0.3682 |
75
- | 0.0689 | 22.96 | 287 | 0.3644 |
76
- | 0.0632 | 24.0 | 300 | 0.3665 |
77
- | 0.0681 | 24.96 | 312 | 0.3709 |
78
- | 0.0627 | 26.0 | 325 | 0.3725 |
79
- | 0.0677 | 26.96 | 337 | 0.3752 |
80
- | 0.0621 | 28.0 | 350 | 0.3774 |
81
- | 0.0645 | 28.8 | 360 | 0.3777 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82
 
83
 
84
  ### Framework versions
85
 
86
- - PEFT 0.10.0
87
  - Transformers 4.38.2
88
  - Pytorch 2.1.2
89
  - Datasets 2.1.0
90
- - Tokenizers 0.15.2
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: google-bert/bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - f1
9
+ - precision
10
+ - recall
11
  model-index:
12
  - name: working
13
  results: []
 
18
 
19
  # working
20
 
21
+ This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.0013
24
+ - Accuracy: 0.9997
25
+ - F1: 0.9997
26
+ - Precision: 0.9997
27
+ - Recall: 0.9997
28
 
29
  ## Model description
30
 
 
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
+ - learning_rate: 5e-05
47
+ - train_batch_size: 32
48
+ - eval_batch_size: 32
49
  - seed: 42
 
 
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 10
53
+ - num_epochs: 10
 
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
59
+ | 1.7386 | 0.05 | 10 | 1.4921 | 0.5744 | 0.5473 | 0.6739 | 0.5638 |
60
+ | 1.2954 | 0.11 | 20 | 0.9534 | 0.9091 | 0.9066 | 0.9148 | 0.9082 |
61
+ | 0.7734 | 0.16 | 30 | 0.4544 | 0.9570 | 0.9567 | 0.9585 | 0.9578 |
62
+ | 0.3456 | 0.21 | 40 | 0.1662 | 0.9907 | 0.9908 | 0.9909 | 0.9907 |
63
+ | 0.1226 | 0.26 | 50 | 0.0743 | 0.9914 | 0.9914 | 0.9916 | 0.9914 |
64
+ | 0.0709 | 0.32 | 60 | 0.0446 | 0.9917 | 0.9917 | 0.9921 | 0.9914 |
65
+ | 0.0565 | 0.37 | 70 | 0.0408 | 0.9897 | 0.9898 | 0.9897 | 0.9900 |
66
+ | 0.0172 | 0.42 | 80 | 0.0214 | 0.9967 | 0.9967 | 0.9966 | 0.9968 |
67
+ | 0.0234 | 0.48 | 90 | 0.0217 | 0.9954 | 0.9954 | 0.9954 | 0.9954 |
68
+ | 0.027 | 0.53 | 100 | 0.0158 | 0.9977 | 0.9977 | 0.9976 | 0.9977 |
69
+ | 0.0346 | 0.58 | 110 | 0.0119 | 0.9980 | 0.9980 | 0.9980 | 0.9981 |
70
+ | 0.0621 | 0.63 | 120 | 0.0093 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
71
+ | 0.0065 | 0.69 | 130 | 0.0156 | 0.9960 | 0.9960 | 0.9960 | 0.9960 |
72
+ | 0.006 | 0.74 | 140 | 0.0081 | 0.9980 | 0.9980 | 0.9979 | 0.9980 |
73
+ | 0.0045 | 0.79 | 150 | 0.0091 | 0.9970 | 0.9970 | 0.9969 | 0.9971 |
74
+ | 0.0108 | 0.85 | 160 | 0.0045 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
75
+ | 0.0128 | 0.9 | 170 | 0.0033 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
76
+ | 0.0166 | 0.95 | 180 | 0.0112 | 0.9974 | 0.9974 | 0.9973 | 0.9974 |
77
+ | 0.0031 | 1.01 | 190 | 0.0117 | 0.9974 | 0.9974 | 0.9973 | 0.9974 |
78
+ | 0.0028 | 1.06 | 200 | 0.0143 | 0.9967 | 0.9967 | 0.9967 | 0.9967 |
79
+ | 0.0031 | 1.11 | 210 | 0.0076 | 0.9987 | 0.9987 | 0.9987 | 0.9987 |
80
+ | 0.0126 | 1.16 | 220 | 0.0051 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
81
+ | 0.0194 | 1.22 | 230 | 0.0048 | 0.9993 | 0.9993 | 0.9993 | 0.9993 |
82
+ | 0.0021 | 1.27 | 240 | 0.0093 | 0.9980 | 0.9980 | 0.9980 | 0.9980 |
83
+ | 0.002 | 1.32 | 250 | 0.0082 | 0.9983 | 0.9983 | 0.9983 | 0.9983 |
84
+ | 0.0187 | 1.38 | 260 | 0.0041 | 0.9993 | 0.9993 | 0.9993 | 0.9993 |
85
+ | 0.0018 | 1.43 | 270 | 0.0049 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
86
+ | 0.0016 | 1.48 | 280 | 0.0050 | 0.9987 | 0.9987 | 0.9987 | 0.9986 |
87
+ | 0.0016 | 1.53 | 290 | 0.0047 | 0.9993 | 0.9993 | 0.9993 | 0.9993 |
88
+ | 0.0014 | 1.59 | 300 | 0.0049 | 0.9993 | 0.9993 | 0.9993 | 0.9993 |
89
+ | 0.0014 | 1.64 | 310 | 0.0050 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
90
+ | 0.0013 | 1.69 | 320 | 0.0051 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
91
+ | 0.0089 | 1.75 | 330 | 0.0360 | 0.9911 | 0.9911 | 0.9913 | 0.9911 |
92
+ | 0.0211 | 1.8 | 340 | 0.0042 | 0.9987 | 0.9987 | 0.9987 | 0.9987 |
93
+ | 0.0014 | 1.85 | 350 | 0.0223 | 0.9957 | 0.9957 | 0.9957 | 0.9957 |
94
+ | 0.003 | 1.9 | 360 | 0.0027 | 0.9993 | 0.9993 | 0.9993 | 0.9993 |
95
+ | 0.0086 | 1.96 | 370 | 0.0026 | 0.9993 | 0.9994 | 0.9994 | 0.9993 |
96
+ | 0.0013 | 2.01 | 380 | 0.0023 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
97
+ | 0.0011 | 2.06 | 390 | 0.0037 | 0.9987 | 0.9987 | 0.9987 | 0.9986 |
98
+ | 0.0047 | 2.12 | 400 | 0.0066 | 0.9980 | 0.9980 | 0.9980 | 0.9980 |
99
+ | 0.0013 | 2.17 | 410 | 0.0037 | 0.9990 | 0.9990 | 0.9990 | 0.9990 |
100
+ | 0.001 | 2.22 | 420 | 0.0020 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
101
+ | 0.0009 | 2.28 | 430 | 0.0014 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
102
+ | 0.0009 | 2.33 | 440 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
103
+ | 0.0008 | 2.38 | 450 | 0.0011 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
104
+ | 0.0008 | 2.43 | 460 | 0.0011 | 0.9997 | 0.9997 | 0.9997 | 0.9996 |
105
+ | 0.0009 | 2.49 | 470 | 0.0010 | 1.0 | 1.0 | 1.0 | 1.0 |
106
+ | 0.0008 | 2.54 | 480 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
107
+ | 0.0007 | 2.59 | 490 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
108
+ | 0.0007 | 2.65 | 500 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
109
+ | 0.0007 | 2.7 | 510 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
110
+ | 0.014 | 2.75 | 520 | 0.0019 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
111
+ | 0.0009 | 2.8 | 530 | 0.0036 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
112
+ | 0.0007 | 2.86 | 540 | 0.0036 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
113
+ | 0.0007 | 2.91 | 550 | 0.0032 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
114
+ | 0.0006 | 2.96 | 560 | 0.0028 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
115
+ | 0.0006 | 3.02 | 570 | 0.0026 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
116
+ | 0.0006 | 3.07 | 580 | 0.0024 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
117
+ | 0.0007 | 3.12 | 590 | 0.0023 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
118
+ | 0.0005 | 3.17 | 600 | 0.0021 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
119
+ | 0.0006 | 3.23 | 610 | 0.0021 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
120
+ | 0.0006 | 3.28 | 620 | 0.0018 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
121
+ | 0.0005 | 3.33 | 630 | 0.0017 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
122
+ | 0.0005 | 3.39 | 640 | 0.0016 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
123
+ | 0.0005 | 3.44 | 650 | 0.0014 | 0.9993 | 0.9993 | 0.9994 | 0.9993 |
124
+ | 0.0005 | 3.49 | 660 | 0.0013 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
125
+ | 0.0005 | 3.54 | 670 | 0.0012 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
126
+ | 0.0005 | 3.6 | 680 | 0.0011 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
127
+ | 0.0004 | 3.65 | 690 | 0.0011 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
128
+ | 0.0004 | 3.7 | 700 | 0.0011 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
129
+ | 0.0004 | 3.76 | 710 | 0.0010 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
130
+ | 0.0004 | 3.81 | 720 | 0.0010 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
131
+ | 0.0004 | 3.86 | 730 | 0.0010 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
132
+ | 0.0004 | 3.92 | 740 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
133
+ | 0.0004 | 3.97 | 750 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
134
+ | 0.0004 | 4.02 | 760 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
135
+ | 0.0004 | 4.07 | 770 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
136
+ | 0.0004 | 4.13 | 780 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
137
+ | 0.0004 | 4.18 | 790 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
138
+ | 0.0004 | 4.23 | 800 | 0.0009 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
139
+ | 0.0004 | 4.29 | 810 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
140
+ | 0.0004 | 4.34 | 820 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
141
+ | 0.0004 | 4.39 | 830 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
142
+ | 0.0004 | 4.44 | 840 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
143
+ | 0.0004 | 4.5 | 850 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
144
+ | 0.0003 | 4.55 | 860 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
145
+ | 0.0004 | 4.6 | 870 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
146
+ | 0.0003 | 4.66 | 880 | 0.0008 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
147
+ | 0.0003 | 4.71 | 890 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
148
+ | 0.0003 | 4.76 | 900 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
149
+ | 0.0003 | 4.81 | 910 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
150
+ | 0.0003 | 4.87 | 920 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
151
+ | 0.0003 | 4.92 | 930 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
152
+ | 0.0003 | 4.97 | 940 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
153
+ | 0.0003 | 5.03 | 950 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
154
+ | 0.0003 | 5.08 | 960 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
155
+ | 0.0003 | 5.13 | 970 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
156
+ | 0.0003 | 5.19 | 980 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
157
+ | 0.0003 | 5.24 | 990 | 0.0007 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
158
+ | 0.0003 | 5.29 | 1000 | 0.0006 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
159
+ | 0.0003 | 5.34 | 1010 | 0.0006 | 0.9997 | 0.9997 | 0.9997 | 0.9997 |
160
+ | 0.0003 | 5.4 | 1020 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
161
+ | 0.0003 | 5.45 | 1030 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
162
+ | 0.0003 | 5.5 | 1040 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
163
+ | 0.0003 | 5.56 | 1050 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
164
+ | 0.0003 | 5.61 | 1060 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
165
+ | 0.0003 | 5.66 | 1070 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
166
+ | 0.0003 | 5.71 | 1080 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
167
+ | 0.0003 | 5.77 | 1090 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
168
+ | 0.0003 | 5.82 | 1100 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
169
+ | 0.0003 | 5.87 | 1110 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
170
+ | 0.0003 | 5.93 | 1120 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
171
+ | 0.0003 | 5.98 | 1130 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
172
+ | 0.0003 | 6.03 | 1140 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
173
+ | 0.0003 | 6.08 | 1150 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
174
+ | 0.0003 | 6.14 | 1160 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
175
+ | 0.0003 | 6.19 | 1170 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
176
+ | 0.0003 | 6.24 | 1180 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
177
+ | 0.0003 | 6.3 | 1190 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
178
+ | 0.0002 | 6.35 | 1200 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
179
+ | 0.0003 | 6.4 | 1210 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
180
+ | 0.0002 | 6.46 | 1220 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
181
+ | 0.0002 | 6.51 | 1230 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
182
+ | 0.0002 | 6.56 | 1240 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
183
+ | 0.0002 | 6.61 | 1250 | 0.0006 | 1.0 | 1.0 | 1.0 | 1.0 |
184
+ | 0.0002 | 6.67 | 1260 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
185
+ | 0.0002 | 6.72 | 1270 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
186
+ | 0.0002 | 6.77 | 1280 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
187
+ | 0.0002 | 6.83 | 1290 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
188
+ | 0.0002 | 6.88 | 1300 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
189
+ | 0.0002 | 6.93 | 1310 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
190
+ | 0.0002 | 6.98 | 1320 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
191
+ | 0.0002 | 7.04 | 1330 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
192
+ | 0.0002 | 7.09 | 1340 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
193
+ | 0.0002 | 7.14 | 1350 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
194
+ | 0.0002 | 7.2 | 1360 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
195
+ | 0.0002 | 7.25 | 1370 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
196
+ | 0.0002 | 7.3 | 1380 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
197
+ | 0.0002 | 7.35 | 1390 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
198
+ | 0.0002 | 7.41 | 1400 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
199
+ | 0.0002 | 7.46 | 1410 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
200
+ | 0.0002 | 7.51 | 1420 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
201
+ | 0.0002 | 7.57 | 1430 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
202
+ | 0.0002 | 7.62 | 1440 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
203
+ | 0.0002 | 7.67 | 1450 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
204
+ | 0.0002 | 7.72 | 1460 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
205
+ | 0.0002 | 7.78 | 1470 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
206
+ | 0.0002 | 7.83 | 1480 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
207
+ | 0.0002 | 7.88 | 1490 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
208
+ | 0.0002 | 7.94 | 1500 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
209
+ | 0.0002 | 7.99 | 1510 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
210
+ | 0.0002 | 8.04 | 1520 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
211
+ | 0.0002 | 8.1 | 1530 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
212
+ | 0.0002 | 8.15 | 1540 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
213
+ | 0.0002 | 8.2 | 1550 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
214
+ | 0.0002 | 8.25 | 1560 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
215
+ | 0.0002 | 8.31 | 1570 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
216
+ | 0.0002 | 8.36 | 1580 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
217
+ | 0.0002 | 8.41 | 1590 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
218
+ | 0.0002 | 8.47 | 1600 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
219
+ | 0.0002 | 8.52 | 1610 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
220
+ | 0.0002 | 8.57 | 1620 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
221
+ | 0.0002 | 8.62 | 1630 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
222
+ | 0.0002 | 8.68 | 1640 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
223
+ | 0.0002 | 8.73 | 1650 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
224
+ | 0.0002 | 8.78 | 1660 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
225
+ | 0.0002 | 8.84 | 1670 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
226
+ | 0.0002 | 8.89 | 1680 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
227
+ | 0.0002 | 8.94 | 1690 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
228
+ | 0.0002 | 8.99 | 1700 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
229
+ | 0.0002 | 9.05 | 1710 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
230
+ | 0.0002 | 9.1 | 1720 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
231
+ | 0.0002 | 9.15 | 1730 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
232
+ | 0.0002 | 9.21 | 1740 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
233
+ | 0.0002 | 9.26 | 1750 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
234
+ | 0.0002 | 9.31 | 1760 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
235
+ | 0.0002 | 9.37 | 1770 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
236
+ | 0.0002 | 9.42 | 1780 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
237
+ | 0.0002 | 9.47 | 1790 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
238
+ | 0.0002 | 9.52 | 1800 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
239
+ | 0.0002 | 9.58 | 1810 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
240
+ | 0.0002 | 9.63 | 1820 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
241
+ | 0.0002 | 9.68 | 1830 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
242
+ | 0.0002 | 9.74 | 1840 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
243
+ | 0.0002 | 9.79 | 1850 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
244
+ | 0.0002 | 9.84 | 1860 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
245
+ | 0.0002 | 9.89 | 1870 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
246
+ | 0.0002 | 9.95 | 1880 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
247
+ | 0.0002 | 10.0 | 1890 | 0.0005 | 1.0 | 1.0 | 1.0 | 1.0 |
248
 
249
 
250
  ### Framework versions
251
 
 
252
  - Transformers 4.38.2
253
  - Pytorch 2.1.2
254
  - Datasets 2.1.0
255
+ - Tokenizers 0.15.2
config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "Install",
14
+ "1": "Compensation",
15
+ "2": "Change",
16
+ "3": "Summarize",
17
+ "4": "Terminate",
18
+ "5": "SBC"
19
+ },
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "Change": 2,
24
+ "Compensation": 1,
25
+ "Install": 0,
26
+ "SBC": 5,
27
+ "Summarize": 3,
28
+ "Terminate": 4
29
+ },
30
+ "layer_norm_eps": 1e-12,
31
+ "max_position_embeddings": 512,
32
+ "model_type": "bert",
33
+ "num_attention_heads": 12,
34
+ "num_hidden_layers": 12,
35
+ "pad_token_id": 0,
36
+ "position_embedding_type": "absolute",
37
+ "problem_type": "single_label_classification",
38
+ "torch_dtype": "float32",
39
+ "transformers_version": "4.38.2",
40
+ "type_vocab_size": 2,
41
+ "use_cache": true,
42
+ "vocab_size": 30522
43
+ }
events.out.tfevents.1714375529.be592b10bdc8.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81a3b183bb1ccf1fdac789d897f34488c3c43e312a40524a9fe80c158f33374a
3
+ size 134109
events.out.tfevents.1714379632.be592b10bdc8.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:947bba194795d6d4484abbef678c02acca862f62d03ea140c687f92c3ddb2108
3
+ size 1504
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a62c4d9f70c329a6afe6e67264509fbef080878b83aa4afd472515ea7e9dc32
3
+ size 437970952
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "max_length": 512,
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca955b9d3c9b82502fe8e64a06e41aa6e80a7a1e0c2c5a62b4ae59a71c03c025
3
- size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6938a7f9f012635cb9368fe55a61a7d36c6dc8845a5f1454e394f1f564fb5b4
3
+ size 4856
vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,40 +1,109 @@
1
- 2024-04-17 07:43:36,940 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
- 2024-04-17 07:43:36,940 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-04-17 07:43:36,940 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-17 07:43:36,940 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240417_074336-4ny3o4w8/logs/debug.log
10
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240417_074336-4ny3o4w8/logs/debug-internal.log
11
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ab2bc9a3ee0>
12
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
- 2024-04-17 07:43:36,941 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
- 2024-04-17 07:43:36,943 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-17 07:43:36,948 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
- 2024-04-17 07:43:36,960 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
- 2024-04-17 07:43:37,610 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
- 2024-04-17 07:43:37,619 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
- 2024-04-17 07:43:37,779 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
- 2024-04-17 07:43:37,868 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-17 07:43:37,868 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
- 2024-04-17 07:44:08,962 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
- 2024-04-17 07:44:08,963 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
- 2024-04-17 07:44:08,963 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
- 2024-04-17 07:44:08,963 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
- 2024-04-17 07:44:08,965 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
- 2024-04-17 07:44:08,973 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'Narkantak/Mistral-2x7b-Instruct-1x2', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'fp4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'bfloat16', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr17_07-43-22_a7400a4704ae', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
- 2024-04-17 09:08:40,460 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
- 2024-04-17 09:08:40,460 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
- 2024-04-17 09:25:14,934 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
- 2024-04-17 09:25:14,958 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
- 2024-04-17 09:25:14,958 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
- 2024-04-17 09:25:42,198 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
- 2024-04-17 09:25:42,199 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
- 2024-04-17 09:25:42,200 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
- 2024-04-17 09:25:43,273 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240429_072705-0t7z1bit/logs/debug.log
10
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240429_072705-0t7z1bit/logs/debug-internal.log
11
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7b7c6cebd510>
12
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-29 07:27:05,567 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-29 07:27:05,570 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-29 07:27:05,587 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-29 07:27:06,060 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-29 07:27:06,063 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-29 07:27:06,249 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-29 07:27:06,331 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-29 07:27:06,331 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-29 07:27:37,477 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-29 07:27:37,477 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-29 07:27:37,478 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-29 07:27:37,478 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-29 07:27:37,479 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-29 07:27:37,487 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'Install', 1: 'Compensation', 2: 'Change', 3: 'Summarize', 4: 'Terminate', 5: 'SBC'}, 'label2id': {'Install': 0, 'Compensation': 1, 'Change': 2, 'Summarize': 3, 'Terminate': 4, 'SBC': 5}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'google-bert/bert-base-uncased', 'transformers_version': '4.38.2', 'gradient_checkpointing': False, 'model_type': 'bert', 'vocab_size': 30522, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 10, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 10, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-29 08:32:33,725 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-29 08:32:33,725 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-29 08:33:23,264 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-29 08:34:18,223 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-29 08:34:18,224 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-29 08:34:43,696 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-29 08:34:43,697 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-04-29 08:34:43,697 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
+ 2024-04-29 08:35:51,215 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
41
+ 2024-04-29 08:35:51,218 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-04-29 08:35:51,218 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
43
+ 2024-04-29 08:35:55,649 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
44
+ 2024-04-29 08:35:58,171 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-04-29 08:35:58,172 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
46
+ 2024-04-29 08:36:23,490 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
47
+ 2024-04-29 08:36:23,493 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-04-29 08:36:23,493 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
49
+ 2024-04-29 08:36:24,474 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
50
+ 2024-04-29 08:36:25,328 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-04-29 08:36:25,329 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
52
+ 2024-04-29 08:36:47,420 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
53
+ 2024-04-29 08:36:47,423 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-29 08:36:47,423 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
55
+ 2024-04-29 08:36:48,363 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
56
+ 2024-04-29 08:36:48,681 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-29 08:36:48,682 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
58
+ 2024-04-29 08:37:26,761 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
59
+ 2024-04-29 08:37:27,658 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-04-29 08:37:27,658 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
61
+ 2024-04-29 08:37:31,816 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
62
+ 2024-04-29 08:37:31,849 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
63
+ 2024-04-29 08:37:31,849 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
64
+ 2024-04-29 08:37:48,262 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
65
+ 2024-04-29 08:37:48,997 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
66
+ 2024-04-29 08:37:48,997 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
67
+ 2024-04-29 08:38:08,126 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
68
+ 2024-04-29 08:38:09,034 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
69
+ 2024-04-29 08:38:09,034 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
70
+ 2024-04-29 08:38:12,014 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
71
+ 2024-04-29 08:38:12,041 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
72
+ 2024-04-29 08:38:12,041 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
73
+ 2024-04-29 08:38:24,975 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
74
+ 2024-04-29 08:38:25,120 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
75
+ 2024-04-29 08:38:25,121 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
76
+ 2024-04-29 08:39:43,660 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
77
+ 2024-04-29 08:39:43,985 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
78
+ 2024-04-29 08:39:43,985 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
79
+ 2024-04-29 08:39:57,431 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
80
+ 2024-04-29 08:39:57,728 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
81
+ 2024-04-29 08:39:57,728 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
82
+ 2024-04-29 08:40:00,496 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
83
+ 2024-04-29 08:40:00,705 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
84
+ 2024-04-29 08:40:00,705 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
85
+ 2024-04-29 08:40:04,198 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
86
+ 2024-04-29 08:40:04,256 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
87
+ 2024-04-29 08:40:04,256 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
88
+ 2024-04-29 08:40:07,104 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
89
+ 2024-04-29 08:40:07,185 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
90
+ 2024-04-29 08:40:07,186 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
91
+ 2024-04-29 08:40:13,433 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
92
+ 2024-04-29 08:40:13,507 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
93
+ 2024-04-29 08:40:13,507 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
94
+ 2024-04-29 08:40:17,190 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
95
+ 2024-04-29 08:40:17,288 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
96
+ 2024-04-29 08:40:17,288 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
97
+ 2024-04-29 08:40:23,262 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
98
+ 2024-04-29 08:40:23,268 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
99
+ 2024-04-29 08:40:23,268 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
100
+ 2024-04-29 08:40:33,903 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
101
+ 2024-04-29 08:40:36,980 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
102
+ 2024-04-29 08:40:36,980 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
103
+ 2024-04-29 08:40:43,896 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
104
+ 2024-04-29 08:40:47,028 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
105
+ 2024-04-29 08:40:47,028 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
106
+ 2024-04-29 08:45:12,809 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
107
+ 2024-04-29 08:45:12,981 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
108
+ 2024-04-29 08:45:12,981 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
109
+ 2024-04-29 08:52:31,675 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
wandb/run-20240429_072705-0t7z1bit/files/conda-environment.yaml ADDED
@@ -0,0 +1,1070 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: base
2
+ channels:
3
+ - pytorch
4
+ - file:///tmp/conda
5
+ - rapidsai
6
+ - nvidia
7
+ - conda-forge
8
+ - defaults
9
+ dependencies:
10
+ - _libgcc_mutex=0.1=conda_forge
11
+ - _openmp_mutex=4.5=2_gnu
12
+ - aiohttp=3.9.1=py310h2372a71_0
13
+ - aiosignal=1.3.1=pyhd8ed1ab_0
14
+ - annotated-types=0.6.0=pyhd8ed1ab_0
15
+ - anyio=4.2.0=pyhd8ed1ab_0
16
+ - archspec=0.2.2=pyhd8ed1ab_0
17
+ - argon2-cffi=23.1.0=pyhd8ed1ab_0
18
+ - argon2-cffi-bindings=21.2.0=py310h2372a71_4
19
+ - arrow=1.3.0=pyhd8ed1ab_0
20
+ - arrow-cpp=11.0.0=ha770c72_9_cpu
21
+ - asttokens=2.4.1=pyhd8ed1ab_0
22
+ - async-timeout=4.0.3=pyhd8ed1ab_0
23
+ - atk-1.0=2.38.0=hd4edc92_1
24
+ - attrs=23.2.0=pyh71513ae_0
25
+ - aws-c-auth=0.6.26=h987a71b_2
26
+ - aws-c-cal=0.5.21=h48707d8_2
27
+ - aws-c-common=0.8.14=h0b41bf4_0
28
+ - aws-c-compression=0.2.16=h03acc5a_5
29
+ - aws-c-event-stream=0.2.20=h00877a2_4
30
+ - aws-c-http=0.7.6=hf342b9f_0
31
+ - aws-c-io=0.13.19=h5b20300_3
32
+ - aws-c-mqtt=0.8.6=hc4349f7_12
33
+ - aws-c-s3=0.2.7=h909e904_1
34
+ - aws-c-sdkutils=0.1.9=h03acc5a_0
35
+ - aws-checksums=0.1.14=h03acc5a_5
36
+ - aws-crt-cpp=0.19.8=hf7fbfca_12
37
+ - aws-sdk-cpp=1.10.57=h17c43bd_8
38
+ - beautifulsoup4=4.12.2=pyha770c72_0
39
+ - bleach=6.1.0=pyhd8ed1ab_0
40
+ - boltons=23.1.1=pyhd8ed1ab_0
41
+ - brotli=1.0.9=h166bdaf_9
42
+ - brotli-bin=1.0.9=h166bdaf_9
43
+ - brotli-python=1.0.9=py310hd8f1fbe_9
44
+ - brotlipy=0.7.0=py310h7f8727e_1002
45
+ - bzip2=1.0.8=h7b6447c_0
46
+ - c-ares=1.25.0=hd590300_0
47
+ - ca-certificates=2024.2.2=hbcca054_0
48
+ - cached-property=1.5.2=hd8ed1ab_1
49
+ - cached_property=1.5.2=pyha770c72_1
50
+ - cairo=1.18.0=h3faef2a_0
51
+ - cartopy=0.22.0=py310hcc13569_1
52
+ - catalogue=2.0.10=py310hff52083_0
53
+ - certifi=2024.2.2=pyhd8ed1ab_0
54
+ - cffi=1.16.0=py310h2fee648_0
55
+ - charset-normalizer=3.3.2=pyhd8ed1ab_0
56
+ - click=8.1.7=unix_pyh707e725_0
57
+ - cloudpathlib=0.16.0=pyhd8ed1ab_0
58
+ - colorama=0.4.6=pyhd8ed1ab_0
59
+ - comm=0.2.1=pyhd8ed1ab_0
60
+ - conda=23.7.4=py310hff52083_0
61
+ - conda-libmamba-solver=23.7.0=pyhd8ed1ab_0
62
+ - conda-package-handling=2.2.0=pyh38be061_0
63
+ - conda-package-streaming=0.9.0=pyhd8ed1ab_0
64
+ - confection=0.1.4=py310h17c5347_0
65
+ - contourpy=1.2.0=py310hd41b1e2_0
66
+ - cryptography=41.0.7=py310hb8475ec_1
67
+ - cuda-cccl=12.4.99=0
68
+ - cuda-cudart=12.4.99=0
69
+ - cuda-cudart-dev=12.4.99=0
70
+ - cuda-nvcc-dev_linux-64=12.1.105=ha770c72_0
71
+ - cuda-nvcc-impl=12.1.105=hd3aeb46_0
72
+ - cuda-nvcc-tools=12.1.105=hd3aeb46_0
73
+ - cuda-nvrtc=12.1.105=hd3aeb46_0
74
+ - cuda-profiler-api=12.4.99=0
75
+ - cuda-python=12.4.0=py310h52dc4f0_0
76
+ - cuda-version=12.1=h1d6eff3_3
77
+ - cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
78
+ - cuml=23.08.00=cuda12_py310_230809_gd7162cdea_0
79
+ - cupy=13.0.0=py310h7aad9d2_3
80
+ - cupy-core=13.0.0=py310had4011e_3
81
+ - curl=8.6.0=hca28451_0
82
+ - cycler=0.12.1=pyhd8ed1ab_0
83
+ - cymem=2.0.8=py310hc6cd4ac_1
84
+ - cython-blis=0.7.10=py310h1f7b6fc_2
85
+ - cytoolz=0.12.3=py310h2372a71_0
86
+ - dask-cuda=23.08.00=py310_230809_gefbd6ca_0
87
+ - dask-cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
88
+ - debugpy=1.8.0=py310hc6cd4ac_1
89
+ - decorator=5.1.1=pyhd8ed1ab_0
90
+ - defusedxml=0.7.1=pyhd8ed1ab_0
91
+ - distributed=2023.7.1=pyhd8ed1ab_0
92
+ - distro=1.9.0=pyhd8ed1ab_0
93
+ - dlenv-tf-2-15-gpu=1.0.20240111=py310ha20f8e0_0
94
+ - dlpack=0.5=h9c3ff4c_0
95
+ - entrypoints=0.4=pyhd8ed1ab_0
96
+ - exceptiongroup=1.2.0=pyhd8ed1ab_2
97
+ - executing=2.0.1=pyhd8ed1ab_0
98
+ - expat=2.6.2=h59595ed_0
99
+ - fastrlock=0.8.2=py310hc6cd4ac_2
100
+ - fftw=3.3.10=nompi_hc118613_108
101
+ - fmt=9.1.0=h924138e_0
102
+ - font-ttf-dejavu-sans-mono=2.37=hab24e00_0
103
+ - font-ttf-inconsolata=3.000=h77eed37_0
104
+ - font-ttf-source-code-pro=2.038=h77eed37_0
105
+ - font-ttf-ubuntu=0.83=h77eed37_1
106
+ - fontconfig=2.14.2=h14ed4e7_0
107
+ - fonts-conda-ecosystem=1=0
108
+ - fonts-conda-forge=1=0
109
+ - fqdn=1.5.1=pyhd8ed1ab_0
110
+ - freetype=2.12.1=h267a509_2
111
+ - fribidi=1.0.10=h36c2ea0_0
112
+ - frozenlist=1.4.1=py310h2372a71_0
113
+ - fsspec=2024.3.0=pyhca7485f_0
114
+ - gdk-pixbuf=2.42.10=h829c605_5
115
+ - geos=3.11.1=h27087fc_0
116
+ - gettext=0.21.1=h27087fc_0
117
+ - gflags=2.2.2=he1b5a44_1004
118
+ - ghostscript=10.03.0=h59595ed_0
119
+ - giflib=5.2.1=h0b41bf4_3
120
+ - glog=0.6.0=h6f12383_0
121
+ - gmock=1.14.0=ha770c72_1
122
+ - gmp=6.3.0=h59595ed_0
123
+ - google-api-core-grpc=2.11.1=hd8ed1ab_0
124
+ - google-auth=2.26.1=pyhca7485f_0
125
+ - google-cloud-core=2.4.1=pyhd8ed1ab_0
126
+ - google-cloud-datastore=2.19.0=pyhd8ed1ab_0
127
+ - googleapis-common-protos=1.62.0=pyhd8ed1ab_0
128
+ - graphite2=1.3.13=h58526e2_1001
129
+ - graphviz=9.0.0=h78e8752_1
130
+ - grpc-cpp=1.51.1=h27aab58_3
131
+ - gtest=1.14.0=h00ab1b0_1
132
+ - gtk2=2.24.33=h280cfa0_4
133
+ - gts=0.7.6=h977cf35_4
134
+ - harfbuzz=8.3.0=h3d44ed6_0
135
+ - icu=73.2=h59595ed_0
136
+ - idna=3.6=pyhd8ed1ab_0
137
+ - imagemagick=7.1.1_29=pl5321hb90aeea_0
138
+ - importlib_metadata=7.0.1=hd8ed1ab_0
139
+ - importlib_resources=6.1.1=pyhd8ed1ab_0
140
+ - intel-openmp=2023.1.0=hdb19cb5_46306
141
+ - ipykernel=6.28.0=pyhd33586a_0
142
+ - ipython=8.20.0=pyh707e725_0
143
+ - ipython_genutils=0.2.0=py_1
144
+ - isoduration=20.11.0=pyhd8ed1ab_0
145
+ - jbig=2.1=h7f98852_2003
146
+ - jedi=0.19.1=pyhd8ed1ab_0
147
+ - jinja2=3.1.2=pyhd8ed1ab_1
148
+ - joblib=1.3.2=pyhd8ed1ab_0
149
+ - jsonpatch=1.33=pyhd8ed1ab_0
150
+ - jsonpointer=2.4=py310hff52083_3
151
+ - jsonschema=4.20.0=pyhd8ed1ab_0
152
+ - jsonschema-specifications=2023.12.1=pyhd8ed1ab_0
153
+ - jsonschema-with-format-nongpl=4.20.0=pyhd8ed1ab_0
154
+ - jupyter_client=8.6.0=pyhd8ed1ab_0
155
+ - jupyter_core=5.7.1=py310hff52083_0
156
+ - jupyter_events=0.9.0=pyhd8ed1ab_0
157
+ - jupyter_server_terminals=0.5.1=pyhd8ed1ab_0
158
+ - jupyterlab_pygments=0.3.0=pyhd8ed1ab_0
159
+ - keyutils=1.6.1=h166bdaf_0
160
+ - kiwisolver=1.4.5=py310hd41b1e2_1
161
+ - krb5=1.21.2=h659d440_0
162
+ - langcodes=3.3.0=pyhd8ed1ab_0
163
+ - lcms2=2.16=hb7c19ff_0
164
+ - ld_impl_linux-64=2.40=h41732ed_0
165
+ - lerc=4.0.0=h27087fc_0
166
+ - libabseil=20230125.0=cxx17_hcb278e6_1
167
+ - libarchive=3.6.2=h039dbb9_1
168
+ - libarrow=11.0.0=h33598ff_9_cpu
169
+ - libblas=3.9.0=21_linux64_openblas
170
+ - libbrotlicommon=1.0.9=h166bdaf_9
171
+ - libbrotlidec=1.0.9=h166bdaf_9
172
+ - libbrotlienc=1.0.9=h166bdaf_9
173
+ - libcblas=3.9.0=21_linux64_openblas
174
+ - libcrc32c=1.1.2=h9c3ff4c_0
175
+ - libcublas=12.1.3.1=hd3aeb46_0
176
+ - libcublas-dev=12.1.3.1=0
177
+ - libcudf=23.08.00=cuda12_230809_g8150d38e08_0
178
+ - libcufft=11.0.2.54=hd3aeb46_0
179
+ - libcufile=1.9.0.20=0
180
+ - libcufile-dev=1.9.0.20=0
181
+ - libcuml=23.08.00=cuda12_230809_gd7162cdea_0
182
+ - libcumlprims=23.08.00=cuda12_230809_g71c0a86_0
183
+ - libcurand=10.3.2.106=hd3aeb46_0
184
+ - libcurand-dev=10.3.2.106=0
185
+ - libcurl=8.6.0=hca28451_0
186
+ - libcusolver=11.4.5.107=hd3aeb46_0
187
+ - libcusolver-dev=11.4.5.107=0
188
+ - libcusparse=12.1.0.106=hd3aeb46_0
189
+ - libcusparse-dev=12.1.0.106=0
190
+ - libdeflate=1.19=hd590300_0
191
+ - libedit=3.1.20191231=he28a2e2_2
192
+ - libev=4.33=hd590300_2
193
+ - libevent=2.1.10=h28343ad_4
194
+ - libexpat=2.6.2=h59595ed_0
195
+ - libffi=3.4.2=h7f98852_5
196
+ - libgcc-ng=13.2.0=h807b86a_3
197
+ - libgd=2.3.3=h119a65a_9
198
+ - libgfortran-ng=13.2.0=h69a702a_5
199
+ - libgfortran5=13.2.0=ha4646dd_5
200
+ - libglib=2.80.0=hf2295e7_0
201
+ - libgomp=13.2.0=h807b86a_3
202
+ - libgoogle-cloud=2.8.0=h3c06191_0
203
+ - libgrpc=1.51.1=hcf146ea_3
204
+ - libhwloc=2.9.3=default_h554bfaf_1009
205
+ - libiconv=1.17=hd590300_2
206
+ - libjpeg-turbo=3.0.0=hd590300_1
207
+ - libkvikio=23.08.00=cuda12_230809_g51a9036_0
208
+ - liblapack=3.9.0=21_linux64_openblas
209
+ - libllvm14=14.0.6=hcd5def8_4
210
+ - libmamba=1.5.0=h658169a_0
211
+ - libmambapy=1.5.0=py310h8aae740_0
212
+ - libnghttp2=1.58.0=h47da74e_1
213
+ - libnsl=2.0.1=hd590300_0
214
+ - libnuma=2.0.18=hd590300_0
215
+ - libnvjitlink=12.1.105=hd3aeb46_0
216
+ - libopenblas=0.3.26=pthreads_h413a1c8_0
217
+ - libpng=1.6.43=h2797004_0
218
+ - libprotobuf=3.21.12=hfc55251_2
219
+ - libraft=23.08.00=cuda12_230809_ge588d7b5_0
220
+ - libraft-headers=23.08.00=cuda12_230809_ge588d7b5_0
221
+ - libraft-headers-only=23.08.00=cuda12_230809_ge588d7b5_0
222
+ - librmm=23.08.00=cuda12_230809_gf3af0e8d_0
223
+ - librsvg=2.56.3=he3f83f7_1
224
+ - libsodium=1.0.18=h36c2ea0_1
225
+ - libsolv=0.7.27=hfc55251_0
226
+ - libsqlite=3.44.2=h2797004_0
227
+ - libssh2=1.11.0=h0841786_0
228
+ - libstdcxx-ng=13.2.0=h7e041cc_3
229
+ - libthrift=0.18.0=h5e4af38_0
230
+ - libtiff=4.6.0=ha9c0a0a_2
231
+ - libutf8proc=2.8.0=h166bdaf_0
232
+ - libuuid=2.38.1=h0b41bf4_0
233
+ - libuv=1.46.0=hd590300_0
234
+ - libwebp=1.3.2=h658648e_1
235
+ - libwebp-base=1.3.2=hd590300_0
236
+ - libxcb=1.15=h0b41bf4_0
237
+ - libxcrypt=4.4.36=hd590300_1
238
+ - libxml2=2.12.6=h232c23b_0
239
+ - libzlib=1.2.13=hd590300_5
240
+ - llvm-openmp=8.0.1=hc9558a2_0
241
+ - locket=1.0.0=pyhd8ed1ab_0
242
+ - lz4=4.3.3=py310h350c4a5_0
243
+ - lz4-c=1.9.4=hcb278e6_0
244
+ - lzo=2.10=h516909a_1000
245
+ - magma-cuda121=2.6.1=1
246
+ - mamba=1.5.0=py310h51d5547_0
247
+ - markdown-it-py=3.0.0=pyhd8ed1ab_0
248
+ - matplotlib-base=3.8.3=py310h62c0568_0
249
+ - matplotlib-inline=0.1.6=pyhd8ed1ab_0
250
+ - mdurl=0.1.2=pyhd8ed1ab_0
251
+ - menuinst=2.0.1=py310hff52083_0
252
+ - mkl=2023.1.0=h213fc3f_46344
253
+ - msgpack-python=1.0.7=py310hd41b1e2_0
254
+ - multidict=6.0.4=py310h2372a71_1
255
+ - munkres=1.1.4=pyh9f0ad1d_0
256
+ - murmurhash=1.0.10=py310hc6cd4ac_1
257
+ - nb_conda=2.2.1=unix_7
258
+ - nb_conda_kernels=2.3.1=pyhd8ed1ab_3
259
+ - nbclassic=1.0.0=pyhb4ecaf3_1
260
+ - nbconvert-pandoc=7.14.0=pyhd8ed1ab_0
261
+ - nbformat=5.9.2=pyhd8ed1ab_0
262
+ - nccl=2.20.5.1=h3a97aeb_0
263
+ - ncurses=6.4=h59595ed_2
264
+ - nest-asyncio=1.5.8=pyhd8ed1ab_0
265
+ - nodejs=20.9.0=hb753e55_0
266
+ - notebook-shim=0.2.3=pyhd8ed1ab_0
267
+ - numpy=1.26.4=py310hb13e2d6_0
268
+ - nvcomp=2.6.1=h10b603f_3
269
+ - nvtx=0.2.10=py310h2372a71_0
270
+ - openjpeg=2.5.2=h488ebb8_0
271
+ - openmp=8.0.1=0
272
+ - openssl=3.2.1=hd590300_0
273
+ - orc=1.8.2=hfdbbad2_2
274
+ - overrides=7.4.0=pyhd8ed1ab_0
275
+ - pandoc=3.1.3=h32600fe_0
276
+ - pandocfilters=1.5.0=pyhd8ed1ab_0
277
+ - pango=1.52.1=ha41ecd1_0
278
+ - parquet-cpp=1.5.1=2
279
+ - parso=0.8.3=pyhd8ed1ab_0
280
+ - partd=1.4.1=pyhd8ed1ab_0
281
+ - pathy=0.10.3=py310h06a4308_0
282
+ - pcre2=10.43=hcad00b1_0
283
+ - perl=5.32.1=7_hd590300_perl5
284
+ - pickleshare=0.7.5=py_1003
285
+ - pip=23.3.2=pyhd8ed1ab_0
286
+ - pixman=0.43.2=h59595ed_0
287
+ - pkg-config=0.29.2=h36c2ea0_1008
288
+ - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_1
289
+ - preshed=3.0.9=py310hc6cd4ac_1
290
+ - proj=9.3.1=h1d62c97_0
291
+ - prometheus_client=0.19.0=pyhd8ed1ab_0
292
+ - proto-plus=1.23.0=pyhd8ed1ab_0
293
+ - pthread-stubs=0.4=h36c2ea0_1001
294
+ - ptyprocess=0.7.0=pyhd3deb0d_0
295
+ - pure_eval=0.2.2=pyhd8ed1ab_0
296
+ - pyarrow=11.0.0=py310h633f555_9_cpu
297
+ - pyasn1=0.5.1=pyhd8ed1ab_0
298
+ - pyasn1-modules=0.3.0=pyhd8ed1ab_0
299
+ - pybind11-abi=4=hd8ed1ab_3
300
+ - pycosat=0.6.6=py310h2372a71_0
301
+ - pygments=2.17.2=pyhd8ed1ab_0
302
+ - pylibraft=23.08.00=cuda12_py310_230809_ge588d7b5_0
303
+ - pynvml=11.4.1=pyhd8ed1ab_0
304
+ - pyopenssl=23.3.0=pyhd8ed1ab_0
305
+ - pyproj=3.6.1=py310hd5c30f3_5
306
+ - pyshp=2.3.1=pyhd8ed1ab_0
307
+ - pysocks=1.7.1=py310h06a4308_0
308
+ - python=3.10.13=hd12c33a_1_cpython
309
+ - python-fastjsonschema=2.19.1=pyhd8ed1ab_0
310
+ - python-json-logger=2.0.7=pyhd8ed1ab_0
311
+ - python_abi=3.10=4_cp310
312
+ - pyu2f=0.1.5=pyhd8ed1ab_0
313
+ - pyyaml=6.0.1=py310h2372a71_1
314
+ - raft-dask=23.08.00=cuda12_py310_230809_ge588d7b5_0
315
+ - rdma-core=28.9=h59595ed_1
316
+ - re2=2023.02.02=hcb278e6_0
317
+ - readline=8.2=h8228510_1
318
+ - referencing=0.32.1=pyhd8ed1ab_0
319
+ - reproc=14.2.4.post0=hd590300_1
320
+ - reproc-cpp=14.2.4.post0=h59595ed_1
321
+ - requests=2.31.0=pyhd8ed1ab_0
322
+ - rfc3339-validator=0.1.4=pyhd8ed1ab_0
323
+ - rfc3986-validator=0.1.1=pyh9f0ad1d_0
324
+ - rmm=23.08.00=cuda12_py310_230809_gf3af0e8d_0
325
+ - rpds-py=0.16.2=py310hcb5633a_0
326
+ - rsa=4.9=pyhd8ed1ab_0
327
+ - ruamel.yaml=0.17.40=py310h2372a71_0
328
+ - ruamel.yaml.clib=0.2.7=py310h2372a71_2
329
+ - ruamel_yaml=0.15.100=py310h7f8727e_0
330
+ - s2n=1.3.41=h3358134_0
331
+ - send2trash=1.8.2=pyh41d4057_0
332
+ - setuptools=69.0.3=pyhd8ed1ab_0
333
+ - shellingham=1.5.4=pyhd8ed1ab_0
334
+ - smart_open=6.4.0=pyhd8ed1ab_0
335
+ - snappy=1.1.10=h9fff704_0
336
+ - sniffio=1.3.0=pyhd8ed1ab_0
337
+ - sortedcontainers=2.4.0=pyhd8ed1ab_0
338
+ - soupsieve=2.5=pyhd8ed1ab_1
339
+ - spacy=3.7.2=py310hcb52e73_0
340
+ - spacy-legacy=3.0.12=pyhd8ed1ab_0
341
+ - spacy-loggers=1.0.5=pyhd8ed1ab_0
342
+ - spdlog=1.11.0=h9b3ece8_1
343
+ - sqlite=3.38.2=hc218d9a_0
344
+ - srsly=2.4.8=py310hc6cd4ac_1
345
+ - stack_data=0.6.2=pyhd8ed1ab_0
346
+ - tblib=3.0.0=pyhd8ed1ab_0
347
+ - terminado=0.18.0=pyh0d859eb_0
348
+ - thinc=8.2.2=py310hcb52e73_0
349
+ - tinycss2=1.2.1=pyhd8ed1ab_0
350
+ - tk=8.6.13=noxft_h4845f30_101
351
+ - toolz=0.12.1=pyhd8ed1ab_0
352
+ - tornado=6.3.3=py310h2372a71_1
353
+ - tqdm=4.66.1=pyhd8ed1ab_0
354
+ - traitlets=5.9.0=pyhd8ed1ab_0
355
+ - treelite=3.2.0=py310h1be96d9_0
356
+ - truststore=0.8.0=pyhd8ed1ab_0
357
+ - typer=0.9.0=pyhd8ed1ab_0
358
+ - types-python-dateutil=2.8.19.20240106=pyhd8ed1ab_0
359
+ - typing-extensions=4.9.0=hd8ed1ab_0
360
+ - typing_extensions=4.9.0=pyha770c72_0
361
+ - typing_utils=0.1.0=pyhd8ed1ab_0
362
+ - ucx=1.14.1=h195a15c_5
363
+ - ucx-proc=1.0.0=gpu
364
+ - ucx-py=0.33.00=py310_230809_gea1eb8f_0
365
+ - unicodedata2=15.1.0=py310h2372a71_0
366
+ - uri-template=1.3.0=pyhd8ed1ab_0
367
+ - wasabi=1.1.2=py310hff52083_0
368
+ - wcwidth=0.2.13=pyhd8ed1ab_0
369
+ - weasel=0.3.4=pyhd8ed1ab_0
370
+ - webcolors=1.13=pyhd8ed1ab_0
371
+ - webencodings=0.5.1=pyhd8ed1ab_2
372
+ - websocket-client=1.7.0=pyhd8ed1ab_0
373
+ - wheel=0.42.0=pyhd8ed1ab_0
374
+ - xorg-kbproto=1.0.7=h7f98852_1002
375
+ - xorg-libice=1.1.1=hd590300_0
376
+ - xorg-libsm=1.2.4=h7391055_0
377
+ - xorg-libx11=1.8.7=h8ee46fc_0
378
+ - xorg-libxau=1.0.11=hd590300_0
379
+ - xorg-libxdmcp=1.1.3=h7f98852_0
380
+ - xorg-libxext=1.3.4=h0b41bf4_2
381
+ - xorg-libxrender=0.9.11=hd590300_0
382
+ - xorg-libxt=1.3.0=hd590300_1
383
+ - xorg-renderproto=0.11.1=h7f98852_1002
384
+ - xorg-xextproto=7.3.0=h0b41bf4_1003
385
+ - xorg-xproto=7.0.31=h7f98852_1007
386
+ - xyzservices=2023.10.1=pyhd8ed1ab_0
387
+ - xz=5.2.6=h166bdaf_0
388
+ - yaml=0.2.5=h7b6447c_0
389
+ - yaml-cpp=0.7.0=h59595ed_3
390
+ - zeromq=4.3.5=h59595ed_0
391
+ - zict=3.0.0=pyhd8ed1ab_0
392
+ - zipp=3.17.0=pyhd8ed1ab_0
393
+ - zlib=1.2.13=hd590300_5
394
+ - zstandard=0.22.0=py310h1275a96_0
395
+ - zstd=1.5.5=hfc55251_0
396
+ - pip:
397
+ - absl-py==1.4.0
398
+ - accelerate==0.28.0
399
+ - access==1.1.9
400
+ - affine==2.4.0
401
+ - aiobotocore==2.12.1
402
+ - aiofiles==22.1.0
403
+ - aiohttp-cors==0.7.0
404
+ - aioitertools==0.11.0
405
+ - aiorwlock==1.3.0
406
+ - aiosqlite==0.19.0
407
+ - albumentations==1.4.0
408
+ - alembic==1.13.1
409
+ - altair==5.2.0
410
+ - annoy==1.17.3
411
+ - apache-beam==2.46.0
412
+ - aplus==0.11.0
413
+ - appdirs==1.4.4
414
+ - array-record==0.5.0
415
+ - arviz==0.17.1
416
+ - astroid==3.0.3
417
+ - astropy==6.0.0
418
+ - astropy-iers-data==0.2024.3.18.0.29.47
419
+ - astunparse==1.6.3
420
+ - async-lru==2.0.4
421
+ - audioread==3.0.1
422
+ - auto-gptq==0.7.1
423
+ - autopep8==2.0.4
424
+ - babel==2.14.0
425
+ - backoff==2.2.1
426
+ - bayesian-optimization==1.4.3
427
+ - beatrix-jupyterlab==2023.128.151533
428
+ - bidict==0.23.1
429
+ - bitsandbytes==0.43.1
430
+ - blake3==0.2.1
431
+ - blessed==1.20.0
432
+ - blinker==1.7.0
433
+ - blosc2==2.5.1
434
+ - bokeh==3.3.4
435
+ - boruta==0.3
436
+ - boto3==1.26.100
437
+ - botocore==1.34.51
438
+ - bqplot==0.12.43
439
+ - branca==0.7.1
440
+ - brewer2mpl==1.4.1
441
+ - cachetools==4.2.4
442
+ - catalyst==22.4
443
+ - catboost==1.2.3
444
+ - category-encoders==2.6.3
445
+ - cesium==0.12.1
446
+ - chex==0.1.85
447
+ - cleverhans==4.0.0
448
+ - click-plugins==1.1.1
449
+ - cligj==0.7.2
450
+ - cloud-tpu-client==0.10
451
+ - cloud-tpu-profiler==2.4.0
452
+ - cloudpickle==2.2.1
453
+ - cmdstanpy==1.2.1
454
+ - cmudict==1.0.21
455
+ - colorcet==3.1.0
456
+ - coloredlogs==15.0.1
457
+ - colorful==0.5.6
458
+ - colorlog==6.8.2
459
+ - colorlover==0.3.0
460
+ - contextily==1.5.2
461
+ - convertdate==2.4.0
462
+ - crcmod==1.7
463
+ - cufflinks==0.17.3
464
+ - cvxcanon==0.1.2
465
+ - cython==3.0.8
466
+ - daal==2024.1.0
467
+ - daal4py==2024.1.0
468
+ - dacite==1.8.1
469
+ - dask==2024.3.1
470
+ - dask-expr==1.0.4
471
+ - dataclasses-json==0.6.4
472
+ - dataproc-jupyter-plugin==0.1.66
473
+ - datasets==2.1.0
474
+ - datashader==0.16.0
475
+ - datatile==1.0.3
476
+ - db-dtypes==1.2.0
477
+ - deap==1.4.1
478
+ - deepdiff==6.7.1
479
+ - deprecated==1.2.14
480
+ - deprecation==2.1.0
481
+ - descartes==1.1.0
482
+ - dill==0.3.8
483
+ - dipy==1.9.0
484
+ - distlib==0.3.8
485
+ - dm-tree==0.1.8
486
+ - docker==7.0.0
487
+ - docker-pycreds==0.4.0
488
+ - docopt==0.6.2
489
+ - docstring-parser==0.15
490
+ - docstring-to-markdown==0.15
491
+ - docutils==0.20.1
492
+ - earthengine-api==0.1.394
493
+ - easydict==1.13
494
+ - easyocr==1.7.1
495
+ - ecos==2.0.13
496
+ - eli5==0.13.0
497
+ - emoji==2.10.1
498
+ - en-core-web-lg==3.7.1
499
+ - en-core-web-sm==3.7.1
500
+ - ephem==4.1.5
501
+ - esda==2.5.1
502
+ - essentia==2.1b6.dev1110
503
+ - et-xmlfile==1.1.0
504
+ - etils==1.6.0
505
+ - explainable-ai-sdk==1.3.3
506
+ - farama-notifications==0.0.4
507
+ - fastai==2.7.14
508
+ - fastapi==0.108.0
509
+ - fastavro==1.9.3
510
+ - fastcore==1.5.29
511
+ - fastdownload==0.0.7
512
+ - fasteners==0.19
513
+ - fastprogress==1.0.3
514
+ - fasttext==0.9.2
515
+ - feather-format==0.4.1
516
+ - featuretools==1.30.0
517
+ - filelock==3.13.1
518
+ - fiona==1.9.6
519
+ - fitter==1.7.0
520
+ - flake8==7.0.0
521
+ - flashtext==2.7
522
+ - flask==3.0.2
523
+ - flatbuffers==23.5.26
524
+ - flax==0.8.2
525
+ - folium==0.16.0
526
+ - fonttools==4.47.0
527
+ - frozendict==2.4.0
528
+ - funcy==2.0
529
+ - fury==0.10.0
530
+ - future==1.0.0
531
+ - fuzzywuzzy==0.18.0
532
+ - gast==0.5.4
533
+ - gatspy==0.3
534
+ - gcsfs==2023.12.2.post1
535
+ - gekko==1.1.1
536
+ - gensim==4.3.2
537
+ - geographiclib==2.0
538
+ - geohash==1.0
539
+ - geojson==3.1.0
540
+ - geopandas==0.14.3
541
+ - geoplot==0.5.1
542
+ - geopy==2.4.1
543
+ - geoviews==1.11.1
544
+ - ggplot==0.11.5
545
+ - giddy==2.3.5
546
+ - gitdb==4.0.11
547
+ - gitpython==3.1.41
548
+ - google-ai-generativelanguage==0.4.0
549
+ - google-api-core==2.17.1
550
+ - google-api-python-client==2.122.0
551
+ - google-apitools==0.5.31
552
+ - google-auth-httplib2==0.1.1
553
+ - google-auth-oauthlib==1.2.0
554
+ - google-cloud-aiplatform==0.6.0a1
555
+ - google-cloud-artifact-registry==1.10.0
556
+ - google-cloud-automl==1.0.1
557
+ - google-cloud-bigquery==2.34.4
558
+ - google-cloud-bigtable==1.7.3
559
+ - google-cloud-dlp==3.14.0
560
+ - google-cloud-jupyter-config==0.0.5
561
+ - google-cloud-language==2.13.3
562
+ - google-cloud-monitoring==2.18.0
563
+ - google-cloud-pubsub==2.19.0
564
+ - google-cloud-pubsublite==1.9.0
565
+ - google-cloud-recommendations-ai==0.7.1
566
+ - google-cloud-resource-manager==1.11.0
567
+ - google-cloud-spanner==3.40.1
568
+ - google-cloud-storage==1.44.0
569
+ - google-cloud-translate==3.12.1
570
+ - google-cloud-videointelligence==2.13.3
571
+ - google-cloud-vision==2.8.0
572
+ - google-crc32c==1.5.0
573
+ - google-generativeai==0.4.1
574
+ - google-pasta==0.2.0
575
+ - google-resumable-media==2.7.0
576
+ - gplearn==0.4.2
577
+ - gpustat==1.0.0
578
+ - gpxpy==1.6.2
579
+ - greenlet==3.0.3
580
+ - grpc-google-iam-v1==0.12.7
581
+ - grpcio==1.60.0
582
+ - grpcio-status==1.48.2
583
+ - gviz-api==1.10.0
584
+ - gym==0.26.2
585
+ - gym-notices==0.0.8
586
+ - gymnasium==0.29.0
587
+ - h11==0.14.0
588
+ - h2o==3.46.0.1
589
+ - h5netcdf==1.3.0
590
+ - h5py==3.10.0
591
+ - haversine==2.8.1
592
+ - hdfs==2.7.3
593
+ - hep-ml==0.7.2
594
+ - hijri-converter==2.3.1
595
+ - hmmlearn==0.3.2
596
+ - holidays==0.24
597
+ - holoviews==1.18.3
598
+ - hpsklearn==0.1.0
599
+ - html5lib==1.1
600
+ - htmlmin==0.1.12
601
+ - httpcore==1.0.4
602
+ - httplib2==0.21.0
603
+ - httptools==0.6.1
604
+ - httpx==0.27.0
605
+ - huggingface-hub==0.21.4
606
+ - humanfriendly==10.0
607
+ - hunspell==0.5.5
608
+ - husl==4.0.3
609
+ - hydra-slayer==0.5.0
610
+ - hyperopt==0.2.7
611
+ - hypertools==0.8.0
612
+ - igraph==0.11.4
613
+ - imagecodecs==2024.1.1
614
+ - imagehash==4.3.1
615
+ - imageio==2.33.1
616
+ - imbalanced-learn==0.12.0
617
+ - imgaug==0.4.0
618
+ - importlib-metadata==6.11.0
619
+ - inequality==1.0.1
620
+ - iniconfig==2.0.0
621
+ - ipydatawidgets==4.3.5
622
+ - ipyleaflet==0.18.2
623
+ - ipympl==0.7.0
624
+ - ipython-genutils==0.2.0
625
+ - ipython-sql==0.5.0
626
+ - ipyvolume==0.6.3
627
+ - ipyvue==1.10.2
628
+ - ipyvuetify==1.9.2
629
+ - ipywebrtc==0.6.0
630
+ - ipywidgets==7.7.1
631
+ - isort==5.13.2
632
+ - isoweek==1.3.3
633
+ - itsdangerous==2.1.2
634
+ - janome==0.5.0
635
+ - jaraco-classes==3.3.0
636
+ - jax==0.4.23
637
+ - jax-jumpy==1.0.0
638
+ - jaxlib==0.4.23.dev20240116
639
+ - jeepney==0.8.0
640
+ - jieba==0.42.1
641
+ - jmespath==1.0.1
642
+ - json5==0.9.14
643
+ - jupyter-client==7.4.9
644
+ - jupyter-console==6.6.3
645
+ - jupyter-http-over-ws==0.0.8
646
+ - jupyter-lsp==1.5.1
647
+ - jupyter-server==2.13.0
648
+ - jupyter-server-fileid==0.9.1
649
+ - jupyter-server-mathjax==0.2.6
650
+ - jupyter-server-proxy==4.1.0
651
+ - jupyter-server-ydoc==0.8.0
652
+ - jupyter-ydoc==0.2.5
653
+ - jupyterlab==4.1.5
654
+ - jupyterlab-git==0.44.0
655
+ - jupyterlab-lsp==5.1.0
656
+ - jupyterlab-server==2.25.2
657
+ - jupyterlab-widgets==3.0.9
658
+ - jupytext==1.16.0
659
+ - kaggle==1.6.6
660
+ - kaggle-environments==1.14.3
661
+ - kagglehub==0.2.0
662
+ - keras==3.0.5
663
+ - keras-cv==0.8.2
664
+ - keras-nlp==0.8.2
665
+ - keras-tuner==1.4.6
666
+ - kernels-mixer==0.0.7
667
+ - keyring==24.3.0
668
+ - keyrings-google-artifactregistry-auth==1.1.2
669
+ - kfp==2.5.0
670
+ - kfp-pipeline-spec==0.2.2
671
+ - kfp-server-api==2.0.5
672
+ - kmapper==2.0.1
673
+ - kmodes==0.12.2
674
+ - korean-lunar-calendar==0.3.1
675
+ - kornia==0.7.2
676
+ - kornia-rs==0.1.2
677
+ - kt-legacy==1.0.5
678
+ - kubernetes==26.1.0
679
+ - langid==1.1.6
680
+ - lazy-loader==0.3
681
+ - learntools==0.3.4
682
+ - leven==1.0.4
683
+ - levenshtein==0.25.0
684
+ - libclang==16.0.6
685
+ - libpysal==4.9.2
686
+ - librosa==0.10.1
687
+ - lightgbm==4.2.0
688
+ - lightning-utilities==0.10.1
689
+ - lime==0.2.0.1
690
+ - line-profiler==4.1.2
691
+ - linkify-it-py==2.0.3
692
+ - llvmlite==0.41.1
693
+ - lml==0.1.0
694
+ - loguru==0.7.2
695
+ - lunarcalendar==0.0.9
696
+ - lxml==5.1.0
697
+ - mako==1.3.2
698
+ - mapclassify==2.6.1
699
+ - markdown==3.5.2
700
+ - markovify==0.9.4
701
+ - markupsafe==2.1.5
702
+ - marshmallow==3.21.1
703
+ - matplotlib==3.7.5
704
+ - matplotlib-venn==0.11.10
705
+ - mccabe==0.7.0
706
+ - mdit-py-plugins==0.4.0
707
+ - memory-profiler==0.61.0
708
+ - mercantile==1.2.1
709
+ - mgwr==2.2.1
710
+ - missingno==0.5.2
711
+ - mistune==0.8.4
712
+ - mizani==0.11.0
713
+ - ml-dtypes==0.2.0
714
+ - mlcrate==0.2.0
715
+ - mlens==0.2.3
716
+ - mlxtend==0.23.1
717
+ - mmh3==4.1.0
718
+ - mne==1.6.1
719
+ - mnist==0.2.2
720
+ - mock==5.1.0
721
+ - momepy==0.7.0
722
+ - more-itertools==10.2.0
723
+ - mpld3==0.5.10
724
+ - mpmath==1.3.0
725
+ - msgpack-numpy==0.4.8
726
+ - multimethod==1.10
727
+ - multipledispatch==1.0.0
728
+ - multiprocess==0.70.16
729
+ - mypy-extensions==1.0.0
730
+ - namex==0.0.7
731
+ - nbclient==0.5.13
732
+ - nbconvert==6.4.5
733
+ - nbdime==3.2.0
734
+ - ndindex==1.8
735
+ - networkx==3.2.1
736
+ - nibabel==5.2.1
737
+ - nilearn==0.10.3
738
+ - ninja==1.11.1.1
739
+ - nltk==3.2.4
740
+ - nose==1.3.7
741
+ - notebook==6.5.6
742
+ - notebook-executor==0.2
743
+ - numba==0.58.1
744
+ - numexpr==2.9.0
745
+ - nvidia-ml-py==11.495.46
746
+ - oauth2client==4.1.3
747
+ - oauthlib==3.2.2
748
+ - objsize==0.6.1
749
+ - odfpy==1.4.1
750
+ - olefile==0.47
751
+ - onnx==1.15.0
752
+ - opencensus==0.11.4
753
+ - opencensus-context==0.1.3
754
+ - opencv-contrib-python==4.9.0.80
755
+ - opencv-python==4.9.0.80
756
+ - opencv-python-headless==4.9.0.80
757
+ - openpyxl==3.1.2
758
+ - openslide-python==1.3.1
759
+ - opentelemetry-api==1.22.0
760
+ - opentelemetry-exporter-otlp==1.22.0
761
+ - opentelemetry-exporter-otlp-proto-common==1.22.0
762
+ - opentelemetry-exporter-otlp-proto-grpc==1.22.0
763
+ - opentelemetry-exporter-otlp-proto-http==1.22.0
764
+ - opentelemetry-proto==1.22.0
765
+ - opentelemetry-sdk==1.22.0
766
+ - opentelemetry-semantic-conventions==0.43b0
767
+ - opt-einsum==3.3.0
768
+ - optax==0.2.1
769
+ - optimum==1.19.1
770
+ - optuna==3.6.0
771
+ - orbax-checkpoint==0.5.6
772
+ - ordered-set==4.1.0
773
+ - orjson==3.9.10
774
+ - ortools==9.4.1874
775
+ - osmnx==1.9.1
776
+ - packaging==21.3
777
+ - pandas==2.1.4
778
+ - pandas-datareader==0.10.0
779
+ - pandas-profiling==3.6.6
780
+ - pandas-summary==0.2.0
781
+ - pandasql==0.7.3
782
+ - panel==1.3.8
783
+ - papermill==2.5.0
784
+ - param==2.0.2
785
+ - path==16.10.0
786
+ - path-py==12.5.0
787
+ - pathos==0.3.2
788
+ - patsy==0.5.6
789
+ - pdf2image==1.17.0
790
+ - peft==0.10.0
791
+ - pettingzoo==1.24.0
792
+ - pexpect==4.9.0
793
+ - phik==0.12.4
794
+ - pillow==9.5.0
795
+ - platformdirs==4.2.0
796
+ - plotly==5.18.0
797
+ - plotly-express==0.4.1
798
+ - plotnine==0.13.2
799
+ - pluggy==1.4.0
800
+ - pointpats==2.4.0
801
+ - polars==0.20.15
802
+ - polyglot==16.7.4
803
+ - pooch==1.8.1
804
+ - pox==0.3.4
805
+ - ppca==0.0.4
806
+ - ppft==1.7.6.8
807
+ - preprocessing==0.1.13
808
+ - prettytable==3.9.0
809
+ - progressbar2==4.4.2
810
+ - promise==2.3
811
+ - prompt-toolkit==3.0.43
812
+ - pronouncing==0.2.0
813
+ - prophet==1.1.1
814
+ - protobuf==3.20.3
815
+ - psutil==5.9.3
816
+ - pudb==2024.1
817
+ - pulp==2.8.0
818
+ - py-cpuinfo==9.0.0
819
+ - py-spy==0.3.14
820
+ - py4j==0.10.9.7
821
+ - pyaml==23.12.0
822
+ - pyarabic==0.6.15
823
+ - pyastronomy==0.21.0
824
+ - pybind11==2.11.1
825
+ - pyclipper==1.3.0.post5
826
+ - pycodestyle==2.11.1
827
+ - pycparser==2.21
828
+ - pycryptodome==3.20.0
829
+ - pyct==0.5.0
830
+ - pycuda==2024.1
831
+ - pydantic==2.5.3
832
+ - pydantic-core==2.14.6
833
+ - pydegensac==0.1.2
834
+ - pydicom==2.4.4
835
+ - pydocstyle==6.3.0
836
+ - pydot==1.4.2
837
+ - pydub==0.25.1
838
+ - pyemd==1.0.0
839
+ - pyerfa==2.0.1.1
840
+ - pyexcel-io==0.6.6
841
+ - pyexcel-ods==0.6.0
842
+ - pyflakes==3.2.0
843
+ - pygltflib==1.16.2
844
+ - pyjwt==2.8.0
845
+ - pykalman==0.9.5
846
+ - pyldavis==3.4.1
847
+ - pylint==3.0.4
848
+ - pymc3==3.11.4
849
+ - pymeeus==0.5.12
850
+ - pymongo==3.13.0
851
+ - pympler==1.0.1
852
+ - pynndescent==0.5.11
853
+ - pynvrtc==9.2
854
+ - pyocr==0.8.5
855
+ - pyparsing==3.1.1
856
+ - pypdf==4.1.0
857
+ - pysal==24.1
858
+ - pytesseract==0.3.10
859
+ - pytest==8.1.1
860
+ - python-bidi==0.4.2
861
+ - python-dateutil==2.9.0.post0
862
+ - python-dotenv==1.0.0
863
+ - python-graphviz==0.20.2
864
+ - python-levenshtein==0.25.0
865
+ - python-louvain==0.16
866
+ - python-lsp-jsonrpc==1.1.2
867
+ - python-lsp-server==1.10.1
868
+ - python-slugify==8.0.4
869
+ - python-utils==3.8.2
870
+ - pythreejs==2.4.2
871
+ - pytoolconfig==1.3.1
872
+ - pytools==2023.1.1
873
+ - pytorch-ignite==0.4.13
874
+ - pytorch-lightning==2.2.1
875
+ - pytz==2023.3.post1
876
+ - pyupset==0.1.1.post7
877
+ - pyviz-comms==3.0.1
878
+ - pywavelets==1.5.0
879
+ - pyzmq==24.0.1
880
+ - qgrid==1.3.1
881
+ - qtconsole==5.5.1
882
+ - qtpy==2.4.1
883
+ - quantecon==0.7.2
884
+ - quantities==0.15.0
885
+ - qudida==0.0.4
886
+ - rapidfuzz==3.6.2
887
+ - rasterio==1.3.9
888
+ - rasterstats==0.19.0
889
+ - ray==2.9.0
890
+ - ray-cpp==2.9.0
891
+ - regex==2023.12.25
892
+ - requests-oauthlib==1.3.1
893
+ - requests-toolbelt==0.10.1
894
+ - responses==0.18.0
895
+ - retrying==1.3.4
896
+ - rgf-python==3.12.0
897
+ - rich==13.7.0
898
+ - rich-click==1.7.4
899
+ - rope==1.12.0
900
+ - rouge==1.0.1
901
+ - rtree==1.2.0
902
+ - s2sphere==0.2.5
903
+ - s3fs==2024.3.0
904
+ - s3transfer==0.6.2
905
+ - safetensors==0.4.2
906
+ - scattertext==0.1.19
907
+ - scikit-image==0.22.0
908
+ - scikit-learn==1.2.2
909
+ - scikit-learn-intelex==2024.1.0
910
+ - scikit-multilearn==0.2.0
911
+ - scikit-optimize==0.10.1
912
+ - scikit-plot==0.3.7
913
+ - scikit-surprise==1.1.3
914
+ - scipy==1.11.4
915
+ - seaborn==0.12.2
916
+ - secretstorage==3.3.3
917
+ - segment-anything==1.0
918
+ - segregation==2.5
919
+ - semver==3.0.2
920
+ - sentencepiece==0.2.0
921
+ - sentry-sdk==1.42.0
922
+ - setproctitle==1.3.3
923
+ - setuptools-git==1.2
924
+ - setuptools-scm==8.0.4
925
+ - shap==0.44.1
926
+ - shapely==2.0.3
927
+ - shimmy==1.3.0
928
+ - simpervisor==1.0.0
929
+ - simpleitk==2.3.1
930
+ - simplejson==3.19.2
931
+ - six==1.16.0
932
+ - sklearn-pandas==2.2.0
933
+ - slicer==0.0.7
934
+ - smmap==5.0.1
935
+ - snowballstemmer==2.2.0
936
+ - snuggs==1.4.7
937
+ - soundfile==0.12.1
938
+ - soxr==0.3.7
939
+ - spaghetti==1.7.5.post1
940
+ - spectral==0.23.1
941
+ - spglm==1.1.0
942
+ - sphinx-rtd-theme==0.2.4
943
+ - spint==1.0.7
944
+ - splot==1.1.5.post1
945
+ - spopt==0.6.0
946
+ - spreg==1.4.2
947
+ - spvcm==0.3.0
948
+ - sqlalchemy==2.0.25
949
+ - sqlparse==0.4.4
950
+ - squarify==0.4.3
951
+ - stable-baselines3==2.1.0
952
+ - stack-data==0.6.3
953
+ - stanio==0.3.0
954
+ - starlette==0.32.0.post1
955
+ - statsmodels==0.14.1
956
+ - stemming==1.0.1
957
+ - stop-words==2018.7.23
958
+ - stopit==1.1.2
959
+ - stumpy==1.12.0
960
+ - sympy==1.12
961
+ - tables==3.9.2
962
+ - tabulate==0.9.0
963
+ - tangled-up-in-unicode==0.2.0
964
+ - tbb==2021.11.0
965
+ - tenacity==8.2.3
966
+ - tensorboard==2.15.1
967
+ - tensorboard-data-server==0.7.2
968
+ - tensorboard-plugin-profile==2.15.0
969
+ - tensorboardx==2.6.2.2
970
+ - tensorflow==2.15.0
971
+ - tensorflow-cloud==0.1.16
972
+ - tensorflow-datasets==4.9.4
973
+ - tensorflow-decision-forests==1.8.1
974
+ - tensorflow-estimator==2.15.0
975
+ - tensorflow-hub==0.16.1
976
+ - tensorflow-io==0.35.0
977
+ - tensorflow-io-gcs-filesystem==0.35.0
978
+ - tensorflow-metadata==0.14.0
979
+ - tensorflow-probability==0.23.0
980
+ - tensorflow-serving-api==2.14.1
981
+ - tensorflow-text==2.15.0
982
+ - tensorflow-transform==0.14.0
983
+ - tensorpack==0.11
984
+ - tensorstore==0.1.56
985
+ - termcolor==2.4.0
986
+ - testpath==0.6.0
987
+ - text-unidecode==1.3
988
+ - textblob==0.18.0.post0
989
+ - texttable==1.7.0
990
+ - tf-keras==2.15.1
991
+ - tfp-nightly==0.24.0.dev0
992
+ - theano==1.0.5
993
+ - theano-pymc==1.1.2
994
+ - threadpoolctl==3.2.0
995
+ - tifffile==2023.12.9
996
+ - timm==0.9.16
997
+ - tobler==0.11.2
998
+ - tokenizers==0.15.2
999
+ - toml==0.10.2
1000
+ - tomli==2.0.1
1001
+ - tomlkit==0.12.4
1002
+ - torch==2.1.2
1003
+ - torchaudio==2.1.2
1004
+ - torchdata==0.7.1
1005
+ - torchinfo==1.8.0
1006
+ - torchmetrics==1.3.2
1007
+ - torchtext==0.16.2
1008
+ - torchvision==0.16.2
1009
+ - tpot==0.12.1
1010
+ - traceml==1.0.8
1011
+ - traittypes==0.2.1
1012
+ - transformers==4.38.2
1013
+ - treelite-runtime==3.2.0
1014
+ - trueskill==0.4.5
1015
+ - trx-python==0.2.9
1016
+ - tsfresh==0.20.2
1017
+ - typeguard==4.1.5
1018
+ - typing-inspect==0.9.0
1019
+ - tzdata==2023.4
1020
+ - uc-micro-py==1.0.3
1021
+ - ujson==5.9.0
1022
+ - umap-learn==0.5.5
1023
+ - unidecode==1.3.8
1024
+ - update-checker==0.18.0
1025
+ - uritemplate==3.0.1
1026
+ - urllib3==1.26.18
1027
+ - urwid==2.6.9
1028
+ - urwid-readline==0.14
1029
+ - uvicorn==0.25.0
1030
+ - uvloop==0.19.0
1031
+ - vaex==4.17.0
1032
+ - vaex-astro==0.9.3
1033
+ - vaex-core==4.17.1
1034
+ - vaex-hdf5==0.14.1
1035
+ - vaex-jupyter==0.8.2
1036
+ - vaex-ml==0.18.3
1037
+ - vaex-server==0.9.0
1038
+ - vaex-viz==0.5.4
1039
+ - vec-noise==1.1.4
1040
+ - vecstack==0.4.0
1041
+ - virtualenv==20.21.0
1042
+ - visions==0.7.5
1043
+ - vowpalwabbit==9.9.0
1044
+ - vtk==9.3.0
1045
+ - wand==0.6.13
1046
+ - wandb==0.16.4
1047
+ - watchfiles==0.21.0
1048
+ - wavio==0.0.8
1049
+ - websockets==12.0
1050
+ - werkzeug==3.0.1
1051
+ - wfdb==4.1.2
1052
+ - whatthepatch==1.0.5
1053
+ - widgetsnbextension==3.6.6
1054
+ - witwidget==1.8.1
1055
+ - woodwork==0.29.0
1056
+ - wordcloud==1.9.3
1057
+ - wordsegment==1.3.1
1058
+ - wrapt==1.14.1
1059
+ - xarray==2024.2.0
1060
+ - xarray-einstats==0.7.0
1061
+ - xgboost==2.0.3
1062
+ - xvfbwrapper==0.2.9
1063
+ - xxhash==3.4.1
1064
+ - y-py==0.6.2
1065
+ - yapf==0.40.2
1066
+ - yarl==1.9.4
1067
+ - ydata-profiling==4.6.4
1068
+ - yellowbrick==1.5
1069
+ - ypy-websocket==0.8.4
1070
+ prefix: /opt/conda
wandb/run-20240429_072705-0t7z1bit/files/config.yaml ADDED
@@ -0,0 +1,722 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.4
8
+ framework: huggingface
9
+ huggingface_version: 4.38.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1714375625.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 105
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 98
41
+ - 105
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.13
46
+ 5: 0.16.4
47
+ 6: 4.38.2
48
+ 8:
49
+ - 1
50
+ - 2
51
+ - 5
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: train/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: train/grad_norm
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: train/learning_rate
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ - 1: train/epoch
72
+ 5: 1
73
+ 6:
74
+ - 1
75
+ - 1: eval/loss
76
+ 5: 1
77
+ 6:
78
+ - 1
79
+ - 1: eval/Accuracy
80
+ 5: 1
81
+ 6:
82
+ - 1
83
+ - 1: eval/F1
84
+ 5: 1
85
+ 6:
86
+ - 1
87
+ - 1: eval/Precision
88
+ 5: 1
89
+ 6:
90
+ - 1
91
+ - 1: eval/Recall
92
+ 5: 1
93
+ 6:
94
+ - 1
95
+ - 1: eval/runtime
96
+ 5: 1
97
+ 6:
98
+ - 1
99
+ - 1: eval/samples_per_second
100
+ 5: 1
101
+ 6:
102
+ - 1
103
+ - 1: eval/steps_per_second
104
+ 5: 1
105
+ 6:
106
+ - 1
107
+ - 1: train/train_runtime
108
+ 5: 1
109
+ 6:
110
+ - 1
111
+ - 1: train/train_samples_per_second
112
+ 5: 1
113
+ 6:
114
+ - 1
115
+ - 1: train/train_steps_per_second
116
+ 5: 1
117
+ 6:
118
+ - 1
119
+ - 1: train/total_flos
120
+ 5: 1
121
+ 6:
122
+ - 1
123
+ - 1: train/train_loss
124
+ 5: 1
125
+ 6:
126
+ - 1
127
+ return_dict:
128
+ desc: null
129
+ value: true
130
+ output_hidden_states:
131
+ desc: null
132
+ value: false
133
+ output_attentions:
134
+ desc: null
135
+ value: false
136
+ torchscript:
137
+ desc: null
138
+ value: false
139
+ torch_dtype:
140
+ desc: null
141
+ value: null
142
+ use_bfloat16:
143
+ desc: null
144
+ value: false
145
+ tf_legacy_loss:
146
+ desc: null
147
+ value: false
148
+ pruned_heads:
149
+ desc: null
150
+ value: {}
151
+ tie_word_embeddings:
152
+ desc: null
153
+ value: true
154
+ chunk_size_feed_forward:
155
+ desc: null
156
+ value: 0
157
+ is_encoder_decoder:
158
+ desc: null
159
+ value: false
160
+ is_decoder:
161
+ desc: null
162
+ value: false
163
+ cross_attention_hidden_size:
164
+ desc: null
165
+ value: null
166
+ add_cross_attention:
167
+ desc: null
168
+ value: false
169
+ tie_encoder_decoder:
170
+ desc: null
171
+ value: false
172
+ max_length:
173
+ desc: null
174
+ value: 20
175
+ min_length:
176
+ desc: null
177
+ value: 0
178
+ do_sample:
179
+ desc: null
180
+ value: false
181
+ early_stopping:
182
+ desc: null
183
+ value: false
184
+ num_beams:
185
+ desc: null
186
+ value: 1
187
+ num_beam_groups:
188
+ desc: null
189
+ value: 1
190
+ diversity_penalty:
191
+ desc: null
192
+ value: 0.0
193
+ temperature:
194
+ desc: null
195
+ value: 1.0
196
+ top_k:
197
+ desc: null
198
+ value: 50
199
+ top_p:
200
+ desc: null
201
+ value: 1.0
202
+ typical_p:
203
+ desc: null
204
+ value: 1.0
205
+ repetition_penalty:
206
+ desc: null
207
+ value: 1.0
208
+ length_penalty:
209
+ desc: null
210
+ value: 1.0
211
+ no_repeat_ngram_size:
212
+ desc: null
213
+ value: 0
214
+ encoder_no_repeat_ngram_size:
215
+ desc: null
216
+ value: 0
217
+ bad_words_ids:
218
+ desc: null
219
+ value: null
220
+ num_return_sequences:
221
+ desc: null
222
+ value: 1
223
+ output_scores:
224
+ desc: null
225
+ value: false
226
+ return_dict_in_generate:
227
+ desc: null
228
+ value: false
229
+ forced_bos_token_id:
230
+ desc: null
231
+ value: null
232
+ forced_eos_token_id:
233
+ desc: null
234
+ value: null
235
+ remove_invalid_values:
236
+ desc: null
237
+ value: false
238
+ exponential_decay_length_penalty:
239
+ desc: null
240
+ value: null
241
+ suppress_tokens:
242
+ desc: null
243
+ value: null
244
+ begin_suppress_tokens:
245
+ desc: null
246
+ value: null
247
+ architectures:
248
+ desc: null
249
+ value:
250
+ - BertForMaskedLM
251
+ finetuning_task:
252
+ desc: null
253
+ value: null
254
+ id2label:
255
+ desc: null
256
+ value:
257
+ '0': Install
258
+ '1': Compensation
259
+ '2': Change
260
+ '3': Summarize
261
+ '4': Terminate
262
+ '5': SBC
263
+ label2id:
264
+ desc: null
265
+ value:
266
+ Install: 0
267
+ Compensation: 1
268
+ Change: 2
269
+ Summarize: 3
270
+ Terminate: 4
271
+ SBC: 5
272
+ tokenizer_class:
273
+ desc: null
274
+ value: null
275
+ prefix:
276
+ desc: null
277
+ value: null
278
+ bos_token_id:
279
+ desc: null
280
+ value: null
281
+ pad_token_id:
282
+ desc: null
283
+ value: 0
284
+ eos_token_id:
285
+ desc: null
286
+ value: null
287
+ sep_token_id:
288
+ desc: null
289
+ value: null
290
+ decoder_start_token_id:
291
+ desc: null
292
+ value: null
293
+ task_specific_params:
294
+ desc: null
295
+ value: null
296
+ problem_type:
297
+ desc: null
298
+ value: null
299
+ _name_or_path:
300
+ desc: null
301
+ value: google-bert/bert-base-uncased
302
+ transformers_version:
303
+ desc: null
304
+ value: 4.38.2
305
+ gradient_checkpointing:
306
+ desc: null
307
+ value: false
308
+ model_type:
309
+ desc: null
310
+ value: bert
311
+ vocab_size:
312
+ desc: null
313
+ value: 30522
314
+ hidden_size:
315
+ desc: null
316
+ value: 768
317
+ num_hidden_layers:
318
+ desc: null
319
+ value: 12
320
+ num_attention_heads:
321
+ desc: null
322
+ value: 12
323
+ hidden_act:
324
+ desc: null
325
+ value: gelu
326
+ intermediate_size:
327
+ desc: null
328
+ value: 3072
329
+ hidden_dropout_prob:
330
+ desc: null
331
+ value: 0.1
332
+ attention_probs_dropout_prob:
333
+ desc: null
334
+ value: 0.1
335
+ max_position_embeddings:
336
+ desc: null
337
+ value: 512
338
+ type_vocab_size:
339
+ desc: null
340
+ value: 2
341
+ initializer_range:
342
+ desc: null
343
+ value: 0.02
344
+ layer_norm_eps:
345
+ desc: null
346
+ value: 1.0e-12
347
+ position_embedding_type:
348
+ desc: null
349
+ value: absolute
350
+ use_cache:
351
+ desc: null
352
+ value: true
353
+ classifier_dropout:
354
+ desc: null
355
+ value: null
356
+ output_dir:
357
+ desc: null
358
+ value: /kaggle/working/
359
+ overwrite_output_dir:
360
+ desc: null
361
+ value: false
362
+ do_train:
363
+ desc: null
364
+ value: true
365
+ do_eval:
366
+ desc: null
367
+ value: true
368
+ do_predict:
369
+ desc: null
370
+ value: false
371
+ evaluation_strategy:
372
+ desc: null
373
+ value: steps
374
+ prediction_loss_only:
375
+ desc: null
376
+ value: false
377
+ per_device_train_batch_size:
378
+ desc: null
379
+ value: 16
380
+ per_device_eval_batch_size:
381
+ desc: null
382
+ value: 16
383
+ per_gpu_train_batch_size:
384
+ desc: null
385
+ value: null
386
+ per_gpu_eval_batch_size:
387
+ desc: null
388
+ value: null
389
+ gradient_accumulation_steps:
390
+ desc: null
391
+ value: 1
392
+ eval_accumulation_steps:
393
+ desc: null
394
+ value: null
395
+ eval_delay:
396
+ desc: null
397
+ value: 0
398
+ learning_rate:
399
+ desc: null
400
+ value: 5.0e-05
401
+ weight_decay:
402
+ desc: null
403
+ value: 0.01
404
+ adam_beta1:
405
+ desc: null
406
+ value: 0.9
407
+ adam_beta2:
408
+ desc: null
409
+ value: 0.999
410
+ adam_epsilon:
411
+ desc: null
412
+ value: 1.0e-08
413
+ max_grad_norm:
414
+ desc: null
415
+ value: 1.0
416
+ num_train_epochs:
417
+ desc: null
418
+ value: 10
419
+ max_steps:
420
+ desc: null
421
+ value: -1
422
+ lr_scheduler_type:
423
+ desc: null
424
+ value: linear
425
+ lr_scheduler_kwargs:
426
+ desc: null
427
+ value: {}
428
+ warmup_ratio:
429
+ desc: null
430
+ value: 0.0
431
+ warmup_steps:
432
+ desc: null
433
+ value: 10
434
+ log_level:
435
+ desc: null
436
+ value: passive
437
+ log_level_replica:
438
+ desc: null
439
+ value: warning
440
+ log_on_each_node:
441
+ desc: null
442
+ value: true
443
+ logging_dir:
444
+ desc: null
445
+ value: /kaggle/working/
446
+ logging_strategy:
447
+ desc: null
448
+ value: steps
449
+ logging_first_step:
450
+ desc: null
451
+ value: false
452
+ logging_steps:
453
+ desc: null
454
+ value: 10
455
+ logging_nan_inf_filter:
456
+ desc: null
457
+ value: true
458
+ save_strategy:
459
+ desc: null
460
+ value: steps
461
+ save_steps:
462
+ desc: null
463
+ value: 500
464
+ save_total_limit:
465
+ desc: null
466
+ value: null
467
+ save_safetensors:
468
+ desc: null
469
+ value: true
470
+ save_on_each_node:
471
+ desc: null
472
+ value: false
473
+ save_only_model:
474
+ desc: null
475
+ value: false
476
+ no_cuda:
477
+ desc: null
478
+ value: false
479
+ use_cpu:
480
+ desc: null
481
+ value: false
482
+ use_mps_device:
483
+ desc: null
484
+ value: false
485
+ seed:
486
+ desc: null
487
+ value: 42
488
+ data_seed:
489
+ desc: null
490
+ value: null
491
+ jit_mode_eval:
492
+ desc: null
493
+ value: false
494
+ use_ipex:
495
+ desc: null
496
+ value: false
497
+ bf16:
498
+ desc: null
499
+ value: false
500
+ fp16:
501
+ desc: null
502
+ value: false
503
+ fp16_opt_level:
504
+ desc: null
505
+ value: O1
506
+ half_precision_backend:
507
+ desc: null
508
+ value: auto
509
+ bf16_full_eval:
510
+ desc: null
511
+ value: false
512
+ fp16_full_eval:
513
+ desc: null
514
+ value: false
515
+ tf32:
516
+ desc: null
517
+ value: null
518
+ local_rank:
519
+ desc: null
520
+ value: 0
521
+ ddp_backend:
522
+ desc: null
523
+ value: null
524
+ tpu_num_cores:
525
+ desc: null
526
+ value: null
527
+ tpu_metrics_debug:
528
+ desc: null
529
+ value: false
530
+ debug:
531
+ desc: null
532
+ value: []
533
+ dataloader_drop_last:
534
+ desc: null
535
+ value: false
536
+ eval_steps:
537
+ desc: null
538
+ value: 10
539
+ dataloader_num_workers:
540
+ desc: null
541
+ value: 0
542
+ dataloader_prefetch_factor:
543
+ desc: null
544
+ value: null
545
+ past_index:
546
+ desc: null
547
+ value: -1
548
+ run_name:
549
+ desc: null
550
+ value: /kaggle/working/
551
+ disable_tqdm:
552
+ desc: null
553
+ value: false
554
+ remove_unused_columns:
555
+ desc: null
556
+ value: true
557
+ label_names:
558
+ desc: null
559
+ value: null
560
+ load_best_model_at_end:
561
+ desc: null
562
+ value: true
563
+ metric_for_best_model:
564
+ desc: null
565
+ value: loss
566
+ greater_is_better:
567
+ desc: null
568
+ value: false
569
+ ignore_data_skip:
570
+ desc: null
571
+ value: false
572
+ fsdp:
573
+ desc: null
574
+ value: []
575
+ fsdp_min_num_params:
576
+ desc: null
577
+ value: 0
578
+ fsdp_config:
579
+ desc: null
580
+ value:
581
+ min_num_params: 0
582
+ xla: false
583
+ xla_fsdp_v2: false
584
+ xla_fsdp_grad_ckpt: false
585
+ fsdp_transformer_layer_cls_to_wrap:
586
+ desc: null
587
+ value: null
588
+ accelerator_config:
589
+ desc: null
590
+ value:
591
+ split_batches: false
592
+ dispatch_batches: null
593
+ even_batches: true
594
+ use_seedable_sampler: true
595
+ deepspeed:
596
+ desc: null
597
+ value: null
598
+ label_smoothing_factor:
599
+ desc: null
600
+ value: 0.0
601
+ optim:
602
+ desc: null
603
+ value: adamw_torch
604
+ optim_args:
605
+ desc: null
606
+ value: null
607
+ adafactor:
608
+ desc: null
609
+ value: false
610
+ group_by_length:
611
+ desc: null
612
+ value: false
613
+ length_column_name:
614
+ desc: null
615
+ value: length
616
+ report_to:
617
+ desc: null
618
+ value:
619
+ - tensorboard
620
+ - wandb
621
+ ddp_find_unused_parameters:
622
+ desc: null
623
+ value: null
624
+ ddp_bucket_cap_mb:
625
+ desc: null
626
+ value: null
627
+ ddp_broadcast_buffers:
628
+ desc: null
629
+ value: null
630
+ dataloader_pin_memory:
631
+ desc: null
632
+ value: true
633
+ dataloader_persistent_workers:
634
+ desc: null
635
+ value: false
636
+ skip_memory_metrics:
637
+ desc: null
638
+ value: true
639
+ use_legacy_prediction_loop:
640
+ desc: null
641
+ value: false
642
+ push_to_hub:
643
+ desc: null
644
+ value: false
645
+ resume_from_checkpoint:
646
+ desc: null
647
+ value: null
648
+ hub_model_id:
649
+ desc: null
650
+ value: null
651
+ hub_strategy:
652
+ desc: null
653
+ value: every_save
654
+ hub_token:
655
+ desc: null
656
+ value: <HUB_TOKEN>
657
+ hub_private_repo:
658
+ desc: null
659
+ value: false
660
+ hub_always_push:
661
+ desc: null
662
+ value: false
663
+ gradient_checkpointing_kwargs:
664
+ desc: null
665
+ value: null
666
+ include_inputs_for_metrics:
667
+ desc: null
668
+ value: false
669
+ fp16_backend:
670
+ desc: null
671
+ value: auto
672
+ push_to_hub_model_id:
673
+ desc: null
674
+ value: null
675
+ push_to_hub_organization:
676
+ desc: null
677
+ value: null
678
+ push_to_hub_token:
679
+ desc: null
680
+ value: <PUSH_TO_HUB_TOKEN>
681
+ mp_parameters:
682
+ desc: null
683
+ value: ''
684
+ auto_find_batch_size:
685
+ desc: null
686
+ value: false
687
+ full_determinism:
688
+ desc: null
689
+ value: false
690
+ torchdynamo:
691
+ desc: null
692
+ value: null
693
+ ray_scope:
694
+ desc: null
695
+ value: last
696
+ ddp_timeout:
697
+ desc: null
698
+ value: 1800
699
+ torch_compile:
700
+ desc: null
701
+ value: false
702
+ torch_compile_backend:
703
+ desc: null
704
+ value: null
705
+ torch_compile_mode:
706
+ desc: null
707
+ value: null
708
+ dispatch_batches:
709
+ desc: null
710
+ value: null
711
+ split_batches:
712
+ desc: null
713
+ value: null
714
+ include_tokens_per_second:
715
+ desc: null
716
+ value: false
717
+ include_num_input_tokens_seen:
718
+ desc: null
719
+ value: false
720
+ neftune_noise_alpha:
721
+ desc: null
722
+ value: null
wandb/run-20240429_072705-0t7z1bit/files/output.log ADDED
@@ -0,0 +1,513 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
2
+ warnings.warn('Was asked to gather along dimension 0, but all '
3
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
4
+ warnings.warn('Was asked to gather along dimension 0, but all '
5
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
6
+ warnings.warn('Was asked to gather along dimension 0, but all '
7
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
8
+ warnings.warn('Was asked to gather along dimension 0, but all '
9
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
10
+ warnings.warn('Was asked to gather along dimension 0, but all '
11
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
12
+ warnings.warn('Was asked to gather along dimension 0, but all '
13
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
14
+ warnings.warn('Was asked to gather along dimension 0, but all '
15
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
16
+ warnings.warn('Was asked to gather along dimension 0, but all '
17
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
18
+ warnings.warn('Was asked to gather along dimension 0, but all '
19
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
20
+ warnings.warn('Was asked to gather along dimension 0, but all '
21
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
22
+ warnings.warn('Was asked to gather along dimension 0, but all '
23
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
24
+ warnings.warn('Was asked to gather along dimension 0, but all '
25
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
26
+ warnings.warn('Was asked to gather along dimension 0, but all '
27
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
28
+ warnings.warn('Was asked to gather along dimension 0, but all '
29
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
30
+ warnings.warn('Was asked to gather along dimension 0, but all '
31
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
32
+ warnings.warn('Was asked to gather along dimension 0, but all '
33
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
34
+ warnings.warn('Was asked to gather along dimension 0, but all '
35
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
36
+ warnings.warn('Was asked to gather along dimension 0, but all '
37
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
38
+ warnings.warn('Was asked to gather along dimension 0, but all '
39
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
40
+ warnings.warn('Was asked to gather along dimension 0, but all '
41
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
42
+ warnings.warn('Was asked to gather along dimension 0, but all '
43
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
44
+ warnings.warn('Was asked to gather along dimension 0, but all '
45
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
46
+ warnings.warn('Was asked to gather along dimension 0, but all '
47
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
48
+ warnings.warn('Was asked to gather along dimension 0, but all '
49
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
50
+ warnings.warn('Was asked to gather along dimension 0, but all '
51
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
52
+ warnings.warn('Was asked to gather along dimension 0, but all '
53
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
54
+ warnings.warn('Was asked to gather along dimension 0, but all '
55
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
56
+ warnings.warn('Was asked to gather along dimension 0, but all '
57
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
58
+ warnings.warn('Was asked to gather along dimension 0, but all '
59
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
60
+ warnings.warn('Was asked to gather along dimension 0, but all '
61
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
62
+ warnings.warn('Was asked to gather along dimension 0, but all '
63
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
64
+ warnings.warn('Was asked to gather along dimension 0, but all '
65
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
66
+ warnings.warn('Was asked to gather along dimension 0, but all '
67
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
68
+ warnings.warn('Was asked to gather along dimension 0, but all '
69
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
70
+ warnings.warn('Was asked to gather along dimension 0, but all '
71
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
72
+ warnings.warn('Was asked to gather along dimension 0, but all '
73
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
74
+ warnings.warn('Was asked to gather along dimension 0, but all '
75
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
76
+ warnings.warn('Was asked to gather along dimension 0, but all '
77
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
78
+ warnings.warn('Was asked to gather along dimension 0, but all '
79
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
80
+ warnings.warn('Was asked to gather along dimension 0, but all '
81
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
82
+ warnings.warn('Was asked to gather along dimension 0, but all '
83
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
84
+ warnings.warn('Was asked to gather along dimension 0, but all '
85
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
86
+ warnings.warn('Was asked to gather along dimension 0, but all '
87
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
88
+ warnings.warn('Was asked to gather along dimension 0, but all '
89
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
90
+ warnings.warn('Was asked to gather along dimension 0, but all '
91
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
92
+ warnings.warn('Was asked to gather along dimension 0, but all '
93
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
94
+ warnings.warn('Was asked to gather along dimension 0, but all '
95
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
96
+ warnings.warn('Was asked to gather along dimension 0, but all '
97
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
98
+ warnings.warn('Was asked to gather along dimension 0, but all '
99
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
100
+ warnings.warn('Was asked to gather along dimension 0, but all '
101
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
102
+ warnings.warn('Was asked to gather along dimension 0, but all '
103
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
104
+ warnings.warn('Was asked to gather along dimension 0, but all '
105
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
106
+ warnings.warn('Was asked to gather along dimension 0, but all '
107
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
108
+ warnings.warn('Was asked to gather along dimension 0, but all '
109
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
110
+ warnings.warn('Was asked to gather along dimension 0, but all '
111
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
112
+ warnings.warn('Was asked to gather along dimension 0, but all '
113
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
114
+ warnings.warn('Was asked to gather along dimension 0, but all '
115
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
116
+ warnings.warn('Was asked to gather along dimension 0, but all '
117
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
118
+ warnings.warn('Was asked to gather along dimension 0, but all '
119
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
120
+ warnings.warn('Was asked to gather along dimension 0, but all '
121
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
122
+ warnings.warn('Was asked to gather along dimension 0, but all '
123
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
124
+ warnings.warn('Was asked to gather along dimension 0, but all '
125
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
126
+ warnings.warn('Was asked to gather along dimension 0, but all '
127
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
128
+ warnings.warn('Was asked to gather along dimension 0, but all '
129
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
130
+ warnings.warn('Was asked to gather along dimension 0, but all '
131
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
132
+ warnings.warn('Was asked to gather along dimension 0, but all '
133
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
134
+ warnings.warn('Was asked to gather along dimension 0, but all '
135
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
136
+ warnings.warn('Was asked to gather along dimension 0, but all '
137
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
138
+ warnings.warn('Was asked to gather along dimension 0, but all '
139
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
140
+ warnings.warn('Was asked to gather along dimension 0, but all '
141
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
142
+ warnings.warn('Was asked to gather along dimension 0, but all '
143
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
144
+ warnings.warn('Was asked to gather along dimension 0, but all '
145
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
146
+ warnings.warn('Was asked to gather along dimension 0, but all '
147
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
148
+ warnings.warn('Was asked to gather along dimension 0, but all '
149
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
150
+ warnings.warn('Was asked to gather along dimension 0, but all '
151
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
152
+ warnings.warn('Was asked to gather along dimension 0, but all '
153
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
154
+ warnings.warn('Was asked to gather along dimension 0, but all '
155
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
156
+ warnings.warn('Was asked to gather along dimension 0, but all '
157
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
158
+ warnings.warn('Was asked to gather along dimension 0, but all '
159
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
160
+ warnings.warn('Was asked to gather along dimension 0, but all '
161
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
162
+ warnings.warn('Was asked to gather along dimension 0, but all '
163
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
164
+ warnings.warn('Was asked to gather along dimension 0, but all '
165
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
166
+ warnings.warn('Was asked to gather along dimension 0, but all '
167
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
168
+ warnings.warn('Was asked to gather along dimension 0, but all '
169
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
170
+ warnings.warn('Was asked to gather along dimension 0, but all '
171
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
172
+ warnings.warn('Was asked to gather along dimension 0, but all '
173
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
174
+ warnings.warn('Was asked to gather along dimension 0, but all '
175
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
176
+ warnings.warn('Was asked to gather along dimension 0, but all '
177
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
178
+ warnings.warn('Was asked to gather along dimension 0, but all '
179
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
180
+ warnings.warn('Was asked to gather along dimension 0, but all '
181
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
182
+ warnings.warn('Was asked to gather along dimension 0, but all '
183
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
184
+ warnings.warn('Was asked to gather along dimension 0, but all '
185
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
186
+ warnings.warn('Was asked to gather along dimension 0, but all '
187
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
188
+ warnings.warn('Was asked to gather along dimension 0, but all '
189
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
190
+ warnings.warn('Was asked to gather along dimension 0, but all '
191
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
192
+ warnings.warn('Was asked to gather along dimension 0, but all '
193
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
194
+ warnings.warn('Was asked to gather along dimension 0, but all '
195
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
196
+ warnings.warn('Was asked to gather along dimension 0, but all '
197
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
198
+ warnings.warn('Was asked to gather along dimension 0, but all '
199
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
200
+ warnings.warn('Was asked to gather along dimension 0, but all '
201
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
202
+ warnings.warn('Was asked to gather along dimension 0, but all '
203
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
204
+ warnings.warn('Was asked to gather along dimension 0, but all '
205
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
206
+ warnings.warn('Was asked to gather along dimension 0, but all '
207
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
208
+ warnings.warn('Was asked to gather along dimension 0, but all '
209
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
210
+ warnings.warn('Was asked to gather along dimension 0, but all '
211
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
212
+ warnings.warn('Was asked to gather along dimension 0, but all '
213
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
214
+ warnings.warn('Was asked to gather along dimension 0, but all '
215
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
216
+ warnings.warn('Was asked to gather along dimension 0, but all '
217
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
218
+ warnings.warn('Was asked to gather along dimension 0, but all '
219
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
220
+ warnings.warn('Was asked to gather along dimension 0, but all '
221
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
222
+ warnings.warn('Was asked to gather along dimension 0, but all '
223
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
224
+ warnings.warn('Was asked to gather along dimension 0, but all '
225
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
226
+ warnings.warn('Was asked to gather along dimension 0, but all '
227
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
228
+ warnings.warn('Was asked to gather along dimension 0, but all '
229
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
230
+ warnings.warn('Was asked to gather along dimension 0, but all '
231
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
232
+ warnings.warn('Was asked to gather along dimension 0, but all '
233
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
234
+ warnings.warn('Was asked to gather along dimension 0, but all '
235
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
236
+ warnings.warn('Was asked to gather along dimension 0, but all '
237
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
238
+ warnings.warn('Was asked to gather along dimension 0, but all '
239
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
240
+ warnings.warn('Was asked to gather along dimension 0, but all '
241
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
242
+ warnings.warn('Was asked to gather along dimension 0, but all '
243
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
244
+ warnings.warn('Was asked to gather along dimension 0, but all '
245
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
246
+ warnings.warn('Was asked to gather along dimension 0, but all '
247
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
248
+ warnings.warn('Was asked to gather along dimension 0, but all '
249
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
250
+ warnings.warn('Was asked to gather along dimension 0, but all '
251
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
252
+ warnings.warn('Was asked to gather along dimension 0, but all '
253
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
254
+ warnings.warn('Was asked to gather along dimension 0, but all '
255
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
256
+ warnings.warn('Was asked to gather along dimension 0, but all '
257
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
258
+ warnings.warn('Was asked to gather along dimension 0, but all '
259
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
260
+ warnings.warn('Was asked to gather along dimension 0, but all '
261
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
262
+ warnings.warn('Was asked to gather along dimension 0, but all '
263
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
264
+ warnings.warn('Was asked to gather along dimension 0, but all '
265
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
266
+ warnings.warn('Was asked to gather along dimension 0, but all '
267
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
268
+ warnings.warn('Was asked to gather along dimension 0, but all '
269
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
270
+ warnings.warn('Was asked to gather along dimension 0, but all '
271
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
272
+ warnings.warn('Was asked to gather along dimension 0, but all '
273
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
274
+ warnings.warn('Was asked to gather along dimension 0, but all '
275
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
276
+ warnings.warn('Was asked to gather along dimension 0, but all '
277
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
278
+ warnings.warn('Was asked to gather along dimension 0, but all '
279
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
280
+ warnings.warn('Was asked to gather along dimension 0, but all '
281
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
282
+ warnings.warn('Was asked to gather along dimension 0, but all '
283
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
284
+ warnings.warn('Was asked to gather along dimension 0, but all '
285
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
286
+ warnings.warn('Was asked to gather along dimension 0, but all '
287
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
288
+ warnings.warn('Was asked to gather along dimension 0, but all '
289
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
290
+ warnings.warn('Was asked to gather along dimension 0, but all '
291
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
292
+ warnings.warn('Was asked to gather along dimension 0, but all '
293
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
294
+ warnings.warn('Was asked to gather along dimension 0, but all '
295
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
296
+ warnings.warn('Was asked to gather along dimension 0, but all '
297
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
298
+ warnings.warn('Was asked to gather along dimension 0, but all '
299
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
300
+ warnings.warn('Was asked to gather along dimension 0, but all '
301
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
302
+ warnings.warn('Was asked to gather along dimension 0, but all '
303
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
304
+ warnings.warn('Was asked to gather along dimension 0, but all '
305
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
306
+ warnings.warn('Was asked to gather along dimension 0, but all '
307
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
308
+ warnings.warn('Was asked to gather along dimension 0, but all '
309
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
310
+ warnings.warn('Was asked to gather along dimension 0, but all '
311
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
312
+ warnings.warn('Was asked to gather along dimension 0, but all '
313
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
314
+ warnings.warn('Was asked to gather along dimension 0, but all '
315
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
316
+ warnings.warn('Was asked to gather along dimension 0, but all '
317
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
318
+ warnings.warn('Was asked to gather along dimension 0, but all '
319
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
320
+ warnings.warn('Was asked to gather along dimension 0, but all '
321
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
322
+ warnings.warn('Was asked to gather along dimension 0, but all '
323
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
324
+ warnings.warn('Was asked to gather along dimension 0, but all '
325
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
326
+ warnings.warn('Was asked to gather along dimension 0, but all '
327
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
328
+ warnings.warn('Was asked to gather along dimension 0, but all '
329
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
330
+ warnings.warn('Was asked to gather along dimension 0, but all '
331
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
332
+ warnings.warn('Was asked to gather along dimension 0, but all '
333
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
334
+ warnings.warn('Was asked to gather along dimension 0, but all '
335
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
336
+ warnings.warn('Was asked to gather along dimension 0, but all '
337
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
338
+ warnings.warn('Was asked to gather along dimension 0, but all '
339
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
340
+ warnings.warn('Was asked to gather along dimension 0, but all '
341
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
342
+ warnings.warn('Was asked to gather along dimension 0, but all '
343
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
344
+ warnings.warn('Was asked to gather along dimension 0, but all '
345
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
346
+ warnings.warn('Was asked to gather along dimension 0, but all '
347
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
348
+ warnings.warn('Was asked to gather along dimension 0, but all '
349
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
350
+ warnings.warn('Was asked to gather along dimension 0, but all '
351
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
352
+ warnings.warn('Was asked to gather along dimension 0, but all '
353
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
354
+ warnings.warn('Was asked to gather along dimension 0, but all '
355
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
356
+ warnings.warn('Was asked to gather along dimension 0, but all '
357
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
358
+ warnings.warn('Was asked to gather along dimension 0, but all '
359
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
360
+ warnings.warn('Was asked to gather along dimension 0, but all '
361
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
362
+ warnings.warn('Was asked to gather along dimension 0, but all '
363
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
364
+ warnings.warn('Was asked to gather along dimension 0, but all '
365
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
366
+ warnings.warn('Was asked to gather along dimension 0, but all '
367
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
368
+ warnings.warn('Was asked to gather along dimension 0, but all '
369
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
370
+ warnings.warn('Was asked to gather along dimension 0, but all '
371
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
372
+ warnings.warn('Was asked to gather along dimension 0, but all '
373
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
374
+ warnings.warn('Was asked to gather along dimension 0, but all '
375
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
376
+ warnings.warn('Was asked to gather along dimension 0, but all '
377
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
378
+ warnings.warn('Was asked to gather along dimension 0, but all '
379
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
380
+ warnings.warn('Was asked to gather along dimension 0, but all '
381
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
382
+ warnings.warn('Was asked to gather along dimension 0, but all '
383
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
384
+ warnings.warn('Was asked to gather along dimension 0, but all '
385
+ 43
386
+ What is the overall deductible?
387
+ SBC
388
+ What are the benefits covered in my plan?
389
+ SBC
390
+ We have a new hire enrollment for Rubi Olivares Cisneros of Spectrum ABA, LLC. Anticipated Effective Date: 3/1/2024. They would like to enroll in the preexisting UnitedHealthcare plans under the same group: 223343
391
+ Install
392
+ I have an employee for the group Guardrail (294930) that is trying to enroll in coverage under policy code CWYN. This is currently a ghost plan for this group. Can you please help us enroll Dinesh Ketumakha under plan CWYN effective 1/1/2024. I have attached this employee's enrollment form. Please let me know if you need anything else to complete this enrollment.
393
+ Install
394
+ What all assistance can you offer?
395
+ SBC
396
+ I have to generate a summary for my benefits data for the group XYZ
397
+ Summarize
398
+ What is the amount i have to pay if i visit the healthcare provider in his clinic. I am going to visit an in network healthcare provider
399
+ SBC
400
+ what is the total premium paid for 2023 under the writing agent 018104-01
401
+ Compensation
402
+ Can you give the list of customers which are tied to producer with producer id 018104
403
+ Compensation
404
+ What is my billed premium for the effective date 01/01/2023 for producer name PROFESSIONAL GROUP PLANS, INC
405
+ Compensation
406
+ The following request has been made by Professional group to remove the member Bhasin from the group ID 88304 since he is no longer part of the employer. can you help in the above request
407
+ Terminate
408
+ What is the cost for outpatient surgery??
409
+ SBC
410
+ In case of an emergency what will be the copay value
411
+ SBC
412
+ What are the services that are not covered under PPO
413
+ SBC
414
+ Is dental services covered in it as well?
415
+ SBC
416
+ I need to extract the total amount paid under medical coverage for customer name Productigear in this quarter
417
+ Compensation
418
+ Please cancel dental coverage effective 1/1/2024 for Franciszek Ratulowski.
419
+ Change
420
+ Can you please update on the following basis: Family to employee + children, waiving dental and keeping vision. Please let me know if you need anything additional.
421
+ Change
422
+ We got a request to update the home address of the following employee: 09Q8404 Previous Address :: 470 Clark St Current Address :: 1012 Sierra Street, Turlock, Stanislaus, CA, USA 95380
423
+ Change
424
+ What are the different issuer states under the broker 018104
425
+ Compensation
426
+ Are the diagnostic tests covered in my current benefit
427
+ SBC
428
+ I was wondering if it's possible to extend benefits for a dismissed employee into February. Is there someone over there who could help?
429
+ Terminate
430
+ Please see the attached employee application to be added to the existing group insurance plan.
431
+ Install
432
+ Please correct member Sabrina Jenkins’ coverage end date from 01/08/2024 to 01/09/2024. An error was made when terminating her with the dates. Thank you and please let me know if anything else is needed to process it
433
+ Change
434
+ I want to know about my out of pocket limit
435
+ SBC
436
+ Please correct member Sabrina Jenkins’ coverage end date from 01/08/2024 to 01/09/2024. An error was made when terminating her with the dates.
437
+ Change
438
+ Please add spouse and son 2/1 for OE.
439
+ Install
440
+ Please process billing contact change to be effective 1/1/24.
441
+ Install
442
+ Can you ensure this member and dependent are moved from COBRA continuation class to active class?
443
+ Install
444
+ please terminate Pamela Escobar effective 3.31.24 as she is no longer employed
445
+ Terminate
446
+ Please process this change for John Morgan, he is changing from Family to employee + children, waiving dental and keeping vision
447
+ Change
448
+ These need to be updated ASAP, so I am sending a list of employees with their new salaries. Please update them on your end and let me know when they have been processed
449
+ Compensation
450
+ Do i need a referral to see a specialist?
451
+ SBC
452
+ We have a Home Address Update for Jessie Fetterling of Big Edition, Inc.
453
+ Change
454
+ I was wondering if it's possible to extend benefits for a dismissed employee into February
455
+ Terminate
456
+ Please cancel dental coverage effective 1/1/2024 for Franciszek Ratulowski
457
+ Change
458
+ They would like to enroll spouse, Allison Bailey-Nouri
459
+ Install
460
+ Please enroll this member and his wife for 1/1/2024 effective date.
461
+ Install
462
+ What are the services provided if I have a hospital stay under my current benefits?
463
+ SBC
464
+ Please process the attached Enrollment request for employer group American Indian OIC
465
+ Install
466
+ What is the average split percentage for the customer ABC?
467
+ Compensation
468
+ Please see attached new hire app effective 11/1/2023.
469
+ Install
470
+ Please enroll new employee, Stephanie Hamilton, with Davis Farr LLP – Group #03V5870, to become effective 10/01/2023
471
+ Install
472
+ SBC
473
+ SBC
474
+ Install
475
+ Install
476
+ SBC
477
+ Summarize
478
+ SBC
479
+ Compensation
480
+ Compensation
481
+ Compensation
482
+ Terminate
483
+ SBC
484
+ SBC
485
+ SBC
486
+ SBC
487
+ Compensation
488
+ Change
489
+ Change
490
+ Change
491
+ Compensation
492
+ SBC
493
+ Terminate
494
+ Install
495
+ Change
496
+ SBC
497
+ Change
498
+ Install
499
+ Install
500
+ Install
501
+ Terminate
502
+ Change
503
+ Compensation
504
+ SBC
505
+ Change
506
+ Terminate
507
+ Change
508
+ Install
509
+ Install
510
+ SBC
511
+ Install
512
+ Compensation
513
+ Install
wandb/run-20240429_072705-0t7z1bit/files/requirements.txt ADDED
@@ -0,0 +1,880 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ Levenshtein==0.25.0
16
+ LunarCalendar==0.0.9
17
+ Mako==1.3.2
18
+ Markdown==3.5.2
19
+ MarkupSafe==2.1.3
20
+ MarkupSafe==2.1.5
21
+ Pillow==9.5.0
22
+ PuLP==2.8.0
23
+ PyArabic==0.6.15
24
+ PyAstronomy==0.21.0
25
+ PyJWT==2.8.0
26
+ PyMeeus==0.5.12
27
+ PySocks==1.7.1
28
+ PyUpSet==0.1.1.post7
29
+ PyWavelets==1.5.0
30
+ PyYAML==6.0.1
31
+ Pygments==2.17.2
32
+ Pympler==1.0.1
33
+ QtPy==2.4.1
34
+ Rtree==1.2.0
35
+ SQLAlchemy==2.0.25
36
+ SecretStorage==3.3.3
37
+ Send2Trash==1.8.2
38
+ Shapely==1.8.5.post1
39
+ Shimmy==1.3.0
40
+ SimpleITK==2.3.1
41
+ TPOT==0.12.1
42
+ Theano-PyMC==1.1.2
43
+ Theano==1.0.5
44
+ Unidecode==1.3.8
45
+ Wand==0.6.13
46
+ Werkzeug==3.0.1
47
+ absl-py==1.4.0
48
+ accelerate==0.28.0
49
+ access==1.1.9
50
+ affine==2.4.0
51
+ aiobotocore==2.12.1
52
+ aiofiles==22.1.0
53
+ aiohttp-cors==0.7.0
54
+ aiohttp==3.9.1
55
+ aioitertools==0.11.0
56
+ aiorwlock==1.3.0
57
+ aiosignal==1.3.1
58
+ aiosqlite==0.19.0
59
+ albumentations==1.4.0
60
+ alembic==1.13.1
61
+ altair==5.2.0
62
+ annotated-types==0.6.0
63
+ annoy==1.17.3
64
+ anyio==4.2.0
65
+ apache-beam==2.46.0
66
+ aplus==0.11.0
67
+ appdirs==1.4.4
68
+ archspec==0.2.2
69
+ argon2-cffi-bindings==21.2.0
70
+ argon2-cffi==23.1.0
71
+ array-record==0.5.0
72
+ arrow==1.3.0
73
+ arviz==0.17.1
74
+ astroid==3.0.3
75
+ astropy-iers-data==0.2024.3.18.0.29.47
76
+ astropy==6.0.0
77
+ asttokens==2.4.1
78
+ astunparse==1.6.3
79
+ async-lru==2.0.4
80
+ async-timeout==4.0.3
81
+ attrs==23.2.0
82
+ audioread==3.0.1
83
+ auto_gptq==0.7.1
84
+ autopep8==2.0.4
85
+ backoff==2.2.1
86
+ bayesian-optimization==1.4.3
87
+ beatrix_jupyterlab==2023.128.151533
88
+ beautifulsoup4==4.12.2
89
+ bidict==0.23.1
90
+ bitsandbytes==0.43.1
91
+ blake3==0.2.1
92
+ bleach==6.1.0
93
+ blessed==1.20.0
94
+ blinker==1.7.0
95
+ blis==0.7.10
96
+ blosc2==2.5.1
97
+ bokeh==3.3.4
98
+ boltons==23.1.1
99
+ boto3==1.26.100
100
+ botocore==1.34.51
101
+ bq_helper==0.4.1
102
+ bqplot==0.12.43
103
+ branca==0.7.1
104
+ brewer2mpl==1.4.1
105
+ brotlipy==0.7.0
106
+ cached-property==1.5.2
107
+ cachetools==4.2.4
108
+ cachetools==5.3.2
109
+ catalogue==2.0.10
110
+ catalyst==22.4
111
+ catboost==1.2.3
112
+ category-encoders==2.6.3
113
+ certifi==2024.2.2
114
+ cesium==0.12.1
115
+ cffi==1.16.0
116
+ charset-normalizer==3.3.2
117
+ chex==0.1.85
118
+ cleverhans==4.0.0
119
+ click-plugins==1.1.1
120
+ click==8.1.7
121
+ cligj==0.7.2
122
+ cloud-tpu-client==0.10
123
+ cloud-tpu-profiler==2.4.0
124
+ cloudpathlib==0.16.0
125
+ cloudpickle==2.2.1
126
+ cloudpickle==3.0.0
127
+ cmdstanpy==1.2.1
128
+ cmudict==1.0.21
129
+ colorama==0.4.6
130
+ colorcet==3.1.0
131
+ coloredlogs==15.0.1
132
+ colorful==0.5.6
133
+ colorlog==6.8.2
134
+ colorlover==0.3.0
135
+ comm==0.2.1
136
+ conda-libmamba-solver==23.7.0
137
+ conda-package-handling==2.2.0
138
+ conda==23.7.4
139
+ conda_package_streaming==0.9.0
140
+ confection==0.1.4
141
+ contextily==1.5.2
142
+ contourpy==1.2.0
143
+ convertdate==2.4.0
144
+ crcmod==1.7
145
+ cryptography==41.0.7
146
+ cuda-python==12.4.0
147
+ cudf==23.8.0
148
+ cufflinks==0.17.3
149
+ cuml==23.8.0
150
+ cupy==13.0.0
151
+ cycler==0.12.1
152
+ cymem==2.0.8
153
+ cytoolz==0.12.3
154
+ daal4py==2024.1.0
155
+ daal==2024.1.0
156
+ dacite==1.8.1
157
+ dask-cuda==23.8.0
158
+ dask-cudf==23.8.0
159
+ dask-expr==1.0.4
160
+ dask==2024.3.1
161
+ dataclasses-json==0.6.4
162
+ dataproc_jupyter_plugin==0.1.66
163
+ datasets==2.1.0
164
+ datashader==0.16.0
165
+ datatile==1.0.3
166
+ db-dtypes==1.2.0
167
+ deap==1.4.1
168
+ debugpy==1.8.0
169
+ decorator==5.1.1
170
+ deepdiff==6.7.1
171
+ defusedxml==0.7.1
172
+ deprecation==2.1.0
173
+ descartes==1.1.0
174
+ dill==0.3.8
175
+ dipy==1.9.0
176
+ distlib==0.3.8
177
+ distributed==2023.7.1
178
+ distro==1.9.0
179
+ dm-tree==0.1.8
180
+ docker-pycreds==0.4.0
181
+ docker==7.0.0
182
+ docopt==0.6.2
183
+ docstring-parser==0.15
184
+ docstring-to-markdown==0.15
185
+ docutils==0.20.1
186
+ earthengine-api==0.1.394
187
+ easydict==1.13
188
+ easyocr==1.7.1
189
+ ecos==2.0.13
190
+ eli5==0.13.0
191
+ emoji==2.10.1
192
+ en-core-web-lg==3.7.1
193
+ en-core-web-sm==3.7.1
194
+ entrypoints==0.4
195
+ ephem==4.1.5
196
+ esda==2.5.1
197
+ essentia==2.1b6.dev1110
198
+ et-xmlfile==1.1.0
199
+ etils==1.6.0
200
+ exceptiongroup==1.2.0
201
+ executing==2.0.1
202
+ explainable-ai-sdk==1.3.3
203
+ fastai==2.7.14
204
+ fastapi==0.108.0
205
+ fastavro==1.9.3
206
+ fastcore==1.5.29
207
+ fastdownload==0.0.7
208
+ fasteners==0.19
209
+ fastjsonschema==2.19.1
210
+ fastprogress==1.0.3
211
+ fastrlock==0.8.2
212
+ fasttext==0.9.2
213
+ feather-format==0.4.1
214
+ featuretools==1.30.0
215
+ filelock==3.13.1
216
+ fiona==1.9.6
217
+ fitter==1.7.0
218
+ flake8==7.0.0
219
+ flashtext==2.7
220
+ flatbuffers==23.5.26
221
+ flax==0.8.2
222
+ folium==0.16.0
223
+ fonttools==4.47.0
224
+ fonttools==4.49.0
225
+ fqdn==1.5.1
226
+ frozendict==2.4.0
227
+ frozenlist==1.4.1
228
+ fsspec==2024.3.0
229
+ funcy==2.0
230
+ fury==0.10.0
231
+ future==1.0.0
232
+ fuzzywuzzy==0.18.0
233
+ gast==0.5.4
234
+ gatspy==0.3
235
+ gcsfs==2023.12.2.post1
236
+ gekko==1.1.1
237
+ gensim==4.3.2
238
+ geographiclib==2.0
239
+ geojson==3.1.0
240
+ geopandas==0.14.3
241
+ geoplot==0.5.1
242
+ geopy==2.4.1
243
+ geoviews==1.11.1
244
+ ggplot==0.11.5
245
+ giddy==2.3.5
246
+ gitdb==4.0.11
247
+ google-ai-generativelanguage==0.4.0
248
+ google-api-core==2.11.1
249
+ google-api-core==2.17.1
250
+ google-api-python-client==2.122.0
251
+ google-apitools==0.5.31
252
+ google-auth-httplib2==0.1.1
253
+ google-auth-oauthlib==1.2.0
254
+ google-auth==2.26.1
255
+ google-cloud-aiplatform==0.6.0a1
256
+ google-cloud-artifact-registry==1.10.0
257
+ google-cloud-automl==1.0.1
258
+ google-cloud-bigquery==2.34.4
259
+ google-cloud-bigtable==1.7.3
260
+ google-cloud-core==2.4.1
261
+ google-cloud-datastore==2.19.0
262
+ google-cloud-dlp==3.14.0
263
+ google-cloud-jupyter-config==0.0.5
264
+ google-cloud-language==2.13.3
265
+ google-cloud-monitoring==2.18.0
266
+ google-cloud-pubsub==2.19.0
267
+ google-cloud-pubsublite==1.9.0
268
+ google-cloud-recommendations-ai==0.7.1
269
+ google-cloud-resource-manager==1.11.0
270
+ google-cloud-spanner==3.40.1
271
+ google-cloud-storage==1.44.0
272
+ google-cloud-translate==3.12.1
273
+ google-cloud-videointelligence==2.13.3
274
+ google-cloud-vision==2.8.0
275
+ google-crc32c==1.5.0
276
+ google-generativeai==0.4.1
277
+ google-pasta==0.2.0
278
+ google-resumable-media==2.7.0
279
+ googleapis-common-protos==1.62.0
280
+ gplearn==0.4.2
281
+ gpustat==1.0.0
282
+ gpxpy==1.6.2
283
+ graphviz==0.20.2
284
+ greenlet==3.0.3
285
+ grpc-google-iam-v1==0.12.7
286
+ grpcio-status==1.48.1
287
+ grpcio-status==1.48.2
288
+ grpcio==1.51.1
289
+ grpcio==1.60.0
290
+ gviz-api==1.10.0
291
+ gym-notices==0.0.8
292
+ gym==0.26.2
293
+ gymnasium==0.29.0
294
+ h11==0.14.0
295
+ h2o==3.46.0.1
296
+ h5netcdf==1.3.0
297
+ h5py==3.10.0
298
+ haversine==2.8.1
299
+ hdfs==2.7.3
300
+ hep-ml==0.7.2
301
+ hijri-converter==2.3.1
302
+ hmmlearn==0.3.2
303
+ holidays==0.24
304
+ holoviews==1.18.3
305
+ hpsklearn==0.1.0
306
+ html5lib==1.1
307
+ htmlmin==0.1.12
308
+ httpcore==1.0.4
309
+ httplib2==0.21.0
310
+ httptools==0.6.1
311
+ httpx==0.27.0
312
+ huggingface-hub==0.21.4
313
+ humanfriendly==10.0
314
+ hunspell==0.5.5
315
+ husl==4.0.3
316
+ hydra-slayer==0.5.0
317
+ hyperopt==0.2.7
318
+ hypertools==0.8.0
319
+ idna==3.6
320
+ igraph==0.11.4
321
+ imagecodecs==2024.1.1
322
+ imageio==2.33.1
323
+ imbalanced-learn==0.12.0
324
+ imgaug==0.4.0
325
+ importlib-metadata==6.11.0
326
+ importlib-metadata==7.0.1
327
+ importlib-resources==6.1.1
328
+ inequality==1.0.1
329
+ iniconfig==2.0.0
330
+ ipydatawidgets==4.3.5
331
+ ipykernel==6.28.0
332
+ ipyleaflet==0.18.2
333
+ ipympl==0.7.0
334
+ ipython-genutils==0.2.0
335
+ ipython-genutils==0.2.0
336
+ ipython-sql==0.5.0
337
+ ipython==8.20.0
338
+ ipyvolume==0.6.3
339
+ ipyvue==1.10.2
340
+ ipyvuetify==1.9.2
341
+ ipywebrtc==0.6.0
342
+ ipywidgets==7.7.1
343
+ isoduration==20.11.0
344
+ isort==5.13.2
345
+ isoweek==1.3.3
346
+ itsdangerous==2.1.2
347
+ jaraco.classes==3.3.0
348
+ jax-jumpy==1.0.0
349
+ jax==0.4.23
350
+ jaxlib==0.4.23.dev20240116
351
+ jedi==0.19.1
352
+ jeepney==0.8.0
353
+ jieba==0.42.1
354
+ jmespath==1.0.1
355
+ joblib==1.3.2
356
+ json5==0.9.14
357
+ jsonpatch==1.33
358
+ jsonpointer==2.4
359
+ jsonschema-specifications==2023.12.1
360
+ jsonschema==4.20.0
361
+ jupyter-console==6.6.3
362
+ jupyter-events==0.9.0
363
+ jupyter-http-over-ws==0.0.8
364
+ jupyter-lsp==1.5.1
365
+ jupyter-server-mathjax==0.2.6
366
+ jupyter-ydoc==0.2.5
367
+ jupyter_client==7.4.9
368
+ jupyter_client==8.6.0
369
+ jupyter_core==5.7.1
370
+ jupyter_server==2.13.0
371
+ jupyter_server_fileid==0.9.1
372
+ jupyter_server_proxy==4.1.0
373
+ jupyter_server_terminals==0.5.1
374
+ jupyter_server_ydoc==0.8.0
375
+ jupyterlab-lsp==5.1.0
376
+ jupyterlab-widgets==3.0.9
377
+ jupyterlab==4.1.5
378
+ jupyterlab_git==0.44.0
379
+ jupyterlab_pygments==0.3.0
380
+ jupyterlab_server==2.25.2
381
+ jupytext==1.16.0
382
+ kaggle-environments==1.14.3
383
+ kaggle==1.6.6
384
+ kagglehub==0.2.0
385
+ keras-cv==0.8.2
386
+ keras-nlp==0.8.2
387
+ keras-tuner==1.4.6
388
+ keras==3.0.5
389
+ kernels-mixer==0.0.7
390
+ keyring==24.3.0
391
+ keyrings.google-artifactregistry-auth==1.1.2
392
+ kfp-pipeline-spec==0.2.2
393
+ kfp-server-api==2.0.5
394
+ kfp==2.5.0
395
+ kiwisolver==1.4.5
396
+ kmapper==2.0.1
397
+ kmodes==0.12.2
398
+ korean-lunar-calendar==0.3.1
399
+ kornia==0.7.2
400
+ kornia_rs==0.1.2
401
+ kt-legacy==1.0.5
402
+ kubernetes==26.1.0
403
+ langcodes==3.3.0
404
+ langid==1.1.6
405
+ lazy_loader==0.3
406
+ learntools==0.3.4
407
+ leven==1.0.4
408
+ libclang==16.0.6
409
+ libmambapy==1.5.0
410
+ libpysal==4.9.2
411
+ librosa==0.10.1
412
+ lightgbm==4.2.0
413
+ lightning-utilities==0.10.1
414
+ lime==0.2.0.1
415
+ line-profiler==4.1.2
416
+ linkify-it-py==2.0.3
417
+ llvmlite==0.41.1
418
+ llvmlite==0.42.0
419
+ lml==0.1.0
420
+ locket==1.0.0
421
+ loguru==0.7.2
422
+ lxml==5.1.0
423
+ lz4==4.3.3
424
+ mamba==1.5.0
425
+ mapclassify==2.6.1
426
+ markdown-it-py==3.0.0
427
+ markovify==0.9.4
428
+ marshmallow==3.21.1
429
+ matplotlib-inline==0.1.6
430
+ matplotlib-venn==0.11.10
431
+ matplotlib==3.7.5
432
+ matplotlib==3.8.3
433
+ mccabe==0.7.0
434
+ mdit-py-plugins==0.4.0
435
+ mdurl==0.1.2
436
+ memory-profiler==0.61.0
437
+ menuinst==2.0.1
438
+ mercantile==1.2.1
439
+ mgwr==2.2.1
440
+ missingno==0.5.2
441
+ mistune==0.8.4
442
+ mizani==0.11.0
443
+ ml-dtypes==0.2.0
444
+ mlcrate==0.2.0
445
+ mlens==0.2.3
446
+ mlxtend==0.23.1
447
+ mmh3==4.1.0
448
+ mne==1.6.1
449
+ mnist==0.2.2
450
+ mock==5.1.0
451
+ momepy==0.7.0
452
+ more-itertools==10.2.0
453
+ mpld3==0.5.10
454
+ mpmath==1.3.0
455
+ msgpack-numpy==0.4.8
456
+ msgpack==1.0.7
457
+ multidict==6.0.4
458
+ multimethod==1.10
459
+ multipledispatch==1.0.0
460
+ multiprocess==0.70.16
461
+ munkres==1.1.4
462
+ murmurhash==1.0.10
463
+ mypy-extensions==1.0.0
464
+ namex==0.0.7
465
+ nb-conda-kernels==2.3.1
466
+ nb_conda==2.2.1
467
+ nbclassic==1.0.0
468
+ nbclient==0.5.13
469
+ nbconvert==6.4.5
470
+ nbdime==3.2.0
471
+ nbformat==5.9.2
472
+ ndindex==1.8
473
+ nest-asyncio==1.5.8
474
+ networkx==3.2.1
475
+ nibabel==5.2.1
476
+ nilearn==0.10.3
477
+ ninja==1.11.1.1
478
+ nltk==3.2.4
479
+ nose==1.3.7
480
+ notebook==6.5.4
481
+ notebook==6.5.6
482
+ notebook_executor==0.2
483
+ notebook_shim==0.2.3
484
+ numba==0.58.1
485
+ numba==0.59.0
486
+ numexpr==2.9.0
487
+ numpy==1.26.4
488
+ nvidia-ml-py==11.495.46
489
+ nvtx==0.2.10
490
+ oauth2client==4.1.3
491
+ oauthlib==3.2.2
492
+ objsize==0.6.1
493
+ odfpy==1.4.1
494
+ olefile==0.47
495
+ onnx==1.15.0
496
+ opencensus-context==0.1.3
497
+ opencensus==0.11.4
498
+ opencv-contrib-python==4.9.0.80
499
+ opencv-python-headless==4.9.0.80
500
+ opencv-python==4.9.0.80
501
+ openpyxl==3.1.2
502
+ openslide-python==1.3.1
503
+ opentelemetry-api==1.22.0
504
+ opentelemetry-exporter-otlp-proto-common==1.22.0
505
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
506
+ opentelemetry-exporter-otlp-proto-http==1.22.0
507
+ opentelemetry-exporter-otlp==1.22.0
508
+ opentelemetry-proto==1.22.0
509
+ opentelemetry-sdk==1.22.0
510
+ opentelemetry-semantic-conventions==0.43b0
511
+ opt-einsum==3.3.0
512
+ optax==0.2.1
513
+ optimum==1.19.1
514
+ optuna==3.6.0
515
+ orbax-checkpoint==0.5.6
516
+ ordered-set==4.1.0
517
+ orjson==3.9.10
518
+ ortools==9.4.1874
519
+ osmnx==1.9.1
520
+ overrides==7.4.0
521
+ packaging==21.3
522
+ pandas-datareader==0.10.0
523
+ pandas-profiling==3.6.6
524
+ pandas-summary==0.2.0
525
+ pandas==2.1.4
526
+ pandas==2.2.1
527
+ pandasql==0.7.3
528
+ pandocfilters==1.5.0
529
+ panel==1.3.8
530
+ papermill==2.5.0
531
+ param==2.0.2
532
+ parso==0.8.3
533
+ partd==1.4.1
534
+ path.py==12.5.0
535
+ path==16.10.0
536
+ pathos==0.3.2
537
+ pathy==0.10.3
538
+ patsy==0.5.6
539
+ pdf2image==1.17.0
540
+ peft==0.10.0
541
+ pettingzoo==1.24.0
542
+ pexpect==4.8.0
543
+ pexpect==4.9.0
544
+ phik==0.12.4
545
+ pickleshare==0.7.5
546
+ pip==23.3.2
547
+ pkgutil_resolve_name==1.3.10
548
+ platformdirs==4.2.0
549
+ plotly-express==0.4.1
550
+ plotly==5.18.0
551
+ plotnine==0.13.2
552
+ pluggy==1.4.0
553
+ pointpats==2.4.0
554
+ polars==0.20.15
555
+ polyglot==16.7.4
556
+ pooch==1.8.1
557
+ pox==0.3.4
558
+ ppca==0.0.4
559
+ ppft==1.7.6.8
560
+ preprocessing==0.1.13
561
+ preshed==3.0.9
562
+ prettytable==3.9.0
563
+ progressbar2==4.4.2
564
+ prometheus-client==0.19.0
565
+ promise==2.3
566
+ prompt-toolkit==3.0.42
567
+ prompt-toolkit==3.0.43
568
+ pronouncing==0.2.0
569
+ prophet==1.1.1
570
+ proto-plus==1.23.0
571
+ protobuf==3.20.3
572
+ protobuf==4.21.12
573
+ psutil==5.9.3
574
+ psutil==5.9.7
575
+ ptyprocess==0.7.0
576
+ pudb==2024.1
577
+ pure-eval==0.2.2
578
+ py-cpuinfo==9.0.0
579
+ py-spy==0.3.14
580
+ py4j==0.10.9.7
581
+ pyLDAvis==3.4.1
582
+ pyOpenSSL==23.3.0
583
+ pyaml==23.12.0
584
+ pyarrow==11.0.0
585
+ pyasn1-modules==0.3.0
586
+ pyasn1==0.5.1
587
+ pybind11==2.11.1
588
+ pyclipper==1.3.0.post5
589
+ pycodestyle==2.11.1
590
+ pycosat==0.6.6
591
+ pycparser==2.21
592
+ pycryptodome==3.20.0
593
+ pyct==0.5.0
594
+ pycuda==2024.1
595
+ pydantic==2.5.3
596
+ pydantic==2.6.4
597
+ pydantic_core==2.14.6
598
+ pydantic_core==2.16.3
599
+ pydegensac==0.1.2
600
+ pydicom==2.4.4
601
+ pydocstyle==6.3.0
602
+ pydot==1.4.2
603
+ pydub==0.25.1
604
+ pyemd==1.0.0
605
+ pyerfa==2.0.1.1
606
+ pyexcel-io==0.6.6
607
+ pyexcel-ods==0.6.0
608
+ pyflakes==3.2.0
609
+ pygltflib==1.16.2
610
+ pykalman==0.9.5
611
+ pylibraft==23.8.0
612
+ pylint==3.0.4
613
+ pymc3==3.11.4
614
+ pymongo==3.13.0
615
+ pynndescent==0.5.11
616
+ pynvml==11.4.1
617
+ pynvrtc==9.2
618
+ pyocr==0.8.5
619
+ pyparsing==3.1.1
620
+ pyparsing==3.1.2
621
+ pypdf==4.1.0
622
+ pyproj==3.6.1
623
+ pysal==24.1
624
+ pyshp==2.3.1
625
+ pytesseract==0.3.10
626
+ pytest==8.1.1
627
+ python-Levenshtein==0.25.0
628
+ python-bidi==0.4.2
629
+ python-dateutil==2.9.0.post0
630
+ python-dotenv==1.0.0
631
+ python-json-logger==2.0.7
632
+ python-louvain==0.16
633
+ python-lsp-jsonrpc==1.1.2
634
+ python-lsp-server==1.10.1
635
+ python-slugify==8.0.4
636
+ python-utils==3.8.2
637
+ pythreejs==2.4.2
638
+ pytoolconfig==1.3.1
639
+ pytools==2023.1.1
640
+ pytorch-ignite==0.4.13
641
+ pytorch-lightning==2.2.1
642
+ pytz==2023.3.post1
643
+ pytz==2024.1
644
+ pyu2f==0.1.5
645
+ pyviz_comms==3.0.1
646
+ pyzmq==24.0.1
647
+ pyzmq==25.1.2
648
+ qgrid==1.3.1
649
+ qtconsole==5.5.1
650
+ quantecon==0.7.2
651
+ quantities==0.15.0
652
+ qudida==0.0.4
653
+ raft-dask==23.8.0
654
+ rapidfuzz==3.6.2
655
+ rasterio==1.3.9
656
+ rasterstats==0.19.0
657
+ ray-cpp==2.9.0
658
+ ray==2.9.0
659
+ referencing==0.32.1
660
+ regex==2023.12.25
661
+ requests-oauthlib==1.3.1
662
+ requests-toolbelt==0.10.1
663
+ requests==2.31.0
664
+ responses==0.18.0
665
+ retrying==1.3.3
666
+ retrying==1.3.4
667
+ rfc3339-validator==0.1.4
668
+ rfc3986-validator==0.1.1
669
+ rgf-python==3.12.0
670
+ rich-click==1.7.4
671
+ rich==13.7.0
672
+ rich==13.7.1
673
+ rmm==23.8.0
674
+ rope==1.12.0
675
+ rouge==1.0.1
676
+ rpds-py==0.16.2
677
+ rsa==4.9
678
+ ruamel-yaml-conda==0.15.100
679
+ ruamel.yaml.clib==0.2.7
680
+ ruamel.yaml==0.17.40
681
+ s2sphere==0.2.5
682
+ s3fs==2024.3.0
683
+ s3transfer==0.6.2
684
+ safetensors==0.4.2
685
+ scattertext==0.1.19
686
+ scikit-image==0.22.0
687
+ scikit-learn-intelex==2024.1.0
688
+ scikit-learn==1.2.2
689
+ scikit-multilearn==0.2.0
690
+ scikit-optimize==0.10.1
691
+ scikit-plot==0.3.7
692
+ scikit-surprise==1.1.3
693
+ scipy==1.11.4
694
+ scipy==1.12.0
695
+ seaborn==0.12.2
696
+ segment_anything==1.0
697
+ segregation==2.5
698
+ semver==3.0.2
699
+ sentencepiece==0.2.0
700
+ sentry-sdk==1.42.0
701
+ setproctitle==1.3.3
702
+ setuptools-git==1.2
703
+ setuptools-scm==8.0.4
704
+ setuptools==69.0.3
705
+ shap==0.44.1
706
+ shapely==2.0.3
707
+ shellingham==1.5.4
708
+ simpervisor==1.0.0
709
+ simplejson==3.19.2
710
+ six==1.16.0
711
+ sklearn-pandas==2.2.0
712
+ slicer==0.0.7
713
+ smart-open==6.4.0
714
+ smmap==5.0.1
715
+ sniffio==1.3.0
716
+ snowballstemmer==2.2.0
717
+ snuggs==1.4.7
718
+ sortedcontainers==2.4.0
719
+ soundfile==0.12.1
720
+ soupsieve==2.5
721
+ soxr==0.3.7
722
+ spacy-legacy==3.0.12
723
+ spacy-loggers==1.0.5
724
+ spacy==3.7.2
725
+ spaghetti==1.7.5.post1
726
+ spectral==0.23.1
727
+ spglm==1.1.0
728
+ sphinx-rtd-theme==0.2.4
729
+ spint==1.0.7
730
+ splot==1.1.5.post1
731
+ spopt==0.6.0
732
+ spreg==1.4.2
733
+ spvcm==0.3.0
734
+ sqlparse==0.4.4
735
+ squarify==0.4.3
736
+ srsly==2.4.8
737
+ stable-baselines3==2.1.0
738
+ stack-data==0.6.2
739
+ stack-data==0.6.3
740
+ stanio==0.3.0
741
+ starlette==0.32.0.post1
742
+ statsmodels==0.14.1
743
+ stemming==1.0.1
744
+ stop-words==2018.7.23
745
+ stopit==1.1.2
746
+ stumpy==1.12.0
747
+ sympy==1.12
748
+ tables==3.9.2
749
+ tabulate==0.9.0
750
+ tangled-up-in-unicode==0.2.0
751
+ tbb==2021.11.0
752
+ tblib==3.0.0
753
+ tenacity==8.2.3
754
+ tensorboard-data-server==0.7.2
755
+ tensorboard-plugin-profile==2.15.0
756
+ tensorboard==2.15.1
757
+ tensorboardX==2.6.2.2
758
+ tensorflow-cloud==0.1.16
759
+ tensorflow-datasets==4.9.4
760
+ tensorflow-decision-forests==1.8.1
761
+ tensorflow-estimator==2.15.0
762
+ tensorflow-hub==0.16.1
763
+ tensorflow-io-gcs-filesystem==0.35.0
764
+ tensorflow-io==0.35.0
765
+ tensorflow-metadata==0.14.0
766
+ tensorflow-probability==0.23.0
767
+ tensorflow-serving-api==2.14.1
768
+ tensorflow-text==2.15.0
769
+ tensorflow-transform==0.14.0
770
+ tensorflow==2.15.0
771
+ tensorpack==0.11
772
+ tensorstore==0.1.56
773
+ termcolor==2.4.0
774
+ terminado==0.18.0
775
+ testpath==0.6.0
776
+ text-unidecode==1.3
777
+ textblob==0.18.0.post0
778
+ texttable==1.7.0
779
+ tf_keras==2.15.1
780
+ tfp-nightly==0.24.0.dev0
781
+ thinc==8.2.2
782
+ threadpoolctl==3.2.0
783
+ tifffile==2023.12.9
784
+ timm==0.9.16
785
+ tinycss2==1.2.1
786
+ tobler==0.11.2
787
+ tokenizers==0.15.2
788
+ toml==0.10.2
789
+ tomli==2.0.1
790
+ tomlkit==0.12.4
791
+ toolz==0.12.1
792
+ torch==2.1.2
793
+ torchaudio==2.1.2
794
+ torchdata==0.7.1
795
+ torchinfo==1.8.0
796
+ torchmetrics==1.3.2
797
+ torchtext==0.16.2
798
+ torchvision==0.16.2
799
+ tornado==6.3.3
800
+ tqdm==4.66.1
801
+ traceml==1.0.8
802
+ traitlets==5.9.0
803
+ traittypes==0.2.1
804
+ transformers==4.38.2
805
+ treelite-runtime==3.2.0
806
+ treelite==3.2.0
807
+ trueskill==0.4.5
808
+ truststore==0.8.0
809
+ trx-python==0.2.9
810
+ tsfresh==0.20.2
811
+ typeguard==4.1.5
812
+ typer==0.9.0
813
+ types-python-dateutil==2.8.19.20240106
814
+ typing-inspect==0.9.0
815
+ typing-utils==0.1.0
816
+ typing_extensions==4.9.0
817
+ tzdata==2023.4
818
+ uc-micro-py==1.0.3
819
+ ucx-py==0.33.0
820
+ ujson==5.9.0
821
+ umap-learn==0.5.5
822
+ unicodedata2==15.1.0
823
+ update-checker==0.18.0
824
+ uri-template==1.3.0
825
+ uritemplate==3.0.1
826
+ urllib3==1.26.18
827
+ urllib3==2.1.0
828
+ urwid==2.6.9
829
+ urwid_readline==0.14
830
+ uvicorn==0.25.0
831
+ uvloop==0.19.0
832
+ vaex-astro==0.9.3
833
+ vaex-core==4.17.1
834
+ vaex-hdf5==0.14.1
835
+ vaex-jupyter==0.8.2
836
+ vaex-ml==0.18.3
837
+ vaex-server==0.9.0
838
+ vaex-viz==0.5.4
839
+ vaex==4.17.0
840
+ vec_noise==1.1.4
841
+ vecstack==0.4.0
842
+ virtualenv==20.21.0
843
+ visions==0.7.5
844
+ vowpalwabbit==9.9.0
845
+ vtk==9.3.0
846
+ wandb==0.16.4
847
+ wasabi==1.1.2
848
+ watchfiles==0.21.0
849
+ wavio==0.0.8
850
+ wcwidth==0.2.13
851
+ weasel==0.3.4
852
+ webcolors==1.13
853
+ webencodings==0.5.1
854
+ websocket-client==1.7.0
855
+ websockets==12.0
856
+ wfdb==4.1.2
857
+ whatthepatch==1.0.5
858
+ wheel==0.42.0
859
+ widgetsnbextension==3.6.6
860
+ witwidget==1.8.1
861
+ woodwork==0.29.0
862
+ wordcloud==1.9.3
863
+ wordsegment==1.3.1
864
+ wrapt==1.14.1
865
+ xarray-einstats==0.7.0
866
+ xarray==2024.2.0
867
+ xgboost==2.0.3
868
+ xvfbwrapper==0.2.9
869
+ xxhash==3.4.1
870
+ xyzservices==2023.10.1
871
+ y-py==0.6.2
872
+ yapf==0.40.2
873
+ yarl==1.9.3
874
+ yarl==1.9.4
875
+ ydata-profiling==4.6.4
876
+ yellowbrick==1.5
877
+ ypy-websocket==0.8.4
878
+ zict==3.0.0
879
+ zipp==3.17.0
880
+ zstandard==0.22.0
wandb/run-20240429_072705-0t7z1bit/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-29T07:27:06.362332",
5
+ "startedAt": "2024-04-29T07:27:05.562699",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "be592b10bdc8",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.132,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.132,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.132,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.132,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.132,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5598.571662902832
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240429_072705-0t7z1bit/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.0002, "train/grad_norm": 0.0017345616361126304, "train/learning_rate": 0.0, "train/epoch": 10.0, "train/global_step": 1890, "_timestamp": 1714379658.2116358, "_runtime": 4032.6411848068237, "_step": 381, "eval/loss": 0.0013308452907949686, "eval/Accuracy": 0.9996693121693122, "eval/F1": 0.9996629468127344, "eval/Precision": 0.9996612466124661, "eval/Recall": 0.999665327978581, "eval/runtime": 9.8512, "eval/samples_per_second": 306.967, "eval/steps_per_second": 9.643, "train/train_runtime": 4024.064, "train/train_samples_per_second": 15.03, "train/train_steps_per_second": 0.47, "train/total_flos": 3854057594526720.0, "train/train_loss": 0.025461199506686103}
wandb/run-20240429_072705-0t7z1bit/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240429_072705-0t7z1bit/logs/debug.log ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240429_072705-0t7z1bit/logs/debug.log
10
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240429_072705-0t7z1bit/logs/debug-internal.log
11
+ 2024-04-29 07:27:05,564 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7b7c6cebd510>
12
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-29 07:27:05,565 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-29 07:27:05,567 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-29 07:27:05,570 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-29 07:27:05,587 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-29 07:27:06,060 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-29 07:27:06,063 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-29 07:27:06,249 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-29 07:27:06,331 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-29 07:27:06,331 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-29 07:27:37,477 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-29 07:27:37,477 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-29 07:27:37,478 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-29 07:27:37,478 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-29 07:27:37,479 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-29 07:27:37,487 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'Install', 1: 'Compensation', 2: 'Change', 3: 'Summarize', 4: 'Terminate', 5: 'SBC'}, 'label2id': {'Install': 0, 'Compensation': 1, 'Change': 2, 'Summarize': 3, 'Terminate': 4, 'SBC': 5}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'google-bert/bert-base-uncased', 'transformers_version': '4.38.2', 'gradient_checkpointing': False, 'model_type': 'bert', 'vocab_size': 30522, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 10, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 10, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-29 08:32:33,725 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-29 08:32:33,725 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-29 08:33:23,264 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-29 08:34:18,223 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-29 08:34:18,224 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-29 08:34:43,696 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-29 08:34:43,697 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-04-29 08:34:43,697 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
+ 2024-04-29 08:35:51,215 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
41
+ 2024-04-29 08:35:51,218 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-04-29 08:35:51,218 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
43
+ 2024-04-29 08:35:55,649 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
44
+ 2024-04-29 08:35:58,171 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-04-29 08:35:58,172 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
46
+ 2024-04-29 08:36:23,490 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
47
+ 2024-04-29 08:36:23,493 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-04-29 08:36:23,493 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
49
+ 2024-04-29 08:36:24,474 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
50
+ 2024-04-29 08:36:25,328 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-04-29 08:36:25,329 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
52
+ 2024-04-29 08:36:47,420 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
53
+ 2024-04-29 08:36:47,423 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-29 08:36:47,423 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
55
+ 2024-04-29 08:36:48,363 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
56
+ 2024-04-29 08:36:48,681 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-29 08:36:48,682 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
58
+ 2024-04-29 08:37:26,761 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
59
+ 2024-04-29 08:37:27,658 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-04-29 08:37:27,658 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
61
+ 2024-04-29 08:37:31,816 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
62
+ 2024-04-29 08:37:31,849 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
63
+ 2024-04-29 08:37:31,849 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
64
+ 2024-04-29 08:37:48,262 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
65
+ 2024-04-29 08:37:48,997 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
66
+ 2024-04-29 08:37:48,997 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
67
+ 2024-04-29 08:38:08,126 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
68
+ 2024-04-29 08:38:09,034 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
69
+ 2024-04-29 08:38:09,034 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
70
+ 2024-04-29 08:38:12,014 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
71
+ 2024-04-29 08:38:12,041 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
72
+ 2024-04-29 08:38:12,041 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
73
+ 2024-04-29 08:38:24,975 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
74
+ 2024-04-29 08:38:25,120 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
75
+ 2024-04-29 08:38:25,121 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
76
+ 2024-04-29 08:39:43,660 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
77
+ 2024-04-29 08:39:43,985 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
78
+ 2024-04-29 08:39:43,985 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
79
+ 2024-04-29 08:39:57,431 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
80
+ 2024-04-29 08:39:57,728 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
81
+ 2024-04-29 08:39:57,728 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
82
+ 2024-04-29 08:40:00,496 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
83
+ 2024-04-29 08:40:00,705 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
84
+ 2024-04-29 08:40:00,705 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
85
+ 2024-04-29 08:40:04,198 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
86
+ 2024-04-29 08:40:04,256 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
87
+ 2024-04-29 08:40:04,256 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
88
+ 2024-04-29 08:40:07,104 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
89
+ 2024-04-29 08:40:07,185 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
90
+ 2024-04-29 08:40:07,186 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
91
+ 2024-04-29 08:40:13,433 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
92
+ 2024-04-29 08:40:13,507 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
93
+ 2024-04-29 08:40:13,507 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
94
+ 2024-04-29 08:40:17,190 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
95
+ 2024-04-29 08:40:17,288 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
96
+ 2024-04-29 08:40:17,288 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
97
+ 2024-04-29 08:40:23,262 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
98
+ 2024-04-29 08:40:23,268 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
99
+ 2024-04-29 08:40:23,268 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
100
+ 2024-04-29 08:40:33,903 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
101
+ 2024-04-29 08:40:36,980 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
102
+ 2024-04-29 08:40:36,980 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
103
+ 2024-04-29 08:40:43,896 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
104
+ 2024-04-29 08:40:47,028 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
105
+ 2024-04-29 08:40:47,028 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
106
+ 2024-04-29 08:45:12,809 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
107
+ 2024-04-29 08:45:12,981 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
108
+ 2024-04-29 08:45:12,981 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
109
+ 2024-04-29 08:52:31,675 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
wandb/run-20240429_072705-0t7z1bit/run-0t7z1bit.wandb ADDED
Binary file (360 kB). View file