Femboyuwu2000 commited on
Commit
5415c3f
1 Parent(s): f4089ce

bloomz_bl_story

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +1 -1
  3. adapter_config.json +1 -1
  4. adapter_model.safetensors +1 -1
  5. bloomz_dam_my/README.md +202 -0
  6. bloomz_dam_my/adapter_config.json +30 -0
  7. bloomz_dam_my/adapter_model.safetensors +3 -0
  8. bloomz_dam_my/special_tokens_map.json +17 -0
  9. bloomz_dam_my/spiece.model +3 -0
  10. bloomz_dam_my/tokenizer.json +3 -0
  11. bloomz_dam_my/tokenizer_config.json +39 -0
  12. bloomz_dam_my/training_args.bin +3 -0
  13. runs/Apr11_04-02-02_321ddc84c049/events.out.tfevents.1712808126.321ddc84c049.414.0 +3 -0
  14. runs/Apr11_04-04-49_321ddc84c049/events.out.tfevents.1712808294.321ddc84c049.740.0 +3 -0
  15. runs/Apr11_04-07-17_321ddc84c049/events.out.tfevents.1712808441.321ddc84c049.1050.0 +3 -0
  16. tokenizer.json +2 -2
  17. training_args.bin +1 -1
  18. wandb/debug-internal.log +0 -0
  19. wandb/debug.log +29 -30
  20. wandb/run-20240411_035909-21esl9hl/files/output.log +1 -0
  21. wandb/run-20240411_035909-21esl9hl/files/wandb-summary.json +1 -1
  22. wandb/run-20240411_035909-21esl9hl/logs/debug-internal.log +158 -0
  23. wandb/run-20240411_035909-21esl9hl/logs/debug.log +2 -0
  24. wandb/run-20240411_035909-21esl9hl/run-21esl9hl.wandb +0 -0
  25. wandb/run-20240411_040208-txpa3x4k/files/conda-environment.yaml +0 -0
  26. wandb/run-20240411_040208-txpa3x4k/files/config.yaml +701 -0
  27. wandb/run-20240411_040208-txpa3x4k/files/output.log +18 -0
  28. wandb/run-20240411_040208-txpa3x4k/files/requirements.txt +864 -0
  29. wandb/run-20240411_040208-txpa3x4k/files/wandb-metadata.json +66 -0
  30. wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json +1 -0
  31. wandb/run-20240411_040208-txpa3x4k/logs/debug-internal.log +273 -0
  32. wandb/run-20240411_040208-txpa3x4k/logs/debug.log +33 -0
  33. wandb/run-20240411_040208-txpa3x4k/run-txpa3x4k.wandb +0 -0
  34. wandb/run-20240411_040455-4dprer7c/files/conda-environment.yaml +0 -0
  35. wandb/run-20240411_040455-4dprer7c/files/config.yaml +701 -0
  36. wandb/run-20240411_040455-4dprer7c/files/output.log +12 -0
  37. wandb/run-20240411_040455-4dprer7c/files/requirements.txt +864 -0
  38. wandb/run-20240411_040455-4dprer7c/files/wandb-metadata.json +66 -0
  39. wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json +1 -0
  40. wandb/run-20240411_040455-4dprer7c/logs/debug-internal.log +254 -0
  41. wandb/run-20240411_040455-4dprer7c/logs/debug.log +33 -0
  42. wandb/run-20240411_040455-4dprer7c/run-4dprer7c.wandb +0 -0
  43. wandb/run-20240411_040722-rimyh9a1/files/conda-environment.yaml +0 -0
  44. wandb/run-20240411_040722-rimyh9a1/files/config.yaml +702 -0
  45. wandb/run-20240411_040722-rimyh9a1/files/output.log +0 -0
  46. wandb/run-20240411_040722-rimyh9a1/files/requirements.txt +864 -0
  47. wandb/run-20240411_040722-rimyh9a1/files/wandb-metadata.json +66 -0
  48. wandb/run-20240411_040722-rimyh9a1/files/wandb-summary.json +1 -0
  49. wandb/run-20240411_040722-rimyh9a1/logs/debug-internal.log +0 -0
  50. wandb/run-20240411_040722-rimyh9a1/logs/debug.log +31 -0
.gitattributes CHANGED
@@ -37,3 +37,4 @@ wandb/run-20240408_203054-ldiuneeg/run-ldiuneeg.wandb filter=lfs diff=lfs merge=
37
  wandb/run-20240410_010250-ft4a6i5j/run-ft4a6i5j.wandb filter=lfs diff=lfs merge=lfs -text
38
  wandb/run-20240410_115017-r9p6sok2/run-r9p6sok2.wandb filter=lfs diff=lfs merge=lfs -text
39
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
37
  wandb/run-20240410_010250-ft4a6i5j/run-ft4a6i5j.wandb filter=lfs diff=lfs merge=lfs -text
38
  wandb/run-20240410_115017-r9p6sok2/run-r9p6sok2.wandb filter=lfs diff=lfs merge=lfs -text
39
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
+ wandb/run-20240411_040722-rimyh9a1/run-rimyh9a1.wandb filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
- - training_steps: 10
46
 
47
  ### Training results
48
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
+ - training_steps: 20000
46
 
47
  ### Training results
48
 
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "lm_head",
24
  "query_key_value",
 
25
  "dense_h_to_4h"
26
  ],
27
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "query_key_value",
24
+ "lm_head",
25
  "dense_h_to_4h"
26
  ],
27
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9baf0d7e0bee813113ef40ed93ff8433a426745af90937aabc260a737c8f5277
3
  size 1042757704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ac730ac4465d0ec9fbe49ed9c5ad2a69ec11ddf9dc1d9cb0f5595f60817af39
3
  size 1042757704
bloomz_dam_my/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ base_model: bigscience/bloomz-560m
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.10.0
bloomz_dam_my/adapter_config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "bigscience/bloomz-560m",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.01,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "lm_head",
24
+ "query_key_value",
25
+ "dense_h_to_4h"
26
+ ],
27
+ "task_type": "CAUSAL_LM",
28
+ "use_dora": false,
29
+ "use_rslora": true
30
+ }
bloomz_dam_my/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baf0d7e0bee813113ef40ed93ff8433a426745af90937aabc260a737c8f5277
3
+ size 1042757704
bloomz_dam_my/special_tokens_map.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": "</s>",
10
+ "unk_token": {
11
+ "content": "<unk>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ }
bloomz_dam_my/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
bloomz_dam_my/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8af11292e1e8b6e7fc393100a39b873778106b40afa8ef4e7c198f0bb5c03f82
3
+ size 16315312
bloomz_dam_my/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<unk>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "additional_special_tokens": [],
29
+ "clean_up_tokenization_spaces": true,
30
+ "eos_token": "</s>",
31
+ "extra_ids": 0,
32
+ "max_length": 1024,
33
+ "model_max_length": 1000000000000000019884624838656,
34
+ "pad_token": "</s>",
35
+ "sp_model_kwargs": {},
36
+ "tokenizer_class": "T5Tokenizer",
37
+ "truncation": true,
38
+ "unk_token": "<unk>"
39
+ }
bloomz_dam_my/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fcb5d40d27446b473f3cf6a16217747341cc71590de85e4a2a9001a3be0bfca
3
+ size 4920
runs/Apr11_04-02-02_321ddc84c049/events.out.tfevents.1712808126.321ddc84c049.414.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dc0de91e150ffa3ae50d49ef29cc00f0d6cd90e2dfce554c500adb81011433e
3
+ size 15910
runs/Apr11_04-04-49_321ddc84c049/events.out.tfevents.1712808294.321ddc84c049.740.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77ccd6bb0384d7f0b531b4689a6d5091c5cf8ef5fe830dd33a06449fc0a4e6c6
3
+ size 15706
runs/Apr11_04-07-17_321ddc84c049/events.out.tfevents.1712808441.321ddc84c049.1050.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39731fc62d56057a7ac2a4a59ab5f57cc8eafb01f7eadaa3b6146959084621fe
3
+ size 233938
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8af11292e1e8b6e7fc393100a39b873778106b40afa8ef4e7c198f0bb5c03f82
3
- size 16315312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54e5c72a5ea09da48b2f316760b8bc5a445683ab9a5bc6b68db5d8db624ecceb
3
+ size 16315213
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fcb5d40d27446b473f3cf6a16217747341cc71590de85e4a2a9001a3be0bfca
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba44607bb84c5fd76d74e8d257eb1a18a8dceedf164c95e0e7b4eb4771e56e1
3
  size 4920
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,32 +1,31 @@
1
- 2024-04-11 03:59:09,934 INFO MainThread:145 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Configure stats pid to 145
3
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_035909-21esl9hl/logs/debug.log
11
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_035909-21esl9hl/logs/debug-internal.log
12
- 2024-04-11 03:59:09,935 INFO MainThread:145 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e1d3f588970>
13
- 2024-04-11 03:59:09,936 INFO MainThread:145 [wandb_init.py:init():567] calling init triggers
14
- 2024-04-11 03:59:09,936 INFO MainThread:145 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
  config: {}
16
- 2024-04-11 03:59:09,936 INFO MainThread:145 [wandb_init.py:init():617] starting backend
17
- 2024-04-11 03:59:09,936 INFO MainThread:145 [wandb_init.py:init():621] setting up manager
18
- 2024-04-11 03:59:09,938 INFO MainThread:145 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
- 2024-04-11 03:59:09,940 INFO MainThread:145 [wandb_init.py:init():629] backend started and connected
20
- 2024-04-11 03:59:09,953 INFO MainThread:145 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
- 2024-04-11 03:59:10,362 INFO MainThread:145 [wandb_init.py:init():721] updated telemetry
22
- 2024-04-11 03:59:10,366 INFO MainThread:145 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
- 2024-04-11 03:59:10,526 INFO MainThread:145 [wandb_run.py:_on_init():2344] communicating current version
24
- 2024-04-11 03:59:10,613 INFO MainThread:145 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
- 2024-04-11 03:59:10,614 INFO MainThread:145 [wandb_init.py:init():805] starting run threads in backend
27
- 2024-04-11 03:59:26,715 INFO MainThread:145 [wandb_run.py:_console_start():2323] atexit reg
28
- 2024-04-11 03:59:26,715 INFO MainThread:145 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
- 2024-04-11 03:59:26,716 INFO MainThread:145 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
- 2024-04-11 03:59:26,716 INFO MainThread:145 [wandb_run.py:_redirect():2268] Redirects installed.
31
- 2024-04-11 03:59:26,717 INFO MainThread:145 [wandb_init.py:init():848] run started, returning control to user process
32
- 2024-04-11 03:59:26,723 INFO MainThread:145 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_03-48-59_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
1
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Configure stats pid to 1050
3
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug.log
10
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug-internal.log
11
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7d563f5f5120>
12
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
 
14
  config: {}
15
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 04:07:22,847 INFO MainThread:1050 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 04:07:22,848 INFO MainThread:1050 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 04:07:22,865 INFO MainThread:1050 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 04:07:23,106 INFO MainThread:1050 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 04:07:23,108 INFO MainThread:1050 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 04:07:23,235 INFO MainThread:1050 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 04:07:23,323 INFO MainThread:1050 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-11 04:07:23,324 INFO MainThread:1050 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 04:07:39,319 INFO MainThread:1050 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 04:07:39,324 INFO MainThread:1050 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_04-07-17_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240411_035909-21esl9hl/files/output.log CHANGED
@@ -101,3 +101,4 @@ To disable this warning, you can either:
101
  - Avoid using `tokenizers` before the fork if possible
102
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
  /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
 
 
101
  - Avoid using `tokenizers` before the fork if possible
102
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
  /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
104
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
wandb/run-20240411_035909-21esl9hl/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"train_runtime": 14.0567, "train_samples_per_second": 2.846, "train_steps_per_second": 0.711, "total_flos": 28907911741440.0, "train_loss": 21.220675659179687, "train/epoch": 0.0, "train/global_step": 10, "_timestamp": 1712808008.9437401, "_runtime": 59.003005027770996, "_step": 0}
 
1
+ {"train_runtime": 14.0567, "train_samples_per_second": 2.846, "train_steps_per_second": 0.711, "total_flos": 28907911741440.0, "train_loss": 21.220675659179687, "train/epoch": 0.0, "train/global_step": 10, "_timestamp": 1712808008.9437401, "_runtime": 59.003005027770996, "_step": 0, "_wandb": {"runtime": 91}}
wandb/run-20240411_035909-21esl9hl/logs/debug-internal.log CHANGED
@@ -116,3 +116,161 @@ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after
116
  2024-04-11 04:00:27,786 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
117
  2024-04-11 04:00:32,787 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
118
  2024-04-11 04:00:37,788 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
116
  2024-04-11 04:00:27,786 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
117
  2024-04-11 04:00:32,787 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
118
  2024-04-11 04:00:37,788 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
119
+ 2024-04-11 04:00:40,648 DEBUG SenderThread:191 [sender.py:send():379] send: stats
120
+ 2024-04-11 04:00:42,143 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: stop_status
121
+ 2024-04-11 04:00:42,144 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: stop_status
122
+ 2024-04-11 04:00:42,184 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: internal_messages
123
+ 2024-04-11 04:00:42,200 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: pause
124
+ 2024-04-11 04:00:42,200 INFO HandlerThread:191 [handler.py:handle_request_pause():708] stopping system metrics thread
125
+ 2024-04-11 04:00:42,200 INFO HandlerThread:191 [system_monitor.py:finish():203] Stopping system monitor
126
+ 2024-04-11 04:00:42,200 DEBUG SystemMonitor:191 [system_monitor.py:_start():179] Finished system metrics aggregation loop
127
+ 2024-04-11 04:00:42,200 DEBUG SystemMonitor:191 [system_monitor.py:_start():183] Publishing last batch of metrics
128
+ 2024-04-11 04:00:42,202 INFO HandlerThread:191 [interfaces.py:finish():202] Joined cpu monitor
129
+ 2024-04-11 04:00:42,202 INFO HandlerThread:191 [interfaces.py:finish():202] Joined disk monitor
130
+ 2024-04-11 04:00:42,211 INFO HandlerThread:191 [interfaces.py:finish():202] Joined gpu monitor
131
+ 2024-04-11 04:00:42,212 INFO HandlerThread:191 [interfaces.py:finish():202] Joined memory monitor
132
+ 2024-04-11 04:00:42,212 INFO HandlerThread:191 [interfaces.py:finish():202] Joined network monitor
133
+ 2024-04-11 04:00:42,234 DEBUG SenderThread:191 [sender.py:send():379] send: stats
134
+ 2024-04-11 04:00:43,235 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-04-11 04:00:48,235 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
136
+ 2024-04-11 04:00:53,236 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
137
+ 2024-04-11 04:00:57,141 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: stop_status
138
+ 2024-04-11 04:00:57,142 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: stop_status
139
+ 2024-04-11 04:00:57,182 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: internal_messages
140
+ 2024-04-11 04:00:58,263 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
141
+ 2024-04-11 04:01:03,264 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
142
+ 2024-04-11 04:01:08,265 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
143
+ 2024-04-11 04:01:12,141 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: stop_status
144
+ 2024-04-11 04:01:12,142 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: stop_status
145
+ 2024-04-11 04:01:12,182 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: internal_messages
146
+ 2024-04-11 04:01:13,266 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
147
+ 2024-04-11 04:01:18,267 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
148
+ 2024-04-11 04:01:23,268 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
149
+ 2024-04-11 04:01:23,392 DEBUG SenderThread:191 [sender.py:send():379] send: exit
150
+ 2024-04-11 04:01:23,392 INFO SenderThread:191 [sender.py:send_exit():586] handling exit code: 0
151
+ 2024-04-11 04:01:23,392 INFO SenderThread:191 [sender.py:send_exit():588] handling runtime: 91
152
+ 2024-04-11 04:01:23,394 INFO SenderThread:191 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
153
+ 2024-04-11 04:01:23,394 INFO SenderThread:191 [sender.py:send_exit():594] send defer
154
+ 2024-04-11 04:01:23,394 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
155
+ 2024-04-11 04:01:23,394 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 0
156
+ 2024-04-11 04:01:23,395 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
157
+ 2024-04-11 04:01:23,395 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 0
158
+ 2024-04-11 04:01:23,395 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 1
159
+ 2024-04-11 04:01:23,395 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
160
+ 2024-04-11 04:01:23,395 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 1
161
+ 2024-04-11 04:01:23,395 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
162
+ 2024-04-11 04:01:23,395 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 1
163
+ 2024-04-11 04:01:23,395 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 2
164
+ 2024-04-11 04:01:23,395 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
165
+ 2024-04-11 04:01:23,396 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 2
166
+ 2024-04-11 04:01:23,396 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
167
+ 2024-04-11 04:01:23,396 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 2
168
+ 2024-04-11 04:01:23,396 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 3
169
+ 2024-04-11 04:01:23,396 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
170
+ 2024-04-11 04:01:23,396 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 3
171
+ 2024-04-11 04:01:23,396 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
172
+ 2024-04-11 04:01:23,396 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 3
173
+ 2024-04-11 04:01:23,396 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 4
174
+ 2024-04-11 04:01:23,396 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
175
+ 2024-04-11 04:01:23,397 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 4
176
+ 2024-04-11 04:01:23,397 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
177
+ 2024-04-11 04:01:23,397 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 4
178
+ 2024-04-11 04:01:23,397 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 5
179
+ 2024-04-11 04:01:23,397 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
180
+ 2024-04-11 04:01:23,397 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 5
181
+ 2024-04-11 04:01:23,397 DEBUG SenderThread:191 [sender.py:send():379] send: summary
182
+ 2024-04-11 04:01:23,398 INFO SenderThread:191 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
183
+ 2024-04-11 04:01:23,398 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
184
+ 2024-04-11 04:01:23,398 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 5
185
+ 2024-04-11 04:01:23,398 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 6
186
+ 2024-04-11 04:01:23,398 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
187
+ 2024-04-11 04:01:23,398 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 6
188
+ 2024-04-11 04:01:23,398 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
189
+ 2024-04-11 04:01:23,399 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 6
190
+ 2024-04-11 04:01:23,399 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 7
191
+ 2024-04-11 04:01:23,399 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-04-11 04:01:23,399 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
193
+ 2024-04-11 04:01:23,399 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 7
194
+ 2024-04-11 04:01:23,399 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
195
+ 2024-04-11 04:01:23,399 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 7
196
+ 2024-04-11 04:01:23,565 INFO Thread-12 :191 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/wandb-summary.json
197
+ 2024-04-11 04:01:24,392 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: poll_exit
198
+ 2024-04-11 04:01:24,961 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 8
199
+ 2024-04-11 04:01:24,961 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: poll_exit
200
+ 2024-04-11 04:01:24,961 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
201
+ 2024-04-11 04:01:24,961 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 8
202
+ 2024-04-11 04:01:24,962 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
203
+ 2024-04-11 04:01:24,962 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 8
204
+ 2024-04-11 04:01:24,962 INFO SenderThread:191 [job_builder.py:build():318] Attempting to build job artifact
205
+ 2024-04-11 04:01:24,964 INFO SenderThread:191 [job_builder.py:_get_source_type():466] no source found
206
+ 2024-04-11 04:01:24,964 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 9
207
+ 2024-04-11 04:01:24,964 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
208
+ 2024-04-11 04:01:24,964 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 9
209
+ 2024-04-11 04:01:24,964 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
210
+ 2024-04-11 04:01:24,964 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 9
211
+ 2024-04-11 04:01:24,964 INFO SenderThread:191 [dir_watcher.py:finish():358] shutting down directory watcher
212
+ 2024-04-11 04:01:25,393 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: poll_exit
213
+ 2024-04-11 04:01:25,566 INFO SenderThread:191 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/output.log
214
+ 2024-04-11 04:01:25,567 INFO SenderThread:191 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files
215
+ 2024-04-11 04:01:25,567 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/config.yaml config.yaml
216
+ 2024-04-11 04:01:25,567 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/requirements.txt requirements.txt
217
+ 2024-04-11 04:01:25,568 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/conda-environment.yaml conda-environment.yaml
218
+ 2024-04-11 04:01:25,574 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/wandb-summary.json wandb-summary.json
219
+ 2024-04-11 04:01:25,574 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/wandb-metadata.json wandb-metadata.json
220
+ 2024-04-11 04:01:25,574 INFO SenderThread:191 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/output.log output.log
221
+ 2024-04-11 04:01:25,575 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 10
222
+ 2024-04-11 04:01:25,575 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: poll_exit
223
+ 2024-04-11 04:01:25,578 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
224
+ 2024-04-11 04:01:25,578 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 10
225
+ 2024-04-11 04:01:25,579 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
226
+ 2024-04-11 04:01:25,579 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 10
227
+ 2024-04-11 04:01:25,579 INFO SenderThread:191 [file_pusher.py:finish():172] shutting down file pusher
228
+ 2024-04-11 04:01:25,725 INFO wandb-upload_1:191 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/requirements.txt
229
+ 2024-04-11 04:01:25,772 INFO wandb-upload_0:191 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/config.yaml
230
+ 2024-04-11 04:01:25,777 INFO wandb-upload_3:191 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/output.log
231
+ 2024-04-11 04:01:25,846 INFO wandb-upload_2:191 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_035909-21esl9hl/files/wandb-summary.json
232
+ 2024-04-11 04:01:26,046 INFO Thread-11 (_thread_body):191 [sender.py:transition_state():614] send defer: 11
233
+ 2024-04-11 04:01:26,046 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
234
+ 2024-04-11 04:01:26,047 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 11
235
+ 2024-04-11 04:01:26,047 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
236
+ 2024-04-11 04:01:26,047 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 11
237
+ 2024-04-11 04:01:26,047 INFO SenderThread:191 [file_pusher.py:join():178] waiting for file pusher
238
+ 2024-04-11 04:01:26,047 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 12
239
+ 2024-04-11 04:01:26,048 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
240
+ 2024-04-11 04:01:26,048 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 12
241
+ 2024-04-11 04:01:26,048 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
242
+ 2024-04-11 04:01:26,048 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 12
243
+ 2024-04-11 04:01:26,048 INFO SenderThread:191 [file_stream.py:finish():614] file stream finish called
244
+ 2024-04-11 04:01:26,118 INFO SenderThread:191 [file_stream.py:finish():618] file stream finish is done
245
+ 2024-04-11 04:01:26,118 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 13
246
+ 2024-04-11 04:01:26,118 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
247
+ 2024-04-11 04:01:26,118 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 13
248
+ 2024-04-11 04:01:26,119 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
249
+ 2024-04-11 04:01:26,119 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 13
250
+ 2024-04-11 04:01:26,119 INFO SenderThread:191 [sender.py:transition_state():614] send defer: 14
251
+ 2024-04-11 04:01:26,119 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: defer
252
+ 2024-04-11 04:01:26,119 INFO HandlerThread:191 [handler.py:handle_request_defer():172] handle defer: 14
253
+ 2024-04-11 04:01:26,120 DEBUG SenderThread:191 [sender.py:send():379] send: final
254
+ 2024-04-11 04:01:26,120 DEBUG SenderThread:191 [sender.py:send():379] send: footer
255
+ 2024-04-11 04:01:26,120 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: defer
256
+ 2024-04-11 04:01:26,120 INFO SenderThread:191 [sender.py:send_request_defer():610] handle sender defer: 14
257
+ 2024-04-11 04:01:26,121 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: poll_exit
258
+ 2024-04-11 04:01:26,121 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: poll_exit
259
+ 2024-04-11 04:01:26,122 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: poll_exit
260
+ 2024-04-11 04:01:26,122 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: poll_exit
261
+ 2024-04-11 04:01:26,123 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: server_info
262
+ 2024-04-11 04:01:26,123 DEBUG SenderThread:191 [sender.py:send_request():406] send_request: server_info
263
+ 2024-04-11 04:01:26,127 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: get_summary
264
+ 2024-04-11 04:01:26,127 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: sampled_history
265
+ 2024-04-11 04:01:26,128 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: internal_messages
266
+ 2024-04-11 04:01:26,139 INFO MainThread:191 [wandb_run.py:_footer_history_summary_info():3920] rendering history
267
+ 2024-04-11 04:01:26,139 INFO MainThread:191 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
268
+ 2024-04-11 04:01:26,139 INFO MainThread:191 [wandb_run.py:_footer_sync_info():3879] logging synced files
269
+ 2024-04-11 04:01:26,140 DEBUG HandlerThread:191 [handler.py:handle_request():146] handle_request: shutdown
270
+ 2024-04-11 04:01:26,140 INFO HandlerThread:191 [handler.py:finish():866] shutting down handler
271
+ 2024-04-11 04:01:27,084 WARNING StreamThr :191 [internal.py:is_dead():414] Internal process exiting, parent pid 145 disappeared
272
+ 2024-04-11 04:01:27,084 ERROR StreamThr :191 [internal.py:wandb_internal():152] Internal process shutdown.
273
+ 2024-04-11 04:01:27,123 INFO WriterThread:191 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_035909-21esl9hl/run-21esl9hl.wandb
274
+ 2024-04-11 04:01:27,139 INFO SenderThread:191 [sender.py:finish():1546] shutting down sender
275
+ 2024-04-11 04:01:27,139 INFO SenderThread:191 [file_pusher.py:finish():172] shutting down file pusher
276
+ 2024-04-11 04:01:27,139 INFO SenderThread:191 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_035909-21esl9hl/logs/debug.log CHANGED
@@ -30,3 +30,5 @@ config: {}
30
  2024-04-11 03:59:26,716 INFO MainThread:145 [wandb_run.py:_redirect():2268] Redirects installed.
31
  2024-04-11 03:59:26,717 INFO MainThread:145 [wandb_init.py:init():848] run started, returning control to user process
32
  2024-04-11 03:59:26,723 INFO MainThread:145 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_03-48-59_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
 
 
30
  2024-04-11 03:59:26,716 INFO MainThread:145 [wandb_run.py:_redirect():2268] Redirects installed.
31
  2024-04-11 03:59:26,717 INFO MainThread:145 [wandb_init.py:init():848] run started, returning control to user process
32
  2024-04-11 03:59:26,723 INFO MainThread:145 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_03-48-59_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-11 04:00:42,199 INFO MainThread:145 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-11 04:00:42,200 INFO MainThread:145 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_035909-21esl9hl/run-21esl9hl.wandb CHANGED
Binary files a/wandb/run-20240411_035909-21esl9hl/run-21esl9hl.wandb and b/wandb/run-20240411_035909-21esl9hl/run-21esl9hl.wandb differ
 
wandb/run-20240411_040208-txpa3x4k/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_040208-txpa3x4k/files/config.yaml ADDED
@@ -0,0 +1,701 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712808128.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ vocab_size:
78
+ desc: null
79
+ value: 250880
80
+ hidden_size:
81
+ desc: null
82
+ value: 1024
83
+ n_layer:
84
+ desc: null
85
+ value: 24
86
+ n_head:
87
+ desc: null
88
+ value: 16
89
+ layer_norm_epsilon:
90
+ desc: null
91
+ value: 1.0e-05
92
+ initializer_range:
93
+ desc: null
94
+ value: 0.02
95
+ use_cache:
96
+ desc: null
97
+ value: false
98
+ pretraining_tp:
99
+ desc: null
100
+ value: 1
101
+ apply_residual_connection_post_layernorm:
102
+ desc: null
103
+ value: false
104
+ hidden_dropout:
105
+ desc: null
106
+ value: 0.0
107
+ attention_dropout:
108
+ desc: null
109
+ value: 0.0
110
+ bos_token_id:
111
+ desc: null
112
+ value: 1
113
+ eos_token_id:
114
+ desc: null
115
+ value: 2
116
+ slow_but_exact:
117
+ desc: null
118
+ value: false
119
+ return_dict:
120
+ desc: null
121
+ value: true
122
+ output_hidden_states:
123
+ desc: null
124
+ value: false
125
+ output_attentions:
126
+ desc: null
127
+ value: false
128
+ torchscript:
129
+ desc: null
130
+ value: false
131
+ torch_dtype:
132
+ desc: null
133
+ value: null
134
+ use_bfloat16:
135
+ desc: null
136
+ value: false
137
+ tf_legacy_loss:
138
+ desc: null
139
+ value: false
140
+ pruned_heads:
141
+ desc: null
142
+ value: {}
143
+ tie_word_embeddings:
144
+ desc: null
145
+ value: true
146
+ chunk_size_feed_forward:
147
+ desc: null
148
+ value: 0
149
+ is_encoder_decoder:
150
+ desc: null
151
+ value: false
152
+ is_decoder:
153
+ desc: null
154
+ value: false
155
+ cross_attention_hidden_size:
156
+ desc: null
157
+ value: null
158
+ add_cross_attention:
159
+ desc: null
160
+ value: false
161
+ tie_encoder_decoder:
162
+ desc: null
163
+ value: false
164
+ max_length:
165
+ desc: null
166
+ value: 20
167
+ min_length:
168
+ desc: null
169
+ value: 0
170
+ do_sample:
171
+ desc: null
172
+ value: false
173
+ early_stopping:
174
+ desc: null
175
+ value: false
176
+ num_beams:
177
+ desc: null
178
+ value: 1
179
+ num_beam_groups:
180
+ desc: null
181
+ value: 1
182
+ diversity_penalty:
183
+ desc: null
184
+ value: 0.0
185
+ temperature:
186
+ desc: null
187
+ value: 1.0
188
+ top_k:
189
+ desc: null
190
+ value: 50
191
+ top_p:
192
+ desc: null
193
+ value: 1.0
194
+ typical_p:
195
+ desc: null
196
+ value: 1.0
197
+ repetition_penalty:
198
+ desc: null
199
+ value: 1.0
200
+ length_penalty:
201
+ desc: null
202
+ value: 1.0
203
+ no_repeat_ngram_size:
204
+ desc: null
205
+ value: 0
206
+ encoder_no_repeat_ngram_size:
207
+ desc: null
208
+ value: 0
209
+ bad_words_ids:
210
+ desc: null
211
+ value: null
212
+ num_return_sequences:
213
+ desc: null
214
+ value: 1
215
+ output_scores:
216
+ desc: null
217
+ value: false
218
+ return_dict_in_generate:
219
+ desc: null
220
+ value: false
221
+ forced_bos_token_id:
222
+ desc: null
223
+ value: null
224
+ forced_eos_token_id:
225
+ desc: null
226
+ value: null
227
+ remove_invalid_values:
228
+ desc: null
229
+ value: false
230
+ exponential_decay_length_penalty:
231
+ desc: null
232
+ value: null
233
+ suppress_tokens:
234
+ desc: null
235
+ value: null
236
+ begin_suppress_tokens:
237
+ desc: null
238
+ value: null
239
+ architectures:
240
+ desc: null
241
+ value:
242
+ - BloomForCausalLM
243
+ finetuning_task:
244
+ desc: null
245
+ value: null
246
+ id2label:
247
+ desc: null
248
+ value:
249
+ '0': LABEL_0
250
+ '1': LABEL_1
251
+ label2id:
252
+ desc: null
253
+ value:
254
+ LABEL_0: 0
255
+ LABEL_1: 1
256
+ tokenizer_class:
257
+ desc: null
258
+ value: null
259
+ prefix:
260
+ desc: null
261
+ value: null
262
+ pad_token_id:
263
+ desc: null
264
+ value: 3
265
+ sep_token_id:
266
+ desc: null
267
+ value: null
268
+ decoder_start_token_id:
269
+ desc: null
270
+ value: null
271
+ task_specific_params:
272
+ desc: null
273
+ value: null
274
+ problem_type:
275
+ desc: null
276
+ value: null
277
+ _name_or_path:
278
+ desc: null
279
+ value: bigscience/bloomz-560m
280
+ transformers_version:
281
+ desc: null
282
+ value: 4.39.3
283
+ attention_softmax_in_fp32:
284
+ desc: null
285
+ value: true
286
+ bias_dropout_fusion:
287
+ desc: null
288
+ value: true
289
+ unk_token_id:
290
+ desc: null
291
+ value: 0
292
+ masked_softmax_fusion:
293
+ desc: null
294
+ value: true
295
+ model_type:
296
+ desc: null
297
+ value: bloom
298
+ n_inner:
299
+ desc: null
300
+ value: null
301
+ offset_alibi:
302
+ desc: null
303
+ value: 100
304
+ seq_length:
305
+ desc: null
306
+ value: 2048
307
+ skip_bias_add:
308
+ desc: null
309
+ value: true
310
+ skip_bias_add_qkv:
311
+ desc: null
312
+ value: false
313
+ quantization_config:
314
+ desc: null
315
+ value:
316
+ quant_method: QuantizationMethod.BITS_AND_BYTES
317
+ _load_in_8bit: false
318
+ _load_in_4bit: true
319
+ llm_int8_threshold: 6.0
320
+ llm_int8_skip_modules: null
321
+ llm_int8_enable_fp32_cpu_offload: false
322
+ llm_int8_has_fp16_weight: false
323
+ bnb_4bit_quant_type: nf4
324
+ bnb_4bit_use_double_quant: false
325
+ bnb_4bit_compute_dtype: float16
326
+ bnb_4bit_quant_storage: uint8
327
+ load_in_4bit: true
328
+ load_in_8bit: false
329
+ output_dir:
330
+ desc: null
331
+ value: /kaggle/working/
332
+ overwrite_output_dir:
333
+ desc: null
334
+ value: false
335
+ do_train:
336
+ desc: null
337
+ value: false
338
+ do_eval:
339
+ desc: null
340
+ value: false
341
+ do_predict:
342
+ desc: null
343
+ value: false
344
+ evaluation_strategy:
345
+ desc: null
346
+ value: 'no'
347
+ prediction_loss_only:
348
+ desc: null
349
+ value: false
350
+ per_device_train_batch_size:
351
+ desc: null
352
+ value: 8
353
+ per_device_eval_batch_size:
354
+ desc: null
355
+ value: 8
356
+ per_gpu_train_batch_size:
357
+ desc: null
358
+ value: null
359
+ per_gpu_eval_batch_size:
360
+ desc: null
361
+ value: null
362
+ gradient_accumulation_steps:
363
+ desc: null
364
+ value: 1
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: null
368
+ eval_delay:
369
+ desc: null
370
+ value: 0
371
+ learning_rate:
372
+ desc: null
373
+ value: 3.0e-05
374
+ weight_decay:
375
+ desc: null
376
+ value: 0.001
377
+ adam_beta1:
378
+ desc: null
379
+ value: 0.9
380
+ adam_beta2:
381
+ desc: null
382
+ value: 0.999
383
+ adam_epsilon:
384
+ desc: null
385
+ value: 1.0e-08
386
+ max_grad_norm:
387
+ desc: null
388
+ value: 0.3
389
+ num_train_epochs:
390
+ desc: null
391
+ value: 5
392
+ max_steps:
393
+ desc: null
394
+ value: 7000
395
+ lr_scheduler_type:
396
+ desc: null
397
+ value: cosine
398
+ lr_scheduler_kwargs:
399
+ desc: null
400
+ value: {}
401
+ warmup_ratio:
402
+ desc: null
403
+ value: 0.03
404
+ warmup_steps:
405
+ desc: null
406
+ value: 0
407
+ log_level:
408
+ desc: null
409
+ value: passive
410
+ log_level_replica:
411
+ desc: null
412
+ value: warning
413
+ log_on_each_node:
414
+ desc: null
415
+ value: true
416
+ logging_dir:
417
+ desc: null
418
+ value: /kaggle/working/runs/Apr11_04-02-02_321ddc84c049
419
+ logging_strategy:
420
+ desc: null
421
+ value: steps
422
+ logging_first_step:
423
+ desc: null
424
+ value: false
425
+ logging_steps:
426
+ desc: null
427
+ value: 20
428
+ logging_nan_inf_filter:
429
+ desc: null
430
+ value: true
431
+ save_strategy:
432
+ desc: null
433
+ value: steps
434
+ save_steps:
435
+ desc: null
436
+ value: 20
437
+ save_total_limit:
438
+ desc: null
439
+ value: 1
440
+ save_safetensors:
441
+ desc: null
442
+ value: true
443
+ save_on_each_node:
444
+ desc: null
445
+ value: false
446
+ save_only_model:
447
+ desc: null
448
+ value: false
449
+ no_cuda:
450
+ desc: null
451
+ value: false
452
+ use_cpu:
453
+ desc: null
454
+ value: false
455
+ use_mps_device:
456
+ desc: null
457
+ value: false
458
+ seed:
459
+ desc: null
460
+ value: 42
461
+ data_seed:
462
+ desc: null
463
+ value: null
464
+ jit_mode_eval:
465
+ desc: null
466
+ value: false
467
+ use_ipex:
468
+ desc: null
469
+ value: false
470
+ bf16:
471
+ desc: null
472
+ value: false
473
+ fp16:
474
+ desc: null
475
+ value: false
476
+ fp16_opt_level:
477
+ desc: null
478
+ value: O1
479
+ half_precision_backend:
480
+ desc: null
481
+ value: auto
482
+ bf16_full_eval:
483
+ desc: null
484
+ value: false
485
+ fp16_full_eval:
486
+ desc: null
487
+ value: false
488
+ tf32:
489
+ desc: null
490
+ value: null
491
+ local_rank:
492
+ desc: null
493
+ value: 0
494
+ ddp_backend:
495
+ desc: null
496
+ value: null
497
+ tpu_num_cores:
498
+ desc: null
499
+ value: null
500
+ tpu_metrics_debug:
501
+ desc: null
502
+ value: false
503
+ debug:
504
+ desc: null
505
+ value: []
506
+ dataloader_drop_last:
507
+ desc: null
508
+ value: false
509
+ eval_steps:
510
+ desc: null
511
+ value: null
512
+ dataloader_num_workers:
513
+ desc: null
514
+ value: 8
515
+ dataloader_prefetch_factor:
516
+ desc: null
517
+ value: null
518
+ past_index:
519
+ desc: null
520
+ value: -1
521
+ run_name:
522
+ desc: null
523
+ value: /kaggle/working/
524
+ disable_tqdm:
525
+ desc: null
526
+ value: false
527
+ remove_unused_columns:
528
+ desc: null
529
+ value: true
530
+ label_names:
531
+ desc: null
532
+ value: null
533
+ load_best_model_at_end:
534
+ desc: null
535
+ value: false
536
+ metric_for_best_model:
537
+ desc: null
538
+ value: null
539
+ greater_is_better:
540
+ desc: null
541
+ value: null
542
+ ignore_data_skip:
543
+ desc: null
544
+ value: false
545
+ fsdp:
546
+ desc: null
547
+ value: []
548
+ fsdp_min_num_params:
549
+ desc: null
550
+ value: 0
551
+ fsdp_config:
552
+ desc: null
553
+ value:
554
+ min_num_params: 0
555
+ xla: false
556
+ xla_fsdp_v2: false
557
+ xla_fsdp_grad_ckpt: false
558
+ fsdp_transformer_layer_cls_to_wrap:
559
+ desc: null
560
+ value: null
561
+ accelerator_config:
562
+ desc: null
563
+ value:
564
+ split_batches: false
565
+ dispatch_batches: null
566
+ even_batches: true
567
+ use_seedable_sampler: true
568
+ deepspeed:
569
+ desc: null
570
+ value: null
571
+ label_smoothing_factor:
572
+ desc: null
573
+ value: 0.0
574
+ optim:
575
+ desc: null
576
+ value: paged_adamw_32bit
577
+ optim_args:
578
+ desc: null
579
+ value: null
580
+ adafactor:
581
+ desc: null
582
+ value: false
583
+ group_by_length:
584
+ desc: null
585
+ value: false
586
+ length_column_name:
587
+ desc: null
588
+ value: length
589
+ report_to:
590
+ desc: null
591
+ value:
592
+ - tensorboard
593
+ - wandb
594
+ ddp_find_unused_parameters:
595
+ desc: null
596
+ value: null
597
+ ddp_bucket_cap_mb:
598
+ desc: null
599
+ value: null
600
+ ddp_broadcast_buffers:
601
+ desc: null
602
+ value: null
603
+ dataloader_pin_memory:
604
+ desc: null
605
+ value: true
606
+ dataloader_persistent_workers:
607
+ desc: null
608
+ value: false
609
+ skip_memory_metrics:
610
+ desc: null
611
+ value: true
612
+ use_legacy_prediction_loop:
613
+ desc: null
614
+ value: false
615
+ push_to_hub:
616
+ desc: null
617
+ value: false
618
+ resume_from_checkpoint:
619
+ desc: null
620
+ value: null
621
+ hub_model_id:
622
+ desc: null
623
+ value: null
624
+ hub_strategy:
625
+ desc: null
626
+ value: every_save
627
+ hub_token:
628
+ desc: null
629
+ value: <HUB_TOKEN>
630
+ hub_private_repo:
631
+ desc: null
632
+ value: false
633
+ hub_always_push:
634
+ desc: null
635
+ value: false
636
+ gradient_checkpointing:
637
+ desc: null
638
+ value: true
639
+ gradient_checkpointing_kwargs:
640
+ desc: null
641
+ value: null
642
+ include_inputs_for_metrics:
643
+ desc: null
644
+ value: false
645
+ fp16_backend:
646
+ desc: null
647
+ value: auto
648
+ push_to_hub_model_id:
649
+ desc: null
650
+ value: null
651
+ push_to_hub_organization:
652
+ desc: null
653
+ value: null
654
+ push_to_hub_token:
655
+ desc: null
656
+ value: <PUSH_TO_HUB_TOKEN>
657
+ mp_parameters:
658
+ desc: null
659
+ value: ''
660
+ auto_find_batch_size:
661
+ desc: null
662
+ value: true
663
+ full_determinism:
664
+ desc: null
665
+ value: false
666
+ torchdynamo:
667
+ desc: null
668
+ value: null
669
+ ray_scope:
670
+ desc: null
671
+ value: last
672
+ ddp_timeout:
673
+ desc: null
674
+ value: 1800
675
+ torch_compile:
676
+ desc: null
677
+ value: false
678
+ torch_compile_backend:
679
+ desc: null
680
+ value: null
681
+ torch_compile_mode:
682
+ desc: null
683
+ value: null
684
+ dispatch_batches:
685
+ desc: null
686
+ value: null
687
+ split_batches:
688
+ desc: null
689
+ value: null
690
+ include_tokens_per_second:
691
+ desc: null
692
+ value: false
693
+ include_num_input_tokens_seen:
694
+ desc: null
695
+ value: false
696
+ neftune_noise_alpha:
697
+ desc: null
698
+ value: null
699
+ optim_target_modules:
700
+ desc: null
701
+ value: null
wandb/run-20240411_040208-txpa3x4k/files/output.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
6
+ warnings.warn('Was asked to gather along dimension 0, but all '
7
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
8
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
9
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
10
+ warnings.warn(
11
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
12
+ warnings.warn('Was asked to gather along dimension 0, but all '
13
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
14
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
15
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
16
+ warnings.warn(
17
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
18
+ warnings.warn('Was asked to gather along dimension 0, but all '
wandb/run-20240411_040208-txpa3x4k/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_040208-txpa3x4k/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T04:02:08.804959",
5
+ "startedAt": "2024-04-11T04:02:08.129415",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "321ddc84c049",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.19,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.19,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.19,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.19,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.19,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5566.961982727051
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 27.1232, "train/grad_norm": 212.8720245361328, "train/learning_rate": 5.7142857142857145e-06, "train/epoch": 0.0, "train/global_step": 40, "_timestamp": 1712808237.341076, "_runtime": 109.20561385154724, "_step": 1, "_wandb": {"runtime": 120}}
wandb/run-20240411_040208-txpa3x4k/logs/debug-internal.log ADDED
@@ -0,0 +1,273 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 04:02:08,136 INFO StreamThr :455 [internal.py:wandb_internal():86] W&B internal server running at pid: 455, started at: 2024-04-11 04:02:08.135570
2
+ 2024-04-11 04:02:08,137 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 04:02:08,532 INFO WriterThread:455 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/run-txpa3x4k.wandb
4
+ 2024-04-11 04:02:08,532 DEBUG SenderThread:455 [sender.py:send():379] send: header
5
+ 2024-04-11 04:02:08,535 DEBUG SenderThread:455 [sender.py:send():379] send: run
6
+ 2024-04-11 04:02:08,682 INFO SenderThread:455 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files
7
+ 2024-04-11 04:02:08,683 INFO SenderThread:455 [sender.py:_start_run_threads():1124] run started: txpa3x4k with start time 1712808128.135462
8
+ 2024-04-11 04:02:08,691 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 04:02:08,691 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 04:02:08,778 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 04:02:08,790 DEBUG HandlerThread:455 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 04:02:08,790 DEBUG HandlerThread:455 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 04:02:08,790 INFO HandlerThread:455 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 04:02:08,790 INFO SystemMonitor:455 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 04:02:08,791 INFO SystemMonitor:455 [interfaces.py:start():190] Started cpu monitoring
16
+ 2024-04-11 04:02:08,791 INFO SystemMonitor:455 [interfaces.py:start():190] Started disk monitoring
17
+ 2024-04-11 04:02:08,792 INFO HandlerThread:455 [system_monitor.py:probe():214] Collecting system info
18
+ 2024-04-11 04:02:08,792 INFO SystemMonitor:455 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 04:02:08,793 INFO SystemMonitor:455 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 04:02:08,793 INFO SystemMonitor:455 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 04:02:08,804 DEBUG HandlerThread:455 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 04:02:08,807 DEBUG HandlerThread:455 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 04:02:08,807 DEBUG HandlerThread:455 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 04:02:08,807 DEBUG HandlerThread:455 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T04:02:08.804959', 'startedAt': '2024-04-11T04:02:08.129415', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '321ddc84c049', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.19, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5566.961982727051}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 04:02:08,807 INFO HandlerThread:455 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 04:02:08,807 INFO HandlerThread:455 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 04:02:08,807 DEBUG HandlerThread:455 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 04:02:09,685 INFO Thread-12 :455 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/conda-environment.yaml
29
+ 2024-04-11 04:02:23,823 ERROR HandlerThread:455 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 04:02:23,824 DEBUG HandlerThread:455 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 04:02:23,825 INFO HandlerThread:455 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 04:02:23,830 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 04:02:23,831 DEBUG SenderThread:455 [sender.py:send():379] send: files
49
+ 2024-04-11 04:02:23,831 INFO SenderThread:455 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 04:02:24,033 INFO wandb-upload_0:455 [upload_job.py:push():131] Uploaded file /tmp/tmp1s3x48hcwandb/4qm5hpvk-wandb-metadata.json
51
+ 2024-04-11 04:02:24,688 INFO Thread-12 :455 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-metadata.json
52
+ 2024-04-11 04:02:24,755 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 04:02:24,756 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 04:02:24,757 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
55
+ 2024-04-11 04:02:24,758 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
56
+ 2024-04-11 04:02:24,795 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
57
+ 2024-04-11 04:02:24,805 DEBUG SenderThread:455 [sender.py:send():379] send: telemetry
58
+ 2024-04-11 04:02:24,805 DEBUG SenderThread:455 [sender.py:send():379] send: config
59
+ 2024-04-11 04:02:24,807 DEBUG SenderThread:455 [sender.py:send():379] send: metric
60
+ 2024-04-11 04:02:24,808 DEBUG SenderThread:455 [sender.py:send():379] send: telemetry
61
+ 2024-04-11 04:02:24,808 DEBUG SenderThread:455 [sender.py:send():379] send: metric
62
+ 2024-04-11 04:02:24,808 WARNING SenderThread:455 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
63
+ 2024-04-11 04:02:24,808 DEBUG SenderThread:455 [sender.py:send():379] send: telemetry
64
+ 2024-04-11 04:02:25,688 INFO Thread-12 :455 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
65
+ 2024-04-11 04:02:25,689 INFO Thread-12 :455 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/requirements.txt
66
+ 2024-04-11 04:02:27,689 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
67
+ 2024-04-11 04:02:29,171 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 04:02:29,690 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
69
+ 2024-04-11 04:02:34,172 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-11 04:02:39,178 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 04:02:39,693 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/config.yaml
72
+ 2024-04-11 04:02:40,355 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
73
+ 2024-04-11 04:02:40,355 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
74
+ 2024-04-11 04:02:40,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
75
+ 2024-04-11 04:02:44,408 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-11 04:02:49,409 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-04-11 04:02:54,410 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
78
+ 2024-04-11 04:02:55,351 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
79
+ 2024-04-11 04:02:55,352 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
80
+ 2024-04-11 04:02:55,392 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
81
+ 2024-04-11 04:03:00,394 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-11 04:03:05,395 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 04:03:08,794 DEBUG SystemMonitor:455 [system_monitor.py:_start():172] Starting system metrics aggregation loop
84
+ 2024-04-11 04:03:08,795 DEBUG SenderThread:455 [sender.py:send():379] send: stats
85
+ 2024-04-11 04:03:10,353 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
86
+ 2024-04-11 04:03:10,353 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
87
+ 2024-04-11 04:03:10,394 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
88
+ 2024-04-11 04:03:10,420 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-04-11 04:03:12,619 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: partial_history
90
+ 2024-04-11 04:03:12,622 DEBUG SenderThread:455 [sender.py:send():379] send: metric
91
+ 2024-04-11 04:03:12,623 DEBUG SenderThread:455 [sender.py:send():379] send: metric
92
+ 2024-04-11 04:03:12,623 DEBUG SenderThread:455 [sender.py:send():379] send: metric
93
+ 2024-04-11 04:03:12,623 DEBUG SenderThread:455 [sender.py:send():379] send: metric
94
+ 2024-04-11 04:03:12,623 DEBUG SenderThread:455 [sender.py:send():379] send: history
95
+ 2024-04-11 04:03:12,623 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: summary_record
96
+ 2024-04-11 04:03:12,623 INFO SenderThread:455 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
97
+ 2024-04-11 04:03:12,709 INFO Thread-12 :455 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json
98
+ 2024-04-11 04:03:15,592 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-11 04:03:15,710 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
100
+ 2024-04-11 04:03:17,711 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
101
+ 2024-04-11 04:03:20,730 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
102
+ 2024-04-11 04:03:25,368 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
103
+ 2024-04-11 04:03:25,368 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
104
+ 2024-04-11 04:03:25,409 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
105
+ 2024-04-11 04:03:26,469 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-04-11 04:03:31,470 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-04-11 04:03:36,471 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-04-11 04:03:38,796 DEBUG SenderThread:455 [sender.py:send():379] send: stats
109
+ 2024-04-11 04:03:40,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
110
+ 2024-04-11 04:03:40,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
111
+ 2024-04-11 04:03:40,359 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
112
+ 2024-04-11 04:03:42,393 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-04-11 04:03:42,721 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/config.yaml
114
+ 2024-04-11 04:03:47,526 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-04-11 04:03:52,527 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 04:03:55,357 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
117
+ 2024-04-11 04:03:55,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
118
+ 2024-04-11 04:03:55,359 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
119
+ 2024-04-11 04:03:57,341 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: partial_history
120
+ 2024-04-11 04:03:57,342 DEBUG SenderThread:455 [sender.py:send():379] send: history
121
+ 2024-04-11 04:03:57,343 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: summary_record
122
+ 2024-04-11 04:03:57,345 INFO SenderThread:455 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
123
+ 2024-04-11 04:03:57,727 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json
124
+ 2024-04-11 04:03:58,374 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
125
+ 2024-04-11 04:03:59,728 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
126
+ 2024-04-11 04:04:01,729 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
127
+ 2024-04-11 04:04:03,528 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-04-11 04:04:03,729 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
129
+ 2024-04-11 04:04:08,529 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
130
+ 2024-04-11 04:04:08,797 DEBUG SenderThread:455 [sender.py:send():379] send: stats
131
+ 2024-04-11 04:04:09,554 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: pause
132
+ 2024-04-11 04:04:09,554 INFO HandlerThread:455 [handler.py:handle_request_pause():708] stopping system metrics thread
133
+ 2024-04-11 04:04:09,554 INFO HandlerThread:455 [system_monitor.py:finish():203] Stopping system monitor
134
+ 2024-04-11 04:04:09,554 DEBUG SystemMonitor:455 [system_monitor.py:_start():179] Finished system metrics aggregation loop
135
+ 2024-04-11 04:04:09,554 DEBUG SystemMonitor:455 [system_monitor.py:_start():183] Publishing last batch of metrics
136
+ 2024-04-11 04:04:09,555 INFO HandlerThread:455 [interfaces.py:finish():202] Joined cpu monitor
137
+ 2024-04-11 04:04:09,555 INFO HandlerThread:455 [interfaces.py:finish():202] Joined disk monitor
138
+ 2024-04-11 04:04:09,567 INFO HandlerThread:455 [interfaces.py:finish():202] Joined gpu monitor
139
+ 2024-04-11 04:04:09,567 INFO HandlerThread:455 [interfaces.py:finish():202] Joined memory monitor
140
+ 2024-04-11 04:04:09,567 INFO HandlerThread:455 [interfaces.py:finish():202] Joined network monitor
141
+ 2024-04-11 04:04:09,568 DEBUG SenderThread:455 [sender.py:send():379] send: stats
142
+ 2024-04-11 04:04:10,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
143
+ 2024-04-11 04:04:10,358 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: stop_status
144
+ 2024-04-11 04:04:10,359 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: stop_status
145
+ 2024-04-11 04:04:14,201 DEBUG SenderThread:455 [sender.py:send():379] send: exit
146
+ 2024-04-11 04:04:14,201 INFO SenderThread:455 [sender.py:send_exit():586] handling exit code: 0
147
+ 2024-04-11 04:04:14,201 INFO SenderThread:455 [sender.py:send_exit():588] handling runtime: 120
148
+ 2024-04-11 04:04:14,201 INFO SenderThread:455 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
149
+ 2024-04-11 04:04:14,202 INFO SenderThread:455 [sender.py:send_exit():594] send defer
150
+ 2024-04-11 04:04:14,202 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
151
+ 2024-04-11 04:04:14,202 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 0
152
+ 2024-04-11 04:04:14,202 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
153
+ 2024-04-11 04:04:14,202 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
154
+ 2024-04-11 04:04:14,202 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 0
155
+ 2024-04-11 04:04:14,202 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 1
156
+ 2024-04-11 04:04:14,203 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
157
+ 2024-04-11 04:04:14,203 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 1
158
+ 2024-04-11 04:04:14,203 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
159
+ 2024-04-11 04:04:14,203 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 1
160
+ 2024-04-11 04:04:14,203 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 2
161
+ 2024-04-11 04:04:14,203 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
162
+ 2024-04-11 04:04:14,203 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 2
163
+ 2024-04-11 04:04:14,203 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
164
+ 2024-04-11 04:04:14,204 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 2
165
+ 2024-04-11 04:04:14,204 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 3
166
+ 2024-04-11 04:04:14,204 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
167
+ 2024-04-11 04:04:14,204 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 3
168
+ 2024-04-11 04:04:14,204 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
169
+ 2024-04-11 04:04:14,204 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 3
170
+ 2024-04-11 04:04:14,204 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 4
171
+ 2024-04-11 04:04:14,204 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
172
+ 2024-04-11 04:04:14,204 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 4
173
+ 2024-04-11 04:04:14,204 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
174
+ 2024-04-11 04:04:14,205 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 4
175
+ 2024-04-11 04:04:14,205 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 5
176
+ 2024-04-11 04:04:14,205 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
177
+ 2024-04-11 04:04:14,205 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 5
178
+ 2024-04-11 04:04:14,205 DEBUG SenderThread:455 [sender.py:send():379] send: summary
179
+ 2024-04-11 04:04:14,206 INFO SenderThread:455 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
180
+ 2024-04-11 04:04:14,206 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
181
+ 2024-04-11 04:04:14,206 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 5
182
+ 2024-04-11 04:04:14,206 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 6
183
+ 2024-04-11 04:04:14,206 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
184
+ 2024-04-11 04:04:14,206 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 6
185
+ 2024-04-11 04:04:14,206 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
186
+ 2024-04-11 04:04:14,206 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 6
187
+ 2024-04-11 04:04:14,206 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 7
188
+ 2024-04-11 04:04:14,207 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: status_report
189
+ 2024-04-11 04:04:14,207 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
190
+ 2024-04-11 04:04:14,207 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 7
191
+ 2024-04-11 04:04:14,207 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
192
+ 2024-04-11 04:04:14,207 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 7
193
+ 2024-04-11 04:04:14,733 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json
194
+ 2024-04-11 04:04:14,942 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 8
195
+ 2024-04-11 04:04:14,942 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
196
+ 2024-04-11 04:04:14,943 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 8
197
+ 2024-04-11 04:04:14,943 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
198
+ 2024-04-11 04:04:14,943 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 8
199
+ 2024-04-11 04:04:14,943 INFO SenderThread:455 [job_builder.py:build():318] Attempting to build job artifact
200
+ 2024-04-11 04:04:14,945 INFO SenderThread:455 [job_builder.py:_get_source_type():466] no source found
201
+ 2024-04-11 04:04:14,945 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 9
202
+ 2024-04-11 04:04:14,945 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
203
+ 2024-04-11 04:04:14,945 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 9
204
+ 2024-04-11 04:04:14,945 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
205
+ 2024-04-11 04:04:14,946 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 9
206
+ 2024-04-11 04:04:14,946 INFO SenderThread:455 [dir_watcher.py:finish():358] shutting down directory watcher
207
+ 2024-04-11 04:04:15,201 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: poll_exit
208
+ 2024-04-11 04:04:15,734 INFO Thread-12 :455 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
209
+ 2024-04-11 04:04:15,735 INFO SenderThread:455 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files
210
+ 2024-04-11 04:04:15,735 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/config.yaml config.yaml
211
+ 2024-04-11 04:04:15,735 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/requirements.txt requirements.txt
212
+ 2024-04-11 04:04:15,738 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/conda-environment.yaml conda-environment.yaml
213
+ 2024-04-11 04:04:15,739 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json wandb-summary.json
214
+ 2024-04-11 04:04:15,739 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-metadata.json wandb-metadata.json
215
+ 2024-04-11 04:04:15,739 INFO SenderThread:455 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log output.log
216
+ 2024-04-11 04:04:15,742 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 10
217
+ 2024-04-11 04:04:15,742 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: poll_exit
218
+ 2024-04-11 04:04:15,743 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
219
+ 2024-04-11 04:04:15,743 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 10
220
+ 2024-04-11 04:04:15,749 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
221
+ 2024-04-11 04:04:15,749 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 10
222
+ 2024-04-11 04:04:15,749 INFO SenderThread:455 [file_pusher.py:finish():172] shutting down file pusher
223
+ 2024-04-11 04:04:15,892 INFO wandb-upload_0:455 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/config.yaml
224
+ 2024-04-11 04:04:15,944 INFO wandb-upload_1:455 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/requirements.txt
225
+ 2024-04-11 04:04:15,965 INFO wandb-upload_3:455 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/output.log
226
+ 2024-04-11 04:04:15,966 INFO wandb-upload_2:455 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040208-txpa3x4k/files/wandb-summary.json
227
+ 2024-04-11 04:04:16,167 INFO Thread-11 (_thread_body):455 [sender.py:transition_state():614] send defer: 11
228
+ 2024-04-11 04:04:16,167 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
229
+ 2024-04-11 04:04:16,168 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 11
230
+ 2024-04-11 04:04:16,168 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
231
+ 2024-04-11 04:04:16,168 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 11
232
+ 2024-04-11 04:04:16,168 INFO SenderThread:455 [file_pusher.py:join():178] waiting for file pusher
233
+ 2024-04-11 04:04:16,169 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 12
234
+ 2024-04-11 04:04:16,169 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
235
+ 2024-04-11 04:04:16,169 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 12
236
+ 2024-04-11 04:04:16,169 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
237
+ 2024-04-11 04:04:16,169 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 12
238
+ 2024-04-11 04:04:16,169 INFO SenderThread:455 [file_stream.py:finish():614] file stream finish called
239
+ 2024-04-11 04:04:16,202 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: poll_exit
240
+ 2024-04-11 04:04:16,257 INFO SenderThread:455 [file_stream.py:finish():618] file stream finish is done
241
+ 2024-04-11 04:04:16,257 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 13
242
+ 2024-04-11 04:04:16,257 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: poll_exit
243
+ 2024-04-11 04:04:16,258 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
244
+ 2024-04-11 04:04:16,258 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 13
245
+ 2024-04-11 04:04:16,258 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
246
+ 2024-04-11 04:04:16,258 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 13
247
+ 2024-04-11 04:04:16,258 INFO SenderThread:455 [sender.py:transition_state():614] send defer: 14
248
+ 2024-04-11 04:04:16,258 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: defer
249
+ 2024-04-11 04:04:16,259 INFO HandlerThread:455 [handler.py:handle_request_defer():172] handle defer: 14
250
+ 2024-04-11 04:04:16,259 DEBUG SenderThread:455 [sender.py:send():379] send: final
251
+ 2024-04-11 04:04:16,259 DEBUG SenderThread:455 [sender.py:send():379] send: footer
252
+ 2024-04-11 04:04:16,259 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: defer
253
+ 2024-04-11 04:04:16,259 INFO SenderThread:455 [sender.py:send_request_defer():610] handle sender defer: 14
254
+ 2024-04-11 04:04:16,260 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: poll_exit
255
+ 2024-04-11 04:04:16,260 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: poll_exit
256
+ 2024-04-11 04:04:16,261 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: poll_exit
257
+ 2024-04-11 04:04:16,261 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: poll_exit
258
+ 2024-04-11 04:04:16,262 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: server_info
259
+ 2024-04-11 04:04:16,262 DEBUG SenderThread:455 [sender.py:send_request():406] send_request: server_info
260
+ 2024-04-11 04:04:16,265 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: get_summary
261
+ 2024-04-11 04:04:16,266 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: sampled_history
262
+ 2024-04-11 04:04:16,267 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: internal_messages
263
+ 2024-04-11 04:04:16,277 INFO MainThread:455 [wandb_run.py:_footer_history_summary_info():3920] rendering history
264
+ 2024-04-11 04:04:16,277 INFO MainThread:455 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
265
+ 2024-04-11 04:04:16,278 INFO MainThread:455 [wandb_run.py:_footer_sync_info():3879] logging synced files
266
+ 2024-04-11 04:04:16,278 DEBUG HandlerThread:455 [handler.py:handle_request():146] handle_request: shutdown
267
+ 2024-04-11 04:04:16,278 INFO HandlerThread:455 [handler.py:finish():866] shutting down handler
268
+ 2024-04-11 04:04:17,262 INFO WriterThread:455 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_040208-txpa3x4k/run-txpa3x4k.wandb
269
+ 2024-04-11 04:04:17,267 WARNING StreamThr :455 [internal.py:is_dead():414] Internal process exiting, parent pid 414 disappeared
270
+ 2024-04-11 04:04:17,268 ERROR StreamThr :455 [internal.py:wandb_internal():152] Internal process shutdown.
271
+ 2024-04-11 04:04:17,277 INFO SenderThread:455 [sender.py:finish():1546] shutting down sender
272
+ 2024-04-11 04:04:17,277 INFO SenderThread:455 [file_pusher.py:finish():172] shutting down file pusher
273
+ 2024-04-11 04:04:17,277 INFO SenderThread:455 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_040208-txpa3x4k/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Configure stats pid to 414
3
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_040208-txpa3x4k/logs/debug.log
10
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_040208-txpa3x4k/logs/debug-internal.log
11
+ 2024-04-11 04:02:08,131 INFO MainThread:414 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x79289283df90>
12
+ 2024-04-11 04:02:08,132 INFO MainThread:414 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 04:02:08,132 INFO MainThread:414 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 04:02:08,132 INFO MainThread:414 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 04:02:08,132 INFO MainThread:414 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 04:02:08,134 INFO MainThread:414 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 04:02:08,135 INFO MainThread:414 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 04:02:08,148 INFO MainThread:414 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 04:02:08,531 INFO MainThread:414 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 04:02:08,534 INFO MainThread:414 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 04:02:08,691 INFO MainThread:414 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 04:02:08,772 INFO MainThread:414 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 04:02:08,772 INFO MainThread:414 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 04:02:24,756 INFO MainThread:414 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 04:02:24,757 INFO MainThread:414 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 04:02:24,757 INFO MainThread:414 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 04:02:24,758 INFO MainThread:414 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 04:02:24,759 INFO MainThread:414 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 04:02:24,765 INFO MainThread:414 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 7000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_04-02-02_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-11 04:04:09,553 INFO MainThread:414 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-11 04:04:09,553 INFO MainThread:414 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_040208-txpa3x4k/run-txpa3x4k.wandb ADDED
Binary file (15 kB). View file
 
wandb/run-20240411_040455-4dprer7c/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_040455-4dprer7c/files/config.yaml ADDED
@@ -0,0 +1,701 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712808295.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ vocab_size:
78
+ desc: null
79
+ value: 250880
80
+ hidden_size:
81
+ desc: null
82
+ value: 1024
83
+ n_layer:
84
+ desc: null
85
+ value: 24
86
+ n_head:
87
+ desc: null
88
+ value: 16
89
+ layer_norm_epsilon:
90
+ desc: null
91
+ value: 1.0e-05
92
+ initializer_range:
93
+ desc: null
94
+ value: 0.02
95
+ use_cache:
96
+ desc: null
97
+ value: false
98
+ pretraining_tp:
99
+ desc: null
100
+ value: 1
101
+ apply_residual_connection_post_layernorm:
102
+ desc: null
103
+ value: false
104
+ hidden_dropout:
105
+ desc: null
106
+ value: 0.0
107
+ attention_dropout:
108
+ desc: null
109
+ value: 0.0
110
+ bos_token_id:
111
+ desc: null
112
+ value: 1
113
+ eos_token_id:
114
+ desc: null
115
+ value: 2
116
+ slow_but_exact:
117
+ desc: null
118
+ value: false
119
+ return_dict:
120
+ desc: null
121
+ value: true
122
+ output_hidden_states:
123
+ desc: null
124
+ value: false
125
+ output_attentions:
126
+ desc: null
127
+ value: false
128
+ torchscript:
129
+ desc: null
130
+ value: false
131
+ torch_dtype:
132
+ desc: null
133
+ value: null
134
+ use_bfloat16:
135
+ desc: null
136
+ value: false
137
+ tf_legacy_loss:
138
+ desc: null
139
+ value: false
140
+ pruned_heads:
141
+ desc: null
142
+ value: {}
143
+ tie_word_embeddings:
144
+ desc: null
145
+ value: true
146
+ chunk_size_feed_forward:
147
+ desc: null
148
+ value: 0
149
+ is_encoder_decoder:
150
+ desc: null
151
+ value: false
152
+ is_decoder:
153
+ desc: null
154
+ value: false
155
+ cross_attention_hidden_size:
156
+ desc: null
157
+ value: null
158
+ add_cross_attention:
159
+ desc: null
160
+ value: false
161
+ tie_encoder_decoder:
162
+ desc: null
163
+ value: false
164
+ max_length:
165
+ desc: null
166
+ value: 20
167
+ min_length:
168
+ desc: null
169
+ value: 0
170
+ do_sample:
171
+ desc: null
172
+ value: false
173
+ early_stopping:
174
+ desc: null
175
+ value: false
176
+ num_beams:
177
+ desc: null
178
+ value: 1
179
+ num_beam_groups:
180
+ desc: null
181
+ value: 1
182
+ diversity_penalty:
183
+ desc: null
184
+ value: 0.0
185
+ temperature:
186
+ desc: null
187
+ value: 1.0
188
+ top_k:
189
+ desc: null
190
+ value: 50
191
+ top_p:
192
+ desc: null
193
+ value: 1.0
194
+ typical_p:
195
+ desc: null
196
+ value: 1.0
197
+ repetition_penalty:
198
+ desc: null
199
+ value: 1.0
200
+ length_penalty:
201
+ desc: null
202
+ value: 1.0
203
+ no_repeat_ngram_size:
204
+ desc: null
205
+ value: 0
206
+ encoder_no_repeat_ngram_size:
207
+ desc: null
208
+ value: 0
209
+ bad_words_ids:
210
+ desc: null
211
+ value: null
212
+ num_return_sequences:
213
+ desc: null
214
+ value: 1
215
+ output_scores:
216
+ desc: null
217
+ value: false
218
+ return_dict_in_generate:
219
+ desc: null
220
+ value: false
221
+ forced_bos_token_id:
222
+ desc: null
223
+ value: null
224
+ forced_eos_token_id:
225
+ desc: null
226
+ value: null
227
+ remove_invalid_values:
228
+ desc: null
229
+ value: false
230
+ exponential_decay_length_penalty:
231
+ desc: null
232
+ value: null
233
+ suppress_tokens:
234
+ desc: null
235
+ value: null
236
+ begin_suppress_tokens:
237
+ desc: null
238
+ value: null
239
+ architectures:
240
+ desc: null
241
+ value:
242
+ - BloomForCausalLM
243
+ finetuning_task:
244
+ desc: null
245
+ value: null
246
+ id2label:
247
+ desc: null
248
+ value:
249
+ '0': LABEL_0
250
+ '1': LABEL_1
251
+ label2id:
252
+ desc: null
253
+ value:
254
+ LABEL_0: 0
255
+ LABEL_1: 1
256
+ tokenizer_class:
257
+ desc: null
258
+ value: null
259
+ prefix:
260
+ desc: null
261
+ value: null
262
+ pad_token_id:
263
+ desc: null
264
+ value: 3
265
+ sep_token_id:
266
+ desc: null
267
+ value: null
268
+ decoder_start_token_id:
269
+ desc: null
270
+ value: null
271
+ task_specific_params:
272
+ desc: null
273
+ value: null
274
+ problem_type:
275
+ desc: null
276
+ value: null
277
+ _name_or_path:
278
+ desc: null
279
+ value: bigscience/bloomz-560m
280
+ transformers_version:
281
+ desc: null
282
+ value: 4.39.3
283
+ attention_softmax_in_fp32:
284
+ desc: null
285
+ value: true
286
+ bias_dropout_fusion:
287
+ desc: null
288
+ value: true
289
+ unk_token_id:
290
+ desc: null
291
+ value: 0
292
+ masked_softmax_fusion:
293
+ desc: null
294
+ value: true
295
+ model_type:
296
+ desc: null
297
+ value: bloom
298
+ n_inner:
299
+ desc: null
300
+ value: null
301
+ offset_alibi:
302
+ desc: null
303
+ value: 100
304
+ seq_length:
305
+ desc: null
306
+ value: 2048
307
+ skip_bias_add:
308
+ desc: null
309
+ value: true
310
+ skip_bias_add_qkv:
311
+ desc: null
312
+ value: false
313
+ quantization_config:
314
+ desc: null
315
+ value:
316
+ quant_method: QuantizationMethod.BITS_AND_BYTES
317
+ _load_in_8bit: false
318
+ _load_in_4bit: true
319
+ llm_int8_threshold: 6.0
320
+ llm_int8_skip_modules: null
321
+ llm_int8_enable_fp32_cpu_offload: false
322
+ llm_int8_has_fp16_weight: false
323
+ bnb_4bit_quant_type: nf4
324
+ bnb_4bit_use_double_quant: false
325
+ bnb_4bit_compute_dtype: float16
326
+ bnb_4bit_quant_storage: uint8
327
+ load_in_4bit: true
328
+ load_in_8bit: false
329
+ output_dir:
330
+ desc: null
331
+ value: /kaggle/working/
332
+ overwrite_output_dir:
333
+ desc: null
334
+ value: false
335
+ do_train:
336
+ desc: null
337
+ value: false
338
+ do_eval:
339
+ desc: null
340
+ value: false
341
+ do_predict:
342
+ desc: null
343
+ value: false
344
+ evaluation_strategy:
345
+ desc: null
346
+ value: 'no'
347
+ prediction_loss_only:
348
+ desc: null
349
+ value: false
350
+ per_device_train_batch_size:
351
+ desc: null
352
+ value: 8
353
+ per_device_eval_batch_size:
354
+ desc: null
355
+ value: 8
356
+ per_gpu_train_batch_size:
357
+ desc: null
358
+ value: null
359
+ per_gpu_eval_batch_size:
360
+ desc: null
361
+ value: null
362
+ gradient_accumulation_steps:
363
+ desc: null
364
+ value: 1
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: null
368
+ eval_delay:
369
+ desc: null
370
+ value: 0
371
+ learning_rate:
372
+ desc: null
373
+ value: 3.0e-05
374
+ weight_decay:
375
+ desc: null
376
+ value: 0.001
377
+ adam_beta1:
378
+ desc: null
379
+ value: 0.9
380
+ adam_beta2:
381
+ desc: null
382
+ value: 0.999
383
+ adam_epsilon:
384
+ desc: null
385
+ value: 1.0e-08
386
+ max_grad_norm:
387
+ desc: null
388
+ value: 0.3
389
+ num_train_epochs:
390
+ desc: null
391
+ value: 5
392
+ max_steps:
393
+ desc: null
394
+ value: 18000
395
+ lr_scheduler_type:
396
+ desc: null
397
+ value: cosine
398
+ lr_scheduler_kwargs:
399
+ desc: null
400
+ value: {}
401
+ warmup_ratio:
402
+ desc: null
403
+ value: 0.03
404
+ warmup_steps:
405
+ desc: null
406
+ value: 0
407
+ log_level:
408
+ desc: null
409
+ value: passive
410
+ log_level_replica:
411
+ desc: null
412
+ value: warning
413
+ log_on_each_node:
414
+ desc: null
415
+ value: true
416
+ logging_dir:
417
+ desc: null
418
+ value: /kaggle/working/runs/Apr11_04-04-49_321ddc84c049
419
+ logging_strategy:
420
+ desc: null
421
+ value: steps
422
+ logging_first_step:
423
+ desc: null
424
+ value: false
425
+ logging_steps:
426
+ desc: null
427
+ value: 20
428
+ logging_nan_inf_filter:
429
+ desc: null
430
+ value: true
431
+ save_strategy:
432
+ desc: null
433
+ value: steps
434
+ save_steps:
435
+ desc: null
436
+ value: 20
437
+ save_total_limit:
438
+ desc: null
439
+ value: 1
440
+ save_safetensors:
441
+ desc: null
442
+ value: true
443
+ save_on_each_node:
444
+ desc: null
445
+ value: false
446
+ save_only_model:
447
+ desc: null
448
+ value: false
449
+ no_cuda:
450
+ desc: null
451
+ value: false
452
+ use_cpu:
453
+ desc: null
454
+ value: false
455
+ use_mps_device:
456
+ desc: null
457
+ value: false
458
+ seed:
459
+ desc: null
460
+ value: 42
461
+ data_seed:
462
+ desc: null
463
+ value: null
464
+ jit_mode_eval:
465
+ desc: null
466
+ value: false
467
+ use_ipex:
468
+ desc: null
469
+ value: false
470
+ bf16:
471
+ desc: null
472
+ value: false
473
+ fp16:
474
+ desc: null
475
+ value: false
476
+ fp16_opt_level:
477
+ desc: null
478
+ value: O1
479
+ half_precision_backend:
480
+ desc: null
481
+ value: auto
482
+ bf16_full_eval:
483
+ desc: null
484
+ value: false
485
+ fp16_full_eval:
486
+ desc: null
487
+ value: false
488
+ tf32:
489
+ desc: null
490
+ value: null
491
+ local_rank:
492
+ desc: null
493
+ value: 0
494
+ ddp_backend:
495
+ desc: null
496
+ value: null
497
+ tpu_num_cores:
498
+ desc: null
499
+ value: null
500
+ tpu_metrics_debug:
501
+ desc: null
502
+ value: false
503
+ debug:
504
+ desc: null
505
+ value: []
506
+ dataloader_drop_last:
507
+ desc: null
508
+ value: false
509
+ eval_steps:
510
+ desc: null
511
+ value: null
512
+ dataloader_num_workers:
513
+ desc: null
514
+ value: 8
515
+ dataloader_prefetch_factor:
516
+ desc: null
517
+ value: null
518
+ past_index:
519
+ desc: null
520
+ value: -1
521
+ run_name:
522
+ desc: null
523
+ value: /kaggle/working/
524
+ disable_tqdm:
525
+ desc: null
526
+ value: false
527
+ remove_unused_columns:
528
+ desc: null
529
+ value: true
530
+ label_names:
531
+ desc: null
532
+ value: null
533
+ load_best_model_at_end:
534
+ desc: null
535
+ value: false
536
+ metric_for_best_model:
537
+ desc: null
538
+ value: null
539
+ greater_is_better:
540
+ desc: null
541
+ value: null
542
+ ignore_data_skip:
543
+ desc: null
544
+ value: false
545
+ fsdp:
546
+ desc: null
547
+ value: []
548
+ fsdp_min_num_params:
549
+ desc: null
550
+ value: 0
551
+ fsdp_config:
552
+ desc: null
553
+ value:
554
+ min_num_params: 0
555
+ xla: false
556
+ xla_fsdp_v2: false
557
+ xla_fsdp_grad_ckpt: false
558
+ fsdp_transformer_layer_cls_to_wrap:
559
+ desc: null
560
+ value: null
561
+ accelerator_config:
562
+ desc: null
563
+ value:
564
+ split_batches: false
565
+ dispatch_batches: null
566
+ even_batches: true
567
+ use_seedable_sampler: true
568
+ deepspeed:
569
+ desc: null
570
+ value: null
571
+ label_smoothing_factor:
572
+ desc: null
573
+ value: 0.0
574
+ optim:
575
+ desc: null
576
+ value: paged_adamw_32bit
577
+ optim_args:
578
+ desc: null
579
+ value: null
580
+ adafactor:
581
+ desc: null
582
+ value: false
583
+ group_by_length:
584
+ desc: null
585
+ value: false
586
+ length_column_name:
587
+ desc: null
588
+ value: length
589
+ report_to:
590
+ desc: null
591
+ value:
592
+ - tensorboard
593
+ - wandb
594
+ ddp_find_unused_parameters:
595
+ desc: null
596
+ value: null
597
+ ddp_bucket_cap_mb:
598
+ desc: null
599
+ value: null
600
+ ddp_broadcast_buffers:
601
+ desc: null
602
+ value: null
603
+ dataloader_pin_memory:
604
+ desc: null
605
+ value: true
606
+ dataloader_persistent_workers:
607
+ desc: null
608
+ value: false
609
+ skip_memory_metrics:
610
+ desc: null
611
+ value: true
612
+ use_legacy_prediction_loop:
613
+ desc: null
614
+ value: false
615
+ push_to_hub:
616
+ desc: null
617
+ value: false
618
+ resume_from_checkpoint:
619
+ desc: null
620
+ value: null
621
+ hub_model_id:
622
+ desc: null
623
+ value: null
624
+ hub_strategy:
625
+ desc: null
626
+ value: every_save
627
+ hub_token:
628
+ desc: null
629
+ value: <HUB_TOKEN>
630
+ hub_private_repo:
631
+ desc: null
632
+ value: false
633
+ hub_always_push:
634
+ desc: null
635
+ value: false
636
+ gradient_checkpointing:
637
+ desc: null
638
+ value: true
639
+ gradient_checkpointing_kwargs:
640
+ desc: null
641
+ value: null
642
+ include_inputs_for_metrics:
643
+ desc: null
644
+ value: false
645
+ fp16_backend:
646
+ desc: null
647
+ value: auto
648
+ push_to_hub_model_id:
649
+ desc: null
650
+ value: null
651
+ push_to_hub_organization:
652
+ desc: null
653
+ value: null
654
+ push_to_hub_token:
655
+ desc: null
656
+ value: <PUSH_TO_HUB_TOKEN>
657
+ mp_parameters:
658
+ desc: null
659
+ value: ''
660
+ auto_find_batch_size:
661
+ desc: null
662
+ value: true
663
+ full_determinism:
664
+ desc: null
665
+ value: false
666
+ torchdynamo:
667
+ desc: null
668
+ value: null
669
+ ray_scope:
670
+ desc: null
671
+ value: last
672
+ ddp_timeout:
673
+ desc: null
674
+ value: 1800
675
+ torch_compile:
676
+ desc: null
677
+ value: false
678
+ torch_compile_backend:
679
+ desc: null
680
+ value: null
681
+ torch_compile_mode:
682
+ desc: null
683
+ value: null
684
+ dispatch_batches:
685
+ desc: null
686
+ value: null
687
+ split_batches:
688
+ desc: null
689
+ value: null
690
+ include_tokens_per_second:
691
+ desc: null
692
+ value: false
693
+ include_num_input_tokens_seen:
694
+ desc: null
695
+ value: false
696
+ neftune_noise_alpha:
697
+ desc: null
698
+ value: null
699
+ optim_target_modules:
700
+ desc: null
701
+ value: null
wandb/run-20240411_040455-4dprer7c/files/output.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
6
+ warnings.warn('Was asked to gather along dimension 0, but all '
7
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
8
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
9
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
10
+ warnings.warn(
11
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
12
+ warnings.warn('Was asked to gather along dimension 0, but all '
wandb/run-20240411_040455-4dprer7c/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_040455-4dprer7c/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T04:04:56.368183",
5
+ "startedAt": "2024-04-11T04:04:55.701503",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "321ddc84c049",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.19,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.19,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.19,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.19,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.19,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5566.962100982666
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 30.0974, "train/grad_norm": 258.19573974609375, "train/learning_rate": 1.111111111111111e-06, "train/epoch": 0.0, "train/global_step": 20, "_timestamp": 1712808360.9619126, "_runtime": 65.25462675094604, "_step": 0, "_wandb": {"runtime": 106}}
wandb/run-20240411_040455-4dprer7c/logs/debug-internal.log ADDED
@@ -0,0 +1,254 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 04:04:55,708 INFO StreamThr :781 [internal.py:wandb_internal():86] W&B internal server running at pid: 781, started at: 2024-04-11 04:04:55.707378
2
+ 2024-04-11 04:04:55,709 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 04:04:56,097 INFO WriterThread:781 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_040455-4dprer7c/run-4dprer7c.wandb
4
+ 2024-04-11 04:04:56,097 DEBUG SenderThread:781 [sender.py:send():379] send: header
5
+ 2024-04-11 04:04:56,100 DEBUG SenderThread:781 [sender.py:send():379] send: run
6
+ 2024-04-11 04:04:56,242 INFO SenderThread:781 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files
7
+ 2024-04-11 04:04:56,242 INFO SenderThread:781 [sender.py:_start_run_threads():1124] run started: 4dprer7c with start time 1712808295.707286
8
+ 2024-04-11 04:04:56,250 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 04:04:56,250 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 04:04:56,338 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 04:04:56,349 DEBUG HandlerThread:781 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 04:04:56,349 DEBUG HandlerThread:781 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 04:04:56,349 INFO HandlerThread:781 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 04:04:56,349 INFO SystemMonitor:781 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 04:04:56,349 INFO HandlerThread:781 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-11 04:04:56,350 INFO SystemMonitor:781 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-11 04:04:56,351 INFO SystemMonitor:781 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-11 04:04:56,352 INFO SystemMonitor:781 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 04:04:56,352 INFO SystemMonitor:781 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 04:04:56,353 INFO SystemMonitor:781 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 04:04:56,368 DEBUG HandlerThread:781 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 04:04:56,370 DEBUG HandlerThread:781 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 04:04:56,370 DEBUG HandlerThread:781 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 04:04:56,370 DEBUG HandlerThread:781 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T04:04:56.368183', 'startedAt': '2024-04-11T04:04:55.701503', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '321ddc84c049', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.19, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}, {'current': 2000.19, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5566.962100982666}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 04:04:56,370 INFO HandlerThread:781 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 04:04:56,370 INFO HandlerThread:781 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 04:04:56,370 DEBUG HandlerThread:781 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 04:04:57,244 INFO Thread-12 :781 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/conda-environment.yaml
29
+ 2024-04-11 04:05:11,386 ERROR HandlerThread:781 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 04:05:11,386 DEBUG HandlerThread:781 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 04:05:11,387 INFO HandlerThread:781 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 04:05:11,392 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 04:05:11,393 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 04:05:11,393 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 04:05:11,393 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 04:05:11,393 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 04:05:11,393 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 04:05:11,393 DEBUG SenderThread:781 [sender.py:send():379] send: files
49
+ 2024-04-11 04:05:11,394 INFO SenderThread:781 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 04:05:11,591 INFO wandb-upload_0:781 [upload_job.py:push():131] Uploaded file /tmp/tmp48h4betowandb/odf1mr1f-wandb-metadata.json
51
+ 2024-04-11 04:05:12,247 INFO Thread-12 :781 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-metadata.json
52
+ 2024-04-11 04:05:12,360 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 04:05:12,360 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 04:05:12,363 DEBUG SenderThread:781 [sender.py:send():379] send: telemetry
55
+ 2024-04-11 04:05:12,373 DEBUG SenderThread:781 [sender.py:send():379] send: config
56
+ 2024-04-11 04:05:12,375 DEBUG SenderThread:781 [sender.py:send():379] send: metric
57
+ 2024-04-11 04:05:12,376 DEBUG SenderThread:781 [sender.py:send():379] send: telemetry
58
+ 2024-04-11 04:05:12,377 DEBUG SenderThread:781 [sender.py:send():379] send: metric
59
+ 2024-04-11 04:05:12,377 WARNING SenderThread:781 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-11 04:05:12,377 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
61
+ 2024-04-11 04:05:12,377 DEBUG SenderThread:781 [sender.py:send():379] send: telemetry
62
+ 2024-04-11 04:05:12,378 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
63
+ 2024-04-11 04:05:12,379 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-04-11 04:05:13,248 INFO Thread-12 :781 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
65
+ 2024-04-11 04:05:13,248 INFO Thread-12 :781 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/requirements.txt
66
+ 2024-04-11 04:05:15,249 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
67
+ 2024-04-11 04:05:16,873 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 04:05:17,250 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
69
+ 2024-04-11 04:05:21,874 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-11 04:05:26,880 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 04:05:27,254 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/config.yaml
72
+ 2024-04-11 04:05:28,163 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
73
+ 2024-04-11 04:05:28,165 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
74
+ 2024-04-11 04:05:28,166 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
75
+ 2024-04-11 04:05:32,273 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-11 04:05:37,274 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-04-11 04:05:42,275 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
78
+ 2024-04-11 04:05:42,872 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
79
+ 2024-04-11 04:05:42,872 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
80
+ 2024-04-11 04:05:42,873 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
81
+ 2024-04-11 04:05:47,920 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-11 04:05:52,920 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 04:05:56,353 DEBUG SystemMonitor:781 [system_monitor.py:_start():172] Starting system metrics aggregation loop
84
+ 2024-04-11 04:05:56,355 DEBUG SenderThread:781 [sender.py:send():379] send: stats
85
+ 2024-04-11 04:05:57,869 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
86
+ 2024-04-11 04:05:57,870 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
87
+ 2024-04-11 04:05:57,903 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
88
+ 2024-04-11 04:05:58,904 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-04-11 04:06:00,962 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: partial_history
90
+ 2024-04-11 04:06:00,965 DEBUG SenderThread:781 [sender.py:send():379] send: metric
91
+ 2024-04-11 04:06:00,965 DEBUG SenderThread:781 [sender.py:send():379] send: metric
92
+ 2024-04-11 04:06:00,965 DEBUG SenderThread:781 [sender.py:send():379] send: metric
93
+ 2024-04-11 04:06:00,965 DEBUG SenderThread:781 [sender.py:send():379] send: metric
94
+ 2024-04-11 04:06:00,966 DEBUG SenderThread:781 [sender.py:send():379] send: history
95
+ 2024-04-11 04:06:00,966 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: summary_record
96
+ 2024-04-11 04:06:00,966 INFO SenderThread:781 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
97
+ 2024-04-11 04:06:01,267 INFO Thread-12 :781 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json
98
+ 2024-04-11 04:06:03,268 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
99
+ 2024-04-11 04:06:04,004 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-04-11 04:06:07,270 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
101
+ 2024-04-11 04:06:09,172 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
102
+ 2024-04-11 04:06:12,869 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
103
+ 2024-04-11 04:06:12,870 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
104
+ 2024-04-11 04:06:12,873 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
105
+ 2024-04-11 04:06:14,934 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-04-11 04:06:19,934 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-04-11 04:06:24,936 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-04-11 04:06:26,356 DEBUG SenderThread:781 [sender.py:send():379] send: stats
109
+ 2024-04-11 04:06:27,870 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
110
+ 2024-04-11 04:06:27,870 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
111
+ 2024-04-11 04:06:27,910 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
112
+ 2024-04-11 04:06:29,974 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-04-11 04:06:30,281 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/config.yaml
114
+ 2024-04-11 04:06:35,067 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-04-11 04:06:40,068 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 04:06:42,978 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: pause
117
+ 2024-04-11 04:06:42,978 INFO HandlerThread:781 [handler.py:handle_request_pause():708] stopping system metrics thread
118
+ 2024-04-11 04:06:42,978 INFO HandlerThread:781 [system_monitor.py:finish():203] Stopping system monitor
119
+ 2024-04-11 04:06:42,978 INFO HandlerThread:781 [interfaces.py:finish():202] Joined cpu monitor
120
+ 2024-04-11 04:06:42,979 INFO HandlerThread:781 [interfaces.py:finish():202] Joined disk monitor
121
+ 2024-04-11 04:06:42,979 DEBUG SystemMonitor:781 [system_monitor.py:_start():179] Finished system metrics aggregation loop
122
+ 2024-04-11 04:06:42,979 DEBUG SystemMonitor:781 [system_monitor.py:_start():183] Publishing last batch of metrics
123
+ 2024-04-11 04:06:42,990 INFO HandlerThread:781 [interfaces.py:finish():202] Joined gpu monitor
124
+ 2024-04-11 04:06:42,990 INFO HandlerThread:781 [interfaces.py:finish():202] Joined memory monitor
125
+ 2024-04-11 04:06:42,990 INFO HandlerThread:781 [interfaces.py:finish():202] Joined network monitor
126
+ 2024-04-11 04:06:42,991 DEBUG SenderThread:781 [sender.py:send():379] send: stats
127
+ 2024-04-11 04:06:43,019 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: stop_status
128
+ 2024-04-11 04:06:43,019 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
129
+ 2024-04-11 04:06:43,019 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: stop_status
130
+ 2024-04-11 04:06:46,066 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
131
+ 2024-04-11 04:06:49,976 DEBUG SenderThread:781 [sender.py:send():379] send: exit
132
+ 2024-04-11 04:06:49,976 INFO SenderThread:781 [sender.py:send_exit():586] handling exit code: 0
133
+ 2024-04-11 04:06:49,976 INFO SenderThread:781 [sender.py:send_exit():588] handling runtime: 106
134
+ 2024-04-11 04:06:49,977 INFO SenderThread:781 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
135
+ 2024-04-11 04:06:49,977 INFO SenderThread:781 [sender.py:send_exit():594] send defer
136
+ 2024-04-11 04:06:49,977 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
137
+ 2024-04-11 04:06:49,977 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 0
138
+ 2024-04-11 04:06:49,977 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
139
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 0
140
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 1
141
+ 2024-04-11 04:06:49,978 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
142
+ 2024-04-11 04:06:49,978 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 1
143
+ 2024-04-11 04:06:49,978 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
144
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 1
145
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 2
146
+ 2024-04-11 04:06:49,978 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
147
+ 2024-04-11 04:06:49,978 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 2
148
+ 2024-04-11 04:06:49,978 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
149
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 2
150
+ 2024-04-11 04:06:49,978 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 3
151
+ 2024-04-11 04:06:49,979 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
152
+ 2024-04-11 04:06:49,979 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 3
153
+ 2024-04-11 04:06:49,979 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
154
+ 2024-04-11 04:06:49,979 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 3
155
+ 2024-04-11 04:06:49,979 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 4
156
+ 2024-04-11 04:06:49,979 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
157
+ 2024-04-11 04:06:49,979 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 4
158
+ 2024-04-11 04:06:49,979 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
159
+ 2024-04-11 04:06:49,979 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 4
160
+ 2024-04-11 04:06:49,979 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 5
161
+ 2024-04-11 04:06:49,979 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
162
+ 2024-04-11 04:06:49,980 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 5
163
+ 2024-04-11 04:06:49,980 DEBUG SenderThread:781 [sender.py:send():379] send: summary
164
+ 2024-04-11 04:06:49,980 INFO SenderThread:781 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
165
+ 2024-04-11 04:06:49,980 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
166
+ 2024-04-11 04:06:49,981 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 5
167
+ 2024-04-11 04:06:49,981 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 6
168
+ 2024-04-11 04:06:49,981 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
169
+ 2024-04-11 04:06:49,981 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 6
170
+ 2024-04-11 04:06:49,981 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
171
+ 2024-04-11 04:06:49,981 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 6
172
+ 2024-04-11 04:06:49,981 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 7
173
+ 2024-04-11 04:06:49,981 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: status_report
174
+ 2024-04-11 04:06:49,981 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
175
+ 2024-04-11 04:06:49,981 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 7
176
+ 2024-04-11 04:06:49,982 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
177
+ 2024-04-11 04:06:49,982 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 7
178
+ 2024-04-11 04:06:50,288 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json
179
+ 2024-04-11 04:06:50,499 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 8
180
+ 2024-04-11 04:06:50,499 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
181
+ 2024-04-11 04:06:50,499 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 8
182
+ 2024-04-11 04:06:50,499 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
183
+ 2024-04-11 04:06:50,499 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 8
184
+ 2024-04-11 04:06:50,500 INFO SenderThread:781 [job_builder.py:build():318] Attempting to build job artifact
185
+ 2024-04-11 04:06:50,501 INFO SenderThread:781 [job_builder.py:_get_source_type():466] no source found
186
+ 2024-04-11 04:06:50,502 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 9
187
+ 2024-04-11 04:06:50,502 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
188
+ 2024-04-11 04:06:50,502 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 9
189
+ 2024-04-11 04:06:50,502 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
190
+ 2024-04-11 04:06:50,502 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 9
191
+ 2024-04-11 04:06:50,502 INFO SenderThread:781 [dir_watcher.py:finish():358] shutting down directory watcher
192
+ 2024-04-11 04:06:50,976 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: poll_exit
193
+ 2024-04-11 04:06:51,289 INFO Thread-12 :781 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
194
+ 2024-04-11 04:06:51,289 INFO SenderThread:781 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files
195
+ 2024-04-11 04:06:51,290 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/config.yaml config.yaml
196
+ 2024-04-11 04:06:51,290 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/requirements.txt requirements.txt
197
+ 2024-04-11 04:06:51,294 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/conda-environment.yaml conda-environment.yaml
198
+ 2024-04-11 04:06:51,298 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json wandb-summary.json
199
+ 2024-04-11 04:06:51,299 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-metadata.json wandb-metadata.json
200
+ 2024-04-11 04:06:51,303 INFO SenderThread:781 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log output.log
201
+ 2024-04-11 04:06:51,304 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 10
202
+ 2024-04-11 04:06:51,305 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: poll_exit
203
+ 2024-04-11 04:06:51,308 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
204
+ 2024-04-11 04:06:51,313 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 10
205
+ 2024-04-11 04:06:51,313 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
206
+ 2024-04-11 04:06:51,313 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 10
207
+ 2024-04-11 04:06:51,313 INFO SenderThread:781 [file_pusher.py:finish():172] shutting down file pusher
208
+ 2024-04-11 04:06:51,460 INFO wandb-upload_0:781 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/config.yaml
209
+ 2024-04-11 04:06:51,532 INFO wandb-upload_1:781 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/requirements.txt
210
+ 2024-04-11 04:06:51,536 INFO wandb-upload_3:781 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/output.log
211
+ 2024-04-11 04:06:51,604 INFO wandb-upload_2:781 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_040455-4dprer7c/files/wandb-summary.json
212
+ 2024-04-11 04:06:51,804 INFO Thread-11 (_thread_body):781 [sender.py:transition_state():614] send defer: 11
213
+ 2024-04-11 04:06:51,805 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
214
+ 2024-04-11 04:06:51,805 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 11
215
+ 2024-04-11 04:06:51,805 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
216
+ 2024-04-11 04:06:51,805 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 11
217
+ 2024-04-11 04:06:51,805 INFO SenderThread:781 [file_pusher.py:join():178] waiting for file pusher
218
+ 2024-04-11 04:06:51,806 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 12
219
+ 2024-04-11 04:06:51,806 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
220
+ 2024-04-11 04:06:51,806 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 12
221
+ 2024-04-11 04:06:51,806 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
222
+ 2024-04-11 04:06:51,806 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 12
223
+ 2024-04-11 04:06:51,806 INFO SenderThread:781 [file_stream.py:finish():614] file stream finish called
224
+ 2024-04-11 04:06:51,868 INFO SenderThread:781 [file_stream.py:finish():618] file stream finish is done
225
+ 2024-04-11 04:06:51,869 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 13
226
+ 2024-04-11 04:06:51,869 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
227
+ 2024-04-11 04:06:51,869 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 13
228
+ 2024-04-11 04:06:51,869 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
229
+ 2024-04-11 04:06:51,869 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 13
230
+ 2024-04-11 04:06:51,869 INFO SenderThread:781 [sender.py:transition_state():614] send defer: 14
231
+ 2024-04-11 04:06:51,869 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: defer
232
+ 2024-04-11 04:06:51,869 INFO HandlerThread:781 [handler.py:handle_request_defer():172] handle defer: 14
233
+ 2024-04-11 04:06:51,870 DEBUG SenderThread:781 [sender.py:send():379] send: final
234
+ 2024-04-11 04:06:51,870 DEBUG SenderThread:781 [sender.py:send():379] send: footer
235
+ 2024-04-11 04:06:51,870 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: defer
236
+ 2024-04-11 04:06:51,870 INFO SenderThread:781 [sender.py:send_request_defer():610] handle sender defer: 14
237
+ 2024-04-11 04:06:51,871 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: poll_exit
238
+ 2024-04-11 04:06:51,872 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: poll_exit
239
+ 2024-04-11 04:06:51,872 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: poll_exit
240
+ 2024-04-11 04:06:51,873 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: poll_exit
241
+ 2024-04-11 04:06:51,873 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: server_info
242
+ 2024-04-11 04:06:51,873 DEBUG SenderThread:781 [sender.py:send_request():406] send_request: server_info
243
+ 2024-04-11 04:06:51,876 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: get_summary
244
+ 2024-04-11 04:06:51,876 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: sampled_history
245
+ 2024-04-11 04:06:51,877 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: internal_messages
246
+ 2024-04-11 04:06:51,888 INFO MainThread:781 [wandb_run.py:_footer_history_summary_info():3920] rendering history
247
+ 2024-04-11 04:06:51,888 INFO MainThread:781 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
248
+ 2024-04-11 04:06:51,889 INFO MainThread:781 [wandb_run.py:_footer_sync_info():3879] logging synced files
249
+ 2024-04-11 04:06:51,889 DEBUG HandlerThread:781 [handler.py:handle_request():146] handle_request: shutdown
250
+ 2024-04-11 04:06:51,889 INFO HandlerThread:781 [handler.py:finish():866] shutting down handler
251
+ 2024-04-11 04:06:52,873 INFO WriterThread:781 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_040455-4dprer7c/run-4dprer7c.wandb
252
+ 2024-04-11 04:06:52,888 INFO SenderThread:781 [sender.py:finish():1546] shutting down sender
253
+ 2024-04-11 04:06:52,888 INFO SenderThread:781 [file_pusher.py:finish():172] shutting down file pusher
254
+ 2024-04-11 04:06:52,888 INFO SenderThread:781 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_040455-4dprer7c/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Configure stats pid to 740
3
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_040455-4dprer7c/logs/debug.log
10
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_040455-4dprer7c/logs/debug-internal.log
11
+ 2024-04-11 04:04:55,703 INFO MainThread:740 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7805e5e25660>
12
+ 2024-04-11 04:04:55,704 INFO MainThread:740 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 04:04:55,704 INFO MainThread:740 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 04:04:55,704 INFO MainThread:740 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 04:04:55,704 INFO MainThread:740 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 04:04:55,706 INFO MainThread:740 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 04:04:55,707 INFO MainThread:740 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 04:04:55,718 INFO MainThread:740 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 04:04:56,096 INFO MainThread:740 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 04:04:56,099 INFO MainThread:740 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 04:04:56,249 INFO MainThread:740 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 04:04:56,332 INFO MainThread:740 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 04:04:56,332 INFO MainThread:740 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 04:05:12,361 INFO MainThread:740 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 04:05:12,361 INFO MainThread:740 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 04:05:12,362 INFO MainThread:740 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 04:05:12,362 INFO MainThread:740 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 04:05:12,363 INFO MainThread:740 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 04:05:12,368 INFO MainThread:740 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 18000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_04-04-49_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-11 04:06:42,977 INFO MainThread:740 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-11 04:06:42,977 INFO MainThread:740 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_040455-4dprer7c/run-4dprer7c.wandb ADDED
Binary file (13 kB). View file
 
wandb/run-20240411_040722-rimyh9a1/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_040722-rimyh9a1/files/config.yaml ADDED
@@ -0,0 +1,702 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712808442.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ - 62
48
+ 4: 3.10.13
49
+ 5: 0.16.5
50
+ 6: 4.39.3
51
+ 8:
52
+ - 1
53
+ - 2
54
+ - 5
55
+ 9:
56
+ 1: transformers_trainer
57
+ 13: linux-x86_64
58
+ m:
59
+ - 1: train/global_step
60
+ 6:
61
+ - 3
62
+ - 1: train/loss
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/grad_norm
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/learning_rate
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: train/epoch
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ vocab_size:
79
+ desc: null
80
+ value: 250880
81
+ hidden_size:
82
+ desc: null
83
+ value: 1024
84
+ n_layer:
85
+ desc: null
86
+ value: 24
87
+ n_head:
88
+ desc: null
89
+ value: 16
90
+ layer_norm_epsilon:
91
+ desc: null
92
+ value: 1.0e-05
93
+ initializer_range:
94
+ desc: null
95
+ value: 0.02
96
+ use_cache:
97
+ desc: null
98
+ value: false
99
+ pretraining_tp:
100
+ desc: null
101
+ value: 1
102
+ apply_residual_connection_post_layernorm:
103
+ desc: null
104
+ value: false
105
+ hidden_dropout:
106
+ desc: null
107
+ value: 0.0
108
+ attention_dropout:
109
+ desc: null
110
+ value: 0.0
111
+ bos_token_id:
112
+ desc: null
113
+ value: 1
114
+ eos_token_id:
115
+ desc: null
116
+ value: 2
117
+ slow_but_exact:
118
+ desc: null
119
+ value: false
120
+ return_dict:
121
+ desc: null
122
+ value: true
123
+ output_hidden_states:
124
+ desc: null
125
+ value: false
126
+ output_attentions:
127
+ desc: null
128
+ value: false
129
+ torchscript:
130
+ desc: null
131
+ value: false
132
+ torch_dtype:
133
+ desc: null
134
+ value: null
135
+ use_bfloat16:
136
+ desc: null
137
+ value: false
138
+ tf_legacy_loss:
139
+ desc: null
140
+ value: false
141
+ pruned_heads:
142
+ desc: null
143
+ value: {}
144
+ tie_word_embeddings:
145
+ desc: null
146
+ value: true
147
+ chunk_size_feed_forward:
148
+ desc: null
149
+ value: 0
150
+ is_encoder_decoder:
151
+ desc: null
152
+ value: false
153
+ is_decoder:
154
+ desc: null
155
+ value: false
156
+ cross_attention_hidden_size:
157
+ desc: null
158
+ value: null
159
+ add_cross_attention:
160
+ desc: null
161
+ value: false
162
+ tie_encoder_decoder:
163
+ desc: null
164
+ value: false
165
+ max_length:
166
+ desc: null
167
+ value: 20
168
+ min_length:
169
+ desc: null
170
+ value: 0
171
+ do_sample:
172
+ desc: null
173
+ value: false
174
+ early_stopping:
175
+ desc: null
176
+ value: false
177
+ num_beams:
178
+ desc: null
179
+ value: 1
180
+ num_beam_groups:
181
+ desc: null
182
+ value: 1
183
+ diversity_penalty:
184
+ desc: null
185
+ value: 0.0
186
+ temperature:
187
+ desc: null
188
+ value: 1.0
189
+ top_k:
190
+ desc: null
191
+ value: 50
192
+ top_p:
193
+ desc: null
194
+ value: 1.0
195
+ typical_p:
196
+ desc: null
197
+ value: 1.0
198
+ repetition_penalty:
199
+ desc: null
200
+ value: 1.0
201
+ length_penalty:
202
+ desc: null
203
+ value: 1.0
204
+ no_repeat_ngram_size:
205
+ desc: null
206
+ value: 0
207
+ encoder_no_repeat_ngram_size:
208
+ desc: null
209
+ value: 0
210
+ bad_words_ids:
211
+ desc: null
212
+ value: null
213
+ num_return_sequences:
214
+ desc: null
215
+ value: 1
216
+ output_scores:
217
+ desc: null
218
+ value: false
219
+ return_dict_in_generate:
220
+ desc: null
221
+ value: false
222
+ forced_bos_token_id:
223
+ desc: null
224
+ value: null
225
+ forced_eos_token_id:
226
+ desc: null
227
+ value: null
228
+ remove_invalid_values:
229
+ desc: null
230
+ value: false
231
+ exponential_decay_length_penalty:
232
+ desc: null
233
+ value: null
234
+ suppress_tokens:
235
+ desc: null
236
+ value: null
237
+ begin_suppress_tokens:
238
+ desc: null
239
+ value: null
240
+ architectures:
241
+ desc: null
242
+ value:
243
+ - BloomForCausalLM
244
+ finetuning_task:
245
+ desc: null
246
+ value: null
247
+ id2label:
248
+ desc: null
249
+ value:
250
+ '0': LABEL_0
251
+ '1': LABEL_1
252
+ label2id:
253
+ desc: null
254
+ value:
255
+ LABEL_0: 0
256
+ LABEL_1: 1
257
+ tokenizer_class:
258
+ desc: null
259
+ value: null
260
+ prefix:
261
+ desc: null
262
+ value: null
263
+ pad_token_id:
264
+ desc: null
265
+ value: 3
266
+ sep_token_id:
267
+ desc: null
268
+ value: null
269
+ decoder_start_token_id:
270
+ desc: null
271
+ value: null
272
+ task_specific_params:
273
+ desc: null
274
+ value: null
275
+ problem_type:
276
+ desc: null
277
+ value: null
278
+ _name_or_path:
279
+ desc: null
280
+ value: bigscience/bloomz-560m
281
+ transformers_version:
282
+ desc: null
283
+ value: 4.39.3
284
+ attention_softmax_in_fp32:
285
+ desc: null
286
+ value: true
287
+ bias_dropout_fusion:
288
+ desc: null
289
+ value: true
290
+ unk_token_id:
291
+ desc: null
292
+ value: 0
293
+ masked_softmax_fusion:
294
+ desc: null
295
+ value: true
296
+ model_type:
297
+ desc: null
298
+ value: bloom
299
+ n_inner:
300
+ desc: null
301
+ value: null
302
+ offset_alibi:
303
+ desc: null
304
+ value: 100
305
+ seq_length:
306
+ desc: null
307
+ value: 2048
308
+ skip_bias_add:
309
+ desc: null
310
+ value: true
311
+ skip_bias_add_qkv:
312
+ desc: null
313
+ value: false
314
+ quantization_config:
315
+ desc: null
316
+ value:
317
+ quant_method: QuantizationMethod.BITS_AND_BYTES
318
+ _load_in_8bit: false
319
+ _load_in_4bit: true
320
+ llm_int8_threshold: 6.0
321
+ llm_int8_skip_modules: null
322
+ llm_int8_enable_fp32_cpu_offload: false
323
+ llm_int8_has_fp16_weight: false
324
+ bnb_4bit_quant_type: nf4
325
+ bnb_4bit_use_double_quant: false
326
+ bnb_4bit_compute_dtype: float16
327
+ bnb_4bit_quant_storage: uint8
328
+ load_in_4bit: true
329
+ load_in_8bit: false
330
+ output_dir:
331
+ desc: null
332
+ value: /kaggle/working/
333
+ overwrite_output_dir:
334
+ desc: null
335
+ value: false
336
+ do_train:
337
+ desc: null
338
+ value: false
339
+ do_eval:
340
+ desc: null
341
+ value: false
342
+ do_predict:
343
+ desc: null
344
+ value: false
345
+ evaluation_strategy:
346
+ desc: null
347
+ value: 'no'
348
+ prediction_loss_only:
349
+ desc: null
350
+ value: false
351
+ per_device_train_batch_size:
352
+ desc: null
353
+ value: 8
354
+ per_device_eval_batch_size:
355
+ desc: null
356
+ value: 8
357
+ per_gpu_train_batch_size:
358
+ desc: null
359
+ value: null
360
+ per_gpu_eval_batch_size:
361
+ desc: null
362
+ value: null
363
+ gradient_accumulation_steps:
364
+ desc: null
365
+ value: 1
366
+ eval_accumulation_steps:
367
+ desc: null
368
+ value: null
369
+ eval_delay:
370
+ desc: null
371
+ value: 0
372
+ learning_rate:
373
+ desc: null
374
+ value: 3.0e-05
375
+ weight_decay:
376
+ desc: null
377
+ value: 0.001
378
+ adam_beta1:
379
+ desc: null
380
+ value: 0.9
381
+ adam_beta2:
382
+ desc: null
383
+ value: 0.999
384
+ adam_epsilon:
385
+ desc: null
386
+ value: 1.0e-08
387
+ max_grad_norm:
388
+ desc: null
389
+ value: 0.3
390
+ num_train_epochs:
391
+ desc: null
392
+ value: 5
393
+ max_steps:
394
+ desc: null
395
+ value: 20000
396
+ lr_scheduler_type:
397
+ desc: null
398
+ value: cosine
399
+ lr_scheduler_kwargs:
400
+ desc: null
401
+ value: {}
402
+ warmup_ratio:
403
+ desc: null
404
+ value: 0.03
405
+ warmup_steps:
406
+ desc: null
407
+ value: 0
408
+ log_level:
409
+ desc: null
410
+ value: passive
411
+ log_level_replica:
412
+ desc: null
413
+ value: warning
414
+ log_on_each_node:
415
+ desc: null
416
+ value: true
417
+ logging_dir:
418
+ desc: null
419
+ value: /kaggle/working/runs/Apr11_04-07-17_321ddc84c049
420
+ logging_strategy:
421
+ desc: null
422
+ value: steps
423
+ logging_first_step:
424
+ desc: null
425
+ value: false
426
+ logging_steps:
427
+ desc: null
428
+ value: 20
429
+ logging_nan_inf_filter:
430
+ desc: null
431
+ value: true
432
+ save_strategy:
433
+ desc: null
434
+ value: steps
435
+ save_steps:
436
+ desc: null
437
+ value: 20
438
+ save_total_limit:
439
+ desc: null
440
+ value: 1
441
+ save_safetensors:
442
+ desc: null
443
+ value: true
444
+ save_on_each_node:
445
+ desc: null
446
+ value: false
447
+ save_only_model:
448
+ desc: null
449
+ value: false
450
+ no_cuda:
451
+ desc: null
452
+ value: false
453
+ use_cpu:
454
+ desc: null
455
+ value: false
456
+ use_mps_device:
457
+ desc: null
458
+ value: false
459
+ seed:
460
+ desc: null
461
+ value: 42
462
+ data_seed:
463
+ desc: null
464
+ value: null
465
+ jit_mode_eval:
466
+ desc: null
467
+ value: false
468
+ use_ipex:
469
+ desc: null
470
+ value: false
471
+ bf16:
472
+ desc: null
473
+ value: false
474
+ fp16:
475
+ desc: null
476
+ value: false
477
+ fp16_opt_level:
478
+ desc: null
479
+ value: O1
480
+ half_precision_backend:
481
+ desc: null
482
+ value: auto
483
+ bf16_full_eval:
484
+ desc: null
485
+ value: false
486
+ fp16_full_eval:
487
+ desc: null
488
+ value: false
489
+ tf32:
490
+ desc: null
491
+ value: null
492
+ local_rank:
493
+ desc: null
494
+ value: 0
495
+ ddp_backend:
496
+ desc: null
497
+ value: null
498
+ tpu_num_cores:
499
+ desc: null
500
+ value: null
501
+ tpu_metrics_debug:
502
+ desc: null
503
+ value: false
504
+ debug:
505
+ desc: null
506
+ value: []
507
+ dataloader_drop_last:
508
+ desc: null
509
+ value: false
510
+ eval_steps:
511
+ desc: null
512
+ value: null
513
+ dataloader_num_workers:
514
+ desc: null
515
+ value: 8
516
+ dataloader_prefetch_factor:
517
+ desc: null
518
+ value: null
519
+ past_index:
520
+ desc: null
521
+ value: -1
522
+ run_name:
523
+ desc: null
524
+ value: /kaggle/working/
525
+ disable_tqdm:
526
+ desc: null
527
+ value: false
528
+ remove_unused_columns:
529
+ desc: null
530
+ value: true
531
+ label_names:
532
+ desc: null
533
+ value: null
534
+ load_best_model_at_end:
535
+ desc: null
536
+ value: false
537
+ metric_for_best_model:
538
+ desc: null
539
+ value: null
540
+ greater_is_better:
541
+ desc: null
542
+ value: null
543
+ ignore_data_skip:
544
+ desc: null
545
+ value: false
546
+ fsdp:
547
+ desc: null
548
+ value: []
549
+ fsdp_min_num_params:
550
+ desc: null
551
+ value: 0
552
+ fsdp_config:
553
+ desc: null
554
+ value:
555
+ min_num_params: 0
556
+ xla: false
557
+ xla_fsdp_v2: false
558
+ xla_fsdp_grad_ckpt: false
559
+ fsdp_transformer_layer_cls_to_wrap:
560
+ desc: null
561
+ value: null
562
+ accelerator_config:
563
+ desc: null
564
+ value:
565
+ split_batches: false
566
+ dispatch_batches: null
567
+ even_batches: true
568
+ use_seedable_sampler: true
569
+ deepspeed:
570
+ desc: null
571
+ value: null
572
+ label_smoothing_factor:
573
+ desc: null
574
+ value: 0.0
575
+ optim:
576
+ desc: null
577
+ value: paged_adamw_32bit
578
+ optim_args:
579
+ desc: null
580
+ value: null
581
+ adafactor:
582
+ desc: null
583
+ value: false
584
+ group_by_length:
585
+ desc: null
586
+ value: false
587
+ length_column_name:
588
+ desc: null
589
+ value: length
590
+ report_to:
591
+ desc: null
592
+ value:
593
+ - tensorboard
594
+ - wandb
595
+ ddp_find_unused_parameters:
596
+ desc: null
597
+ value: null
598
+ ddp_bucket_cap_mb:
599
+ desc: null
600
+ value: null
601
+ ddp_broadcast_buffers:
602
+ desc: null
603
+ value: null
604
+ dataloader_pin_memory:
605
+ desc: null
606
+ value: true
607
+ dataloader_persistent_workers:
608
+ desc: null
609
+ value: false
610
+ skip_memory_metrics:
611
+ desc: null
612
+ value: true
613
+ use_legacy_prediction_loop:
614
+ desc: null
615
+ value: false
616
+ push_to_hub:
617
+ desc: null
618
+ value: false
619
+ resume_from_checkpoint:
620
+ desc: null
621
+ value: null
622
+ hub_model_id:
623
+ desc: null
624
+ value: null
625
+ hub_strategy:
626
+ desc: null
627
+ value: every_save
628
+ hub_token:
629
+ desc: null
630
+ value: <HUB_TOKEN>
631
+ hub_private_repo:
632
+ desc: null
633
+ value: false
634
+ hub_always_push:
635
+ desc: null
636
+ value: false
637
+ gradient_checkpointing:
638
+ desc: null
639
+ value: true
640
+ gradient_checkpointing_kwargs:
641
+ desc: null
642
+ value: null
643
+ include_inputs_for_metrics:
644
+ desc: null
645
+ value: false
646
+ fp16_backend:
647
+ desc: null
648
+ value: auto
649
+ push_to_hub_model_id:
650
+ desc: null
651
+ value: null
652
+ push_to_hub_organization:
653
+ desc: null
654
+ value: null
655
+ push_to_hub_token:
656
+ desc: null
657
+ value: <PUSH_TO_HUB_TOKEN>
658
+ mp_parameters:
659
+ desc: null
660
+ value: ''
661
+ auto_find_batch_size:
662
+ desc: null
663
+ value: true
664
+ full_determinism:
665
+ desc: null
666
+ value: false
667
+ torchdynamo:
668
+ desc: null
669
+ value: null
670
+ ray_scope:
671
+ desc: null
672
+ value: last
673
+ ddp_timeout:
674
+ desc: null
675
+ value: 1800
676
+ torch_compile:
677
+ desc: null
678
+ value: false
679
+ torch_compile_backend:
680
+ desc: null
681
+ value: null
682
+ torch_compile_mode:
683
+ desc: null
684
+ value: null
685
+ dispatch_batches:
686
+ desc: null
687
+ value: null
688
+ split_batches:
689
+ desc: null
690
+ value: null
691
+ include_tokens_per_second:
692
+ desc: null
693
+ value: false
694
+ include_num_input_tokens_seen:
695
+ desc: null
696
+ value: false
697
+ neftune_noise_alpha:
698
+ desc: null
699
+ value: null
700
+ optim_target_modules:
701
+ desc: null
702
+ value: null
wandb/run-20240411_040722-rimyh9a1/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240411_040722-rimyh9a1/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_040722-rimyh9a1/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T04:07:23.361534",
5
+ "startedAt": "2024-04-11T04:07:22.840819",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "321ddc84c049",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.19,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.19,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.19,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.19,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.19,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5566.962131500244
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_040722-rimyh9a1/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 3.4153, "train/grad_norm": 0.0, "train/learning_rate": 0.0, "train/epoch": 0.09, "train/global_step": 20000, "_timestamp": 1712827654.7326186, "_runtime": 19211.883972644806, "_step": 1006, "train_runtime": 18957.1753, "train_samples_per_second": 2.11, "train_steps_per_second": 1.055, "total_flos": 1.0807103426248704e+16, "train_loss": 3.9423419631004335}
wandb/run-20240411_040722-rimyh9a1/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240411_040722-rimyh9a1/logs/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Configure stats pid to 1050
3
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug.log
10
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug-internal.log
11
+ 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7d563f5f5120>
12
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 04:07:22,847 INFO MainThread:1050 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 04:07:22,848 INFO MainThread:1050 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 04:07:22,865 INFO MainThread:1050 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 04:07:23,106 INFO MainThread:1050 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 04:07:23,108 INFO MainThread:1050 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 04:07:23,235 INFO MainThread:1050 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 04:07:23,323 INFO MainThread:1050 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 04:07:23,324 INFO MainThread:1050 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 04:07:39,319 INFO MainThread:1050 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 04:07:39,324 INFO MainThread:1050 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_04-07-17_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}