Narkantak commited on
Commit
cef7e76
1 Parent(s): 05259f8

Narkantak/mistral-7b-Labelled-SLM

Browse files
README.md CHANGED
@@ -16,7 +16,12 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.4531
 
 
 
 
 
20
 
21
  ## Model description
22
 
@@ -36,68 +41,17 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
- - train_batch_size: 6
40
- - eval_batch_size: 6
41
  - seed: 42
42
- - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 24
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - lr_scheduler_warmup_steps: 2
47
  - num_epochs: 50
48
  - mixed_precision_training: Native AMP
49
 
50
- ### Training results
51
-
52
- | Training Loss | Epoch | Step | Validation Loss |
53
- |:-------------:|:-----:|:----:|:---------------:|
54
- | 5.1986 | 0.89 | 6 | 3.6826 |
55
- | 2.8565 | 1.93 | 13 | 2.2686 |
56
- | 1.6562 | 2.96 | 20 | 1.1494 |
57
- | 0.919 | 4.0 | 27 | 0.8274 |
58
- | 0.8226 | 4.89 | 33 | 0.7219 |
59
- | 0.6004 | 5.93 | 40 | 0.6329 |
60
- | 0.5166 | 6.96 | 47 | 0.5777 |
61
- | 0.4662 | 8.0 | 54 | 0.5476 |
62
- | 0.497 | 8.89 | 60 | 0.5241 |
63
- | 0.3909 | 9.93 | 67 | 0.5034 |
64
- | 0.3628 | 10.96 | 74 | 0.4779 |
65
- | 0.3381 | 12.0 | 81 | 0.4718 |
66
- | 0.3768 | 12.89 | 87 | 0.4562 |
67
- | 0.3121 | 13.93 | 94 | 0.4557 |
68
- | 0.3022 | 14.96 | 101 | 0.4528 |
69
- | 0.2995 | 16.0 | 108 | 0.4449 |
70
- | 0.3423 | 16.89 | 114 | 0.4521 |
71
- | 0.2929 | 17.93 | 121 | 0.4319 |
72
- | 0.2899 | 18.96 | 128 | 0.4354 |
73
- | 0.2873 | 20.0 | 135 | 0.4402 |
74
- | 0.3325 | 20.89 | 141 | 0.4465 |
75
- | 0.2828 | 21.93 | 148 | 0.4420 |
76
- | 0.2813 | 22.96 | 155 | 0.4426 |
77
- | 0.2817 | 24.0 | 162 | 0.4408 |
78
- | 0.3255 | 24.89 | 168 | 0.4417 |
79
- | 0.2793 | 25.93 | 175 | 0.4404 |
80
- | 0.2772 | 26.96 | 182 | 0.4385 |
81
- | 0.279 | 28.0 | 189 | 0.4409 |
82
- | 0.3247 | 28.89 | 195 | 0.4429 |
83
- | 0.2775 | 29.93 | 202 | 0.4421 |
84
- | 0.2769 | 30.96 | 209 | 0.4409 |
85
- | 0.2779 | 32.0 | 216 | 0.4423 |
86
- | 0.322 | 32.89 | 222 | 0.4450 |
87
- | 0.2767 | 33.93 | 229 | 0.4484 |
88
- | 0.2771 | 34.96 | 236 | 0.4455 |
89
- | 0.2761 | 36.0 | 243 | 0.4476 |
90
- | 0.3212 | 36.89 | 249 | 0.4485 |
91
- | 0.2751 | 37.93 | 256 | 0.4503 |
92
- | 0.2766 | 38.96 | 263 | 0.4522 |
93
- | 0.2768 | 40.0 | 270 | 0.4488 |
94
- | 0.321 | 40.89 | 276 | 0.4498 |
95
- | 0.2743 | 41.93 | 283 | 0.4518 |
96
- | 0.2746 | 42.96 | 290 | 0.4527 |
97
- | 0.2751 | 44.0 | 297 | 0.4532 |
98
- | 0.2867 | 44.44 | 300 | 0.4531 |
99
-
100
-
101
  ### Framework versions
102
 
103
  - PEFT 0.10.0
 
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: 1.0970
20
+ - eval_runtime: 1.8296
21
+ - eval_samples_per_second: 1.093
22
+ - eval_steps_per_second: 0.547
23
+ - epoch: 12.0
24
+ - step: 16
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 0.0002
44
+ - train_batch_size: 5
45
+ - eval_batch_size: 5
46
  - seed: 42
47
+ - gradient_accumulation_steps: 3
48
+ - total_train_batch_size: 15
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_steps: 1
52
  - num_epochs: 50
53
  - mixed_precision_training: Native AMP
54
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
  ### Framework versions
56
 
57
  - PEFT 0.10.0
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "TheBloke/Mistral-7B-Instruct-v0.2-GPTQ",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5cf3d53248b547264b821c9dd7dfd8c91324a24627c604948ecd8d54e85171a
3
- size 8398144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bb4ad88bb45adc123b696e91c70736275c8d379919cb30b6f99cb1320a0d768
3
+ size 8397056
runs/Apr01_09-37-28_148e249fd8ed/events.out.tfevents.1711964249.148e249fd8ed.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ab97f8aa92b5252740098c69e920c3c7b1c7bd3c8daa021f0be3274aab8dd61
3
+ size 6184
runs/Apr01_09-39-21_148e249fd8ed/events.out.tfevents.1711964361.148e249fd8ed.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8856799e1c3e22262e6b31a8e4c90d8a3394e20f85b9902f8da1b55debd8d199
3
+ size 6184
runs/Apr01_09-41-00_148e249fd8ed/events.out.tfevents.1711964468.148e249fd8ed.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:210747c3e10d00768f172398957c3a46490b1caba1d98135e51bdbadffb98d8e
3
+ size 5238
runs/Apr01_09-41-30_148e249fd8ed/events.out.tfevents.1711964491.148e249fd8ed.34.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e59c04580b1e9d1dd0c9e137685be59deaeb5a46531f310163ab292a304180c
3
+ size 5238
runs/Apr01_09-41-45_148e249fd8ed/events.out.tfevents.1711964509.148e249fd8ed.34.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f7703b12c03d82d72356c7ae44ffa5e053a67fd8d5c8411572ad63c44213019
3
+ size 10914
test.csv ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ example
2
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
3
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
4
+ Document Type: Document Type 7
5
+
6
+ , context associated to the document type is
7
+
8
+ No Information Available [/INST] "
9
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
10
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
11
+ Document Type: Document Type 1
12
+
13
+ , context associated to the document type is
14
+
15
+ (DT - OMB control number: 1545-0047/Expiration Date: 12/31/2019)(DOL - OMB control number: 1210-0147/Expiration date: 5/31/2022) Page 1 of 5
16
+ (HHS - OMB control number: 0938-1146/Expiration date: 10/31/2022)
17
+ Summary of Benefits and Coverage: What this Plan Covers & What You Pay for Covered Services Coverage Period: Beginning on or after 01/01/2021
18
+ Community Blue – Faculty/Acad Active: Michigan State University Coverage for: Individual/Family | PlanType:PPO
19
+ The Summary of Benefits and Coverage (SBC) document will help you choose a health plan. The SBC shows you how you and the plan would share the cost for covered health care services. NOTE: Information about the cost of this plan (called the premium) will be provided separately. This is only a summary. For more information about your coverage, or to get a copy of the complete terms of coverage, visit www.bcbsm.com or call 888-288-1726. For general definitions of common terms, such as allowed amount, balance billing, coinsurance, copayment, deductible, provider, or other underlined terms, see the Glossary. You can view the Glossary at https://www.healthcare.gov/sbc-glossary or call 888-288-1726 to request a copy.
20
+ Important Questions Answers Why This Matters: In-Network Out-of-Network
21
+ What is the overall deductible?
22
+ $100 Individual/ $200 Family
23
+ $250 Individual/
24
+ $500 Family
25
+ Generally, you must pay all of the costs from provider's up to the deductible amount before this plan begins to pay. If you have other family members on the plan, each family member must meet their own individual deductible until the total amount of deductible expenses paid by all family members meets the overall family deductible. Are there services covered before you meet your deductible? Yes This plan covers some items and services even if you haven’t yet met the deductible amount. But a copayment or coinsurance may apply. For example, this plan covers certain preventive services without cost-sharing and before you meet your deductible. See a list of covered preventive services at https://www.healthcare.gov/coverage/preventive-care-benefits/
26
+ Are there other deductibles for specific services?
27
+ No
28
+ You don’t have to meet deductibles for specific services. What is the out-of-pocket limit for this plan? $2,000 Individual/ $4,000 Family $2,000 Individual/ $4,000 Family Plus deductible The out-of-pocket limit is the most you could pay in a year for covered services. If you have other family members in this plan, they have to meet their own out-of-pocket limits until the overall family out-of-pocket limit has been met.
29
+ What is not included in the out-of-pocket limit?
30
+ Premiums, balance-billed charges, any pharmacy penalty and health care this plan doesn’t cover.
31
+ Even though you pay these expenses, they don’t count toward the out-of-pocket limit. Will you pay less if you use a network provider? Yes. For a list of in-network providers, see www.bcbsm.com or call the number on the back of your BCBSM ID card. This plan uses a provider network. You will pay less if you use a provider in the plan’s network. You will pay the most if you use an out-of-network provider, and you might receive a bill from a provider for the difference between the provider’s charge and what your plan pays (balance billing).
32
+ Do you need a referral to see a specialist?
33
+ No
34
+ You can see the specialist you choose without a referral. [/INST] "
test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b71e3328ddf3c264c43471b314f5f9dd852688fedfae9484b9ee9f836b062f28
3
+ size 16156
train.csv ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ example
2
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
3
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
4
+ Document Type: Document Type 8
5
+
6
+ , context associated to the document type is
7
+
8
+ EcoGuardians is a visionary environmental organization dedicated to safeguarding our planet for future generations. Committed to promoting sustainability, conservation, and environmental awareness, EcoGuardians strives to protect biodiversity, combat climate change, and advocate for eco-friendly practices.
9
+ Through innovative projects, educational initiatives, and community engagement, EcoGuardians empowers individuals to make a positive impact on the environment. By fostering a culture of environmental stewardship, EcoGuardians aims to inspire collective action towards a greener, more sustainable world. [/INST] "
10
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
11
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
12
+ Document Type: Document Type 9
13
+
14
+ , context associated to the document type is
15
+
16
+ No Information Available [/INST] "
17
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
18
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
19
+ Document Type: Document Type 6
20
+
21
+ , context associated to the document type is
22
+
23
+ No Information Available [/INST] "
24
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
25
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
26
+ Document Type: Document Type 5
27
+
28
+ , context associated to the document type is
29
+
30
+ 1 of 5
31
+ Summary of Benefits and Coverage: What this Plan Covers & What You Pay For Covered Services Coverage Period: 01/01/2018 – 12/31/2018
32
+ Insurance Company 1: Plan Option 1 Coverage for: Family | Plan Type: PPO
33
+ The Summary of Benefits and Coverage (SBC) document will help you choose a health plan. The SBC shows you how you and the plan would share the cost for covered health care services. NOTE: Information about the cost of this plan (called the premium) will be provided separately. This is only a summary. For more information about your coverage, or to get a copy of the complete terms of coverage, [insert contact information]. For general definitions of common terms, such as allowed amount, balance billing, coinsurance, copayment, deductible, provider, or other underlined terms see the Glossary. You can view the Glossary at www.[insert].com or call 1-800-[insert] to request a copy.
34
+ Important Questions Answers Why This Matters:
35
+ What is the overall deductible?
36
+ $500/Individual or $1,000/family
37
+ Generally, you must pay all of the costs from providers up to the deductible amount before this plan begins to pay. If you have other family members on the plan, each family member must meet their own individual deductible until the total amount of deductible expenses paid by all family members meets the overall family deductible. Are there services covered before you meet your deductible? Yes. Preventive care and primary care services are covered before you meet your deductible. This plan covers some items and services even if you haven’t yet met the deductible amount. But a copayment or coinsurance may apply. For example, this plan covers certain preventive services without cost-sharing and before you meet your deductible. See a list of covered preventive services at https://www.healthcare.gov/coverage/preventive-care-benefits/. Are there other deductibles for specific services? Yes. $300 for prescription drug coverage and $300 for occupational therapy services. You must pay all of the costs for these services up to the specific deductible amount before this plan begins to pay for these services. What is the out-of-pocket limit for this plan? For network providers $2,500 individual / $5,000 family; for out-of-network providers $4,000 individual / $8,000 family The out-of-pocket limit is the most you could pay in a year for covered services. If you have other family members in this plan, they have to meet their own out-of-pocket limits until the overall family out-of-pocket limit has been met.
38
+ What is not included in
39
+ the out-of-pocket limit?
40
+ Copayments for certain services, premiums, balance-billing charges, and health care this plan doesn’t cover.
41
+ Even though you pay these expenses, they don’t count toward the out–of–pocket limit. Will you pay less if you use a network provider? Yes. See www.[insert].com or call 1-800-[insert] for a list of network providers. This plan uses a provider network. You will pay less if you use a provider in the plan’s network. You will pay the most if you use an out-of-network provider, and you might receive a bill from a provider for the difference between the provider’s charge and what your plan pays (balance billing). Be aware, your network provider might use an out-of-network provider for some services (such as lab work). Check with your provider before you get services. Do you need a referral to see a specialist? Yes. This plan will pay some or all of the costs to see a specialist for covered services but only if you have a referral before you see the specialist.
42
+ OMB Control Numbers 1545-2229, 1210-0147, and 0938-1146
43
+ Released on April 6, 2016 [/INST] "
44
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
45
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
46
+ Document Type: Document Type 3
47
+
48
+ , context associated to the document type is
49
+
50
+ 1 of 5
51
+ Summary of Benefits and Coverage: What this Plan Covers & What You Pay For Covered Services Coverage Period: 01/01/2018 – 12/31/2018
52
+ Insurance Company 1: Plan Option 1 Coverage for: Family | Plan Type: PPO
53
+ The Summary of Benefits and Coverage (SBC) document will help you choose a health plan. The SBC shows you how you and the plan would share the cost for covered health care services. NOTE: Information about the cost of this plan (called the premium) will be provided separately. This is only a summary. For more information about your coverage, or to get a copy of the complete terms of coverage, [insert contact information]. For general definitions of common terms, such as allowed amount, balance billing, coinsurance, copayment, deductible, provider, or other underlined terms see the Glossary. You can view the Glossary at www.[insert].com or call 1-800-[insert] to request a copy.
54
+ Important Questions Answers Why This Matters:
55
+ What is the overall deductible?
56
+ $500/Individual or $1,000/family
57
+ Generally, you must pay all of the costs from providers up to the deductible amount before this plan begins to pay. If you have other family members on the plan, each family member must meet their own individual deductible until the total amount of deductible expenses paid by all family members meets the overall family deductible. Are there services covered before you meet your deductible? Yes. Preventive care and primary care services are covered before you meet your deductible. This plan covers some items and services even if you haven’t yet met the deductible amount. But a copayment or coinsurance may apply. For example, this plan covers certain preventive services without cost-sharing and before you meet your deductible. See a list of covered preventive services at https://www.healthcare.gov/coverage/preventive-care-benefits/. Are there other deductibles for specific services? Yes. $300 for prescription drug coverage and $300 for occupational therapy services. You must pay all of the costs for these services up to the specific deductible amount before this plan begins to pay for these services. What is the out-of-pocket limit for this plan? For network providers $2,500 individual / $5,000 family; for out-of-network providers $4,000 individual / $8,000 family The out-of-pocket limit is the most you could pay in a year for covered services. If you have other family members in this plan, they have to meet their own out-of-pocket limits until the overall family out-of-pocket limit has been met.
58
+ What is not included in
59
+ the out-of-pocket limit?
60
+ Copayments for certain services, premiums, balance-billing charges, and health care this plan doesn’t cover.
61
+ Even though you pay these expenses, they don’t count toward the out–of–pocket limit. Will you pay less if you use a network provider? Yes. See www.[insert].com or call 1-800-[insert] for a list of network providers. This plan uses a provider network. You will pay less if you use a provider in the plan’s network. You will pay the most if you use an out-of-network provider, and you might receive a bill from a provider for the difference between the provider’s charge and what your plan pays (balance billing). Be aware, your network provider might use an out-of-network provider for some services (such as lab work). Check with your provider before you get services. Do you need a referral to see a specialist? Yes. This plan will pay some or all of the costs to see a specialist for covered services but only if you have a referral before you see the specialist.
62
+ OMB Control Numbers 1545-2229, 1210-0147, and 0938-1146
63
+ Released on April 6, 2016 [/INST] "
64
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
65
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
66
+ Document Type: Document Type 4
67
+
68
+ , context associated to the document type is
69
+
70
+ (DT - OMB control number: 1545-0047/Expiration Date: 12/31/2019)(DOL - OMB control number: 1210-0147/Expiration date: 5/31/2022) Page 1 of 5
71
+ (HHS - OMB control number: 0938-1146/Expiration date: 10/31/2022)
72
+ Summary of Benefits and Coverage: What this Plan Covers & What You Pay for Covered Services Coverage Period: Beginning on or after 01/01/2021
73
+ Community Blue – Faculty/Acad Active: Michigan State University Coverage for: Individual/Family | PlanType:PPO
74
+ The Summary of Benefits and Coverage (SBC) document will help you choose a health plan. The SBC shows you how you and the plan would share the cost for covered health care services. NOTE: Information about the cost of this plan (called the premium) will be provided separately. This is only a summary. For more information about your coverage, or to get a copy of the complete terms of coverage, visit www.bcbsm.com or call 888-288-1726. For general definitions of common terms, such as allowed amount, balance billing, coinsurance, copayment, deductible, provider, or other underlined terms, see the Glossary. You can view the Glossary at https://www.healthcare.gov/sbc-glossary or call 888-288-1726 to request a copy.
75
+ Important Questions Answers Why This Matters: In-Network Out-of-Network
76
+ What is the overall deductible?
77
+ $100 Individual/ $200 Family
78
+ $250 Individual/
79
+ $500 Family
80
+ Generally, you must pay all of the costs from provider's up to the deductible amount before this plan begins to pay. If you have other family members on the plan, each family member must meet their own individual deductible until the total amount of deductible expenses paid by all family members meets the overall family deductible. Are there services covered before you meet your deductible? Yes This plan covers some items and services even if you haven’t yet met the deductible amount. But a copayment or coinsurance may apply. For example, this plan covers certain preventive services without cost-sharing and before you meet your deductible. See a list of covered preventive services at https://www.healthcare.gov/coverage/preventive-care-benefits/
81
+ Are there other deductibles for specific services?
82
+ No
83
+ You don’t have to meet deductibles for specific services. What is the out-of-pocket limit for this plan? $2,000 Individual/ $4,000 Family $2,000 Individual/ $4,000 Family Plus deductible The out-of-pocket limit is the most you could pay in a year for covered services. If you have other family members in this plan, they have to meet their own out-of-pocket limits until the overall family out-of-pocket limit has been met.
84
+ What is not included in the out-of-pocket limit?
85
+ Premiums, balance-billed charges, any pharmacy penalty and health care this plan doesn’t cover.
86
+ Even though you pay these expenses, they don’t count toward the out-of-pocket limit. Will you pay less if you use a network provider? Yes. For a list of in-network providers, see www.bcbsm.com or call the number on the back of your BCBSM ID card. This plan uses a provider network. You will pay less if you use a provider in the plan’s network. You will pay the most if you use an out-of-network provider, and you might receive a bill from a provider for the difference between the provider’s charge and what your plan pays (balance billing).
87
+ Do you need a referral to see a specialist?
88
+ No
89
+ You can see the specialist you choose without a referral. [/INST] "
90
+ "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
91
+ Orion is known for its best precision and world class accuracy with the financial knowledge, \
92
+ Document Type: Document Type 2
93
+
94
+ , context associated to the document type is
95
+
96
+ No Information Available [/INST] "
train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3787510ef2fd47560b05aa86e99d6eb5c174a1b8d7ac31670e8fc45bf8fb5c39
3
+ size 7445
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1275110a846ce806cc22d527c35ea9e1824ff82afe6ebdb4216b869931436b5
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10ec4641923a16c59aa087019cc91de846e9c70e43dec12655934e34195ad3fe
3
  size 4920
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,40 +1,68 @@
1
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240401_065413-yzd7dxs3/logs/debug.log
10
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240401_065413-yzd7dxs3/logs/debug-internal.log
11
- 2024-04-01 06:54:13,060 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78c2e1a431c0>
12
- 2024-04-01 06:54:13,061 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
- 2024-04-01 06:54:13,061 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-01 06:54:13,061 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
- 2024-04-01 06:54:13,061 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
- 2024-04-01 06:54:13,063 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-01 06:54:13,066 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
- 2024-04-01 06:54:13,078 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
- 2024-04-01 06:54:13,454 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
- 2024-04-01 06:54:13,458 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
- 2024-04-01 06:54:13,586 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
- 2024-04-01 06:54:13,675 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-01 06:54:13,675 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
- 2024-04-01 06:54:44,704 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
- 2024-04-01 06:54:44,705 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
- 2024-04-01 06:54:44,705 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
- 2024-04-01 06:54:44,705 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
- 2024-04-01 06:54:44,707 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
- 2024-04-01 06:54:44,713 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_06-53-57_e5cddbcf0d54', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
- 2024-04-01 08:30:55,358 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
- 2024-04-01 08:30:55,358 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
- 2024-04-01 08:52:13,855 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
- 2024-04-01 08:52:13,887 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
- 2024-04-01 08:52:13,888 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
- 2024-04-01 08:52:43,723 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
- 2024-04-01 08:52:43,725 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
- 2024-04-01 08:52:43,726 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
- 2024-04-01 08:52:46,758 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug.log
10
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug-internal.log
11
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e30ad2cb760>
12
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-01 09:37:43,081 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-01 09:37:43,085 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-01 09:37:43,097 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-01 09:37:43,596 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-01 09:37:43,599 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-01 09:37:43,765 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-01 09:38:14,888 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-01 09:38:14,891 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-01 09:38:14,896 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-37-28_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-01 09:38:55,054 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-01 09:38:55,054 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-01 09:39:21,055 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-01 09:39:21,061 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-01 09:39:21,061 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-01 09:39:21,457 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-01 09:39:22,007 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-39-21_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
39
+ 2024-04-01 09:39:59,405 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
40
+ 2024-04-01 09:39:59,405 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
41
+ 2024-04-01 09:41:00,163 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
42
+ 2024-04-01 09:41:00,169 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
43
+ 2024-04-01 09:41:00,169 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
44
+ 2024-04-01 09:41:07,654 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
45
+ 2024-04-01 09:41:08,198 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-00_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
46
+ 2024-04-01 09:41:12,441 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
+ 2024-04-01 09:41:12,441 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
48
+ 2024-04-01 09:41:30,402 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
49
+ 2024-04-01 09:41:30,410 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
+ 2024-04-01 09:41:30,410 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
51
+ 2024-04-01 09:41:30,806 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
52
+ 2024-04-01 09:41:31,300 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-30_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
53
+ 2024-04-01 09:41:35,576 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-01 09:41:35,576 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
55
+ 2024-04-01 09:41:45,080 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
56
+ 2024-04-01 09:41:45,084 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-01 09:41:45,085 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
58
+ 2024-04-01 09:41:48,751 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
59
+ 2024-04-01 09:41:49,276 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 5, 'per_device_eval_batch_size': 5, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-45_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
60
+ 2024-04-01 09:44:51,613 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
61
+ 2024-04-01 09:44:51,613 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
62
+ 2024-04-01 09:45:32,015 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
63
+ 2024-04-01 09:45:32,039 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
64
+ 2024-04-01 09:45:32,039 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
65
+ 2024-04-01 09:45:54,716 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
66
+ 2024-04-01 09:45:54,719 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
67
+ 2024-04-01 09:45:54,719 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
68
+ 2024-04-01 09:45:57,946 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
wandb/run-20240401_093743-ie4jv7u5/files/conda-environment.yaml ADDED
@@ -0,0 +1,1072 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: base
2
+ channels:
3
+ - pytorch
4
+ - file:///tmp/conda
5
+ - rapidsai
6
+ - nvidia
7
+ - conda-forge
8
+ - defaults
9
+ dependencies:
10
+ - _libgcc_mutex=0.1=conda_forge
11
+ - _openmp_mutex=4.5=2_gnu
12
+ - aiohttp=3.9.1=py310h2372a71_0
13
+ - aiosignal=1.3.1=pyhd8ed1ab_0
14
+ - annotated-types=0.6.0=pyhd8ed1ab_0
15
+ - anyio=4.2.0=pyhd8ed1ab_0
16
+ - archspec=0.2.2=pyhd8ed1ab_0
17
+ - argon2-cffi=23.1.0=pyhd8ed1ab_0
18
+ - argon2-cffi-bindings=21.2.0=py310h2372a71_4
19
+ - arrow=1.3.0=pyhd8ed1ab_0
20
+ - arrow-cpp=11.0.0=ha770c72_9_cpu
21
+ - asttokens=2.4.1=pyhd8ed1ab_0
22
+ - async-timeout=4.0.3=pyhd8ed1ab_0
23
+ - atk-1.0=2.38.0=hd4edc92_1
24
+ - attrs=23.2.0=pyh71513ae_0
25
+ - aws-c-auth=0.6.26=h987a71b_2
26
+ - aws-c-cal=0.5.21=h48707d8_2
27
+ - aws-c-common=0.8.14=h0b41bf4_0
28
+ - aws-c-compression=0.2.16=h03acc5a_5
29
+ - aws-c-event-stream=0.2.20=h00877a2_4
30
+ - aws-c-http=0.7.6=hf342b9f_0
31
+ - aws-c-io=0.13.19=h5b20300_3
32
+ - aws-c-mqtt=0.8.6=hc4349f7_12
33
+ - aws-c-s3=0.2.7=h909e904_1
34
+ - aws-c-sdkutils=0.1.9=h03acc5a_0
35
+ - aws-checksums=0.1.14=h03acc5a_5
36
+ - aws-crt-cpp=0.19.8=hf7fbfca_12
37
+ - aws-sdk-cpp=1.10.57=h17c43bd_8
38
+ - beautifulsoup4=4.12.2=pyha770c72_0
39
+ - bleach=6.1.0=pyhd8ed1ab_0
40
+ - boltons=23.1.1=pyhd8ed1ab_0
41
+ - brotli=1.0.9=h166bdaf_9
42
+ - brotli-bin=1.0.9=h166bdaf_9
43
+ - brotli-python=1.0.9=py310hd8f1fbe_9
44
+ - brotlipy=0.7.0=py310h7f8727e_1002
45
+ - bzip2=1.0.8=h7b6447c_0
46
+ - c-ares=1.25.0=hd590300_0
47
+ - ca-certificates=2024.2.2=hbcca054_0
48
+ - cached-property=1.5.2=hd8ed1ab_1
49
+ - cached_property=1.5.2=pyha770c72_1
50
+ - cairo=1.18.0=h3faef2a_0
51
+ - cartopy=0.22.0=py310hcc13569_1
52
+ - catalogue=2.0.10=py310hff52083_0
53
+ - certifi=2024.2.2=pyhd8ed1ab_0
54
+ - cffi=1.16.0=py310h2fee648_0
55
+ - charset-normalizer=3.3.2=pyhd8ed1ab_0
56
+ - click=8.1.7=unix_pyh707e725_0
57
+ - cloudpathlib=0.16.0=pyhd8ed1ab_0
58
+ - colorama=0.4.6=pyhd8ed1ab_0
59
+ - comm=0.2.1=pyhd8ed1ab_0
60
+ - conda=23.7.4=py310hff52083_0
61
+ - conda-libmamba-solver=23.7.0=pyhd8ed1ab_0
62
+ - conda-package-handling=2.2.0=pyh38be061_0
63
+ - conda-package-streaming=0.9.0=pyhd8ed1ab_0
64
+ - confection=0.1.4=py310h17c5347_0
65
+ - contourpy=1.2.0=py310hd41b1e2_0
66
+ - cryptography=41.0.7=py310hb8475ec_1
67
+ - cuda-cccl=12.4.99=0
68
+ - cuda-cudart=12.4.99=0
69
+ - cuda-cudart-dev=12.4.99=0
70
+ - cuda-nvcc-dev_linux-64=12.1.105=ha770c72_0
71
+ - cuda-nvcc-impl=12.1.105=hd3aeb46_0
72
+ - cuda-nvcc-tools=12.1.105=hd3aeb46_0
73
+ - cuda-nvrtc=12.1.105=hd3aeb46_0
74
+ - cuda-profiler-api=12.4.99=0
75
+ - cuda-python=12.4.0=py310h52dc4f0_0
76
+ - cuda-version=12.1=h1d6eff3_3
77
+ - cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
78
+ - cuml=23.08.00=cuda12_py310_230809_gd7162cdea_0
79
+ - cupy=13.0.0=py310h7aad9d2_3
80
+ - cupy-core=13.0.0=py310had4011e_3
81
+ - curl=8.6.0=hca28451_0
82
+ - cycler=0.12.1=pyhd8ed1ab_0
83
+ - cymem=2.0.8=py310hc6cd4ac_1
84
+ - cython-blis=0.7.10=py310h1f7b6fc_2
85
+ - cytoolz=0.12.3=py310h2372a71_0
86
+ - dask-cuda=23.08.00=py310_230809_gefbd6ca_0
87
+ - dask-cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
88
+ - debugpy=1.8.0=py310hc6cd4ac_1
89
+ - decorator=5.1.1=pyhd8ed1ab_0
90
+ - defusedxml=0.7.1=pyhd8ed1ab_0
91
+ - distributed=2023.7.1=pyhd8ed1ab_0
92
+ - distro=1.9.0=pyhd8ed1ab_0
93
+ - dlenv-tf-2-15-gpu=1.0.20240111=py310ha20f8e0_0
94
+ - dlpack=0.5=h9c3ff4c_0
95
+ - entrypoints=0.4=pyhd8ed1ab_0
96
+ - exceptiongroup=1.2.0=pyhd8ed1ab_2
97
+ - executing=2.0.1=pyhd8ed1ab_0
98
+ - expat=2.6.2=h59595ed_0
99
+ - fastrlock=0.8.2=py310hc6cd4ac_2
100
+ - fftw=3.3.10=nompi_hc118613_108
101
+ - fmt=9.1.0=h924138e_0
102
+ - font-ttf-dejavu-sans-mono=2.37=hab24e00_0
103
+ - font-ttf-inconsolata=3.000=h77eed37_0
104
+ - font-ttf-source-code-pro=2.038=h77eed37_0
105
+ - font-ttf-ubuntu=0.83=h77eed37_1
106
+ - fontconfig=2.14.2=h14ed4e7_0
107
+ - fonts-conda-ecosystem=1=0
108
+ - fonts-conda-forge=1=0
109
+ - fqdn=1.5.1=pyhd8ed1ab_0
110
+ - freetype=2.12.1=h267a509_2
111
+ - fribidi=1.0.10=h36c2ea0_0
112
+ - frozenlist=1.4.1=py310h2372a71_0
113
+ - fsspec=2024.3.0=pyhca7485f_0
114
+ - gdk-pixbuf=2.42.10=h829c605_5
115
+ - geos=3.11.1=h27087fc_0
116
+ - gettext=0.21.1=h27087fc_0
117
+ - gflags=2.2.2=he1b5a44_1004
118
+ - ghostscript=10.03.0=h59595ed_0
119
+ - giflib=5.2.1=h0b41bf4_3
120
+ - glog=0.6.0=h6f12383_0
121
+ - gmock=1.14.0=ha770c72_1
122
+ - gmp=6.3.0=h59595ed_0
123
+ - google-api-core-grpc=2.11.1=hd8ed1ab_0
124
+ - google-auth=2.26.1=pyhca7485f_0
125
+ - google-cloud-core=2.4.1=pyhd8ed1ab_0
126
+ - google-cloud-datastore=2.19.0=pyhd8ed1ab_0
127
+ - googleapis-common-protos=1.62.0=pyhd8ed1ab_0
128
+ - graphite2=1.3.13=h58526e2_1001
129
+ - graphviz=9.0.0=h78e8752_1
130
+ - grpc-cpp=1.51.1=h27aab58_3
131
+ - gtest=1.14.0=h00ab1b0_1
132
+ - gtk2=2.24.33=h280cfa0_4
133
+ - gts=0.7.6=h977cf35_4
134
+ - harfbuzz=8.3.0=h3d44ed6_0
135
+ - icu=73.2=h59595ed_0
136
+ - idna=3.6=pyhd8ed1ab_0
137
+ - imagemagick=7.1.1_29=pl5321hb90aeea_0
138
+ - importlib_metadata=7.0.1=hd8ed1ab_0
139
+ - importlib_resources=6.1.1=pyhd8ed1ab_0
140
+ - intel-openmp=2023.1.0=hdb19cb5_46306
141
+ - ipykernel=6.28.0=pyhd33586a_0
142
+ - ipython=8.20.0=pyh707e725_0
143
+ - ipython_genutils=0.2.0=py_1
144
+ - isoduration=20.11.0=pyhd8ed1ab_0
145
+ - jbig=2.1=h7f98852_2003
146
+ - jedi=0.19.1=pyhd8ed1ab_0
147
+ - jinja2=3.1.2=pyhd8ed1ab_1
148
+ - joblib=1.3.2=pyhd8ed1ab_0
149
+ - jsonpatch=1.33=pyhd8ed1ab_0
150
+ - jsonpointer=2.4=py310hff52083_3
151
+ - jsonschema=4.20.0=pyhd8ed1ab_0
152
+ - jsonschema-specifications=2023.12.1=pyhd8ed1ab_0
153
+ - jsonschema-with-format-nongpl=4.20.0=pyhd8ed1ab_0
154
+ - jupyter_client=8.6.0=pyhd8ed1ab_0
155
+ - jupyter_core=5.7.1=py310hff52083_0
156
+ - jupyter_events=0.9.0=pyhd8ed1ab_0
157
+ - jupyter_server_terminals=0.5.1=pyhd8ed1ab_0
158
+ - jupyterlab_pygments=0.3.0=pyhd8ed1ab_0
159
+ - keyutils=1.6.1=h166bdaf_0
160
+ - kiwisolver=1.4.5=py310hd41b1e2_1
161
+ - krb5=1.21.2=h659d440_0
162
+ - langcodes=3.3.0=pyhd8ed1ab_0
163
+ - lcms2=2.16=hb7c19ff_0
164
+ - ld_impl_linux-64=2.40=h41732ed_0
165
+ - lerc=4.0.0=h27087fc_0
166
+ - libabseil=20230125.0=cxx17_hcb278e6_1
167
+ - libarchive=3.6.2=h039dbb9_1
168
+ - libarrow=11.0.0=h33598ff_9_cpu
169
+ - libblas=3.9.0=21_linux64_openblas
170
+ - libbrotlicommon=1.0.9=h166bdaf_9
171
+ - libbrotlidec=1.0.9=h166bdaf_9
172
+ - libbrotlienc=1.0.9=h166bdaf_9
173
+ - libcblas=3.9.0=21_linux64_openblas
174
+ - libcrc32c=1.1.2=h9c3ff4c_0
175
+ - libcublas=12.1.3.1=hd3aeb46_0
176
+ - libcublas-dev=12.1.3.1=0
177
+ - libcudf=23.08.00=cuda12_230809_g8150d38e08_0
178
+ - libcufft=11.0.2.54=hd3aeb46_0
179
+ - libcufile=1.9.0.20=0
180
+ - libcufile-dev=1.9.0.20=0
181
+ - libcuml=23.08.00=cuda12_230809_gd7162cdea_0
182
+ - libcumlprims=23.08.00=cuda12_230809_g71c0a86_0
183
+ - libcurand=10.3.2.106=hd3aeb46_0
184
+ - libcurand-dev=10.3.2.106=0
185
+ - libcurl=8.6.0=hca28451_0
186
+ - libcusolver=11.4.5.107=hd3aeb46_0
187
+ - libcusolver-dev=11.4.5.107=0
188
+ - libcusparse=12.1.0.106=hd3aeb46_0
189
+ - libcusparse-dev=12.1.0.106=0
190
+ - libdeflate=1.19=hd590300_0
191
+ - libedit=3.1.20191231=he28a2e2_2
192
+ - libev=4.33=hd590300_2
193
+ - libevent=2.1.10=h28343ad_4
194
+ - libexpat=2.6.2=h59595ed_0
195
+ - libffi=3.4.2=h7f98852_5
196
+ - libgcc-ng=13.2.0=h807b86a_3
197
+ - libgd=2.3.3=h119a65a_9
198
+ - libgfortran-ng=13.2.0=h69a702a_5
199
+ - libgfortran5=13.2.0=ha4646dd_5
200
+ - libglib=2.80.0=hf2295e7_0
201
+ - libgomp=13.2.0=h807b86a_3
202
+ - libgoogle-cloud=2.8.0=h3c06191_0
203
+ - libgrpc=1.51.1=hcf146ea_3
204
+ - libhwloc=2.9.3=default_h554bfaf_1009
205
+ - libiconv=1.17=hd590300_2
206
+ - libjpeg-turbo=3.0.0=hd590300_1
207
+ - libkvikio=23.08.00=cuda12_230809_g51a9036_0
208
+ - liblapack=3.9.0=21_linux64_openblas
209
+ - libllvm14=14.0.6=hcd5def8_4
210
+ - libmamba=1.5.0=h658169a_0
211
+ - libmambapy=1.5.0=py310h8aae740_0
212
+ - libnghttp2=1.58.0=h47da74e_1
213
+ - libnsl=2.0.1=hd590300_0
214
+ - libnuma=2.0.18=hd590300_0
215
+ - libnvjitlink=12.1.105=hd3aeb46_0
216
+ - libopenblas=0.3.26=pthreads_h413a1c8_0
217
+ - libpng=1.6.43=h2797004_0
218
+ - libprotobuf=3.21.12=hfc55251_2
219
+ - libraft=23.08.00=cuda12_230809_ge588d7b5_0
220
+ - libraft-headers=23.08.00=cuda12_230809_ge588d7b5_0
221
+ - libraft-headers-only=23.08.00=cuda12_230809_ge588d7b5_0
222
+ - librmm=23.08.00=cuda12_230809_gf3af0e8d_0
223
+ - librsvg=2.56.3=he3f83f7_1
224
+ - libsodium=1.0.18=h36c2ea0_1
225
+ - libsolv=0.7.27=hfc55251_0
226
+ - libsqlite=3.44.2=h2797004_0
227
+ - libssh2=1.11.0=h0841786_0
228
+ - libstdcxx-ng=13.2.0=h7e041cc_3
229
+ - libthrift=0.18.0=h5e4af38_0
230
+ - libtiff=4.6.0=ha9c0a0a_2
231
+ - libutf8proc=2.8.0=h166bdaf_0
232
+ - libuuid=2.38.1=h0b41bf4_0
233
+ - libuv=1.46.0=hd590300_0
234
+ - libwebp=1.3.2=h658648e_1
235
+ - libwebp-base=1.3.2=hd590300_0
236
+ - libxcb=1.15=h0b41bf4_0
237
+ - libxcrypt=4.4.36=hd590300_1
238
+ - libxml2=2.12.6=h232c23b_0
239
+ - libzlib=1.2.13=hd590300_5
240
+ - llvm-openmp=8.0.1=hc9558a2_0
241
+ - locket=1.0.0=pyhd8ed1ab_0
242
+ - lz4=4.3.3=py310h350c4a5_0
243
+ - lz4-c=1.9.4=hcb278e6_0
244
+ - lzo=2.10=h516909a_1000
245
+ - magma-cuda121=2.6.1=1
246
+ - mamba=1.5.0=py310h51d5547_0
247
+ - markdown-it-py=3.0.0=pyhd8ed1ab_0
248
+ - matplotlib-base=3.8.3=py310h62c0568_0
249
+ - matplotlib-inline=0.1.6=pyhd8ed1ab_0
250
+ - mdurl=0.1.2=pyhd8ed1ab_0
251
+ - menuinst=2.0.1=py310hff52083_0
252
+ - mkl=2023.1.0=h213fc3f_46344
253
+ - msgpack-python=1.0.7=py310hd41b1e2_0
254
+ - multidict=6.0.4=py310h2372a71_1
255
+ - munkres=1.1.4=pyh9f0ad1d_0
256
+ - murmurhash=1.0.10=py310hc6cd4ac_1
257
+ - nb_conda=2.2.1=unix_7
258
+ - nb_conda_kernels=2.3.1=pyhd8ed1ab_3
259
+ - nbclassic=1.0.0=pyhb4ecaf3_1
260
+ - nbconvert-pandoc=7.14.0=pyhd8ed1ab_0
261
+ - nbformat=5.9.2=pyhd8ed1ab_0
262
+ - nccl=2.20.5.1=h3a97aeb_0
263
+ - ncurses=6.4=h59595ed_2
264
+ - nest-asyncio=1.5.8=pyhd8ed1ab_0
265
+ - nodejs=20.9.0=hb753e55_0
266
+ - notebook-shim=0.2.3=pyhd8ed1ab_0
267
+ - numpy=1.26.4=py310hb13e2d6_0
268
+ - nvcomp=2.6.1=h10b603f_3
269
+ - nvtx=0.2.10=py310h2372a71_0
270
+ - openjpeg=2.5.2=h488ebb8_0
271
+ - openmp=8.0.1=0
272
+ - openssl=3.2.1=hd590300_0
273
+ - orc=1.8.2=hfdbbad2_2
274
+ - overrides=7.4.0=pyhd8ed1ab_0
275
+ - pandoc=3.1.3=h32600fe_0
276
+ - pandocfilters=1.5.0=pyhd8ed1ab_0
277
+ - pango=1.52.1=ha41ecd1_0
278
+ - parquet-cpp=1.5.1=2
279
+ - parso=0.8.3=pyhd8ed1ab_0
280
+ - partd=1.4.1=pyhd8ed1ab_0
281
+ - pathy=0.10.3=py310h06a4308_0
282
+ - pcre2=10.43=hcad00b1_0
283
+ - perl=5.32.1=7_hd590300_perl5
284
+ - pickleshare=0.7.5=py_1003
285
+ - pip=23.3.2=pyhd8ed1ab_0
286
+ - pixman=0.43.2=h59595ed_0
287
+ - pkg-config=0.29.2=h36c2ea0_1008
288
+ - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_1
289
+ - preshed=3.0.9=py310hc6cd4ac_1
290
+ - proj=9.3.1=h1d62c97_0
291
+ - prometheus_client=0.19.0=pyhd8ed1ab_0
292
+ - proto-plus=1.23.0=pyhd8ed1ab_0
293
+ - pthread-stubs=0.4=h36c2ea0_1001
294
+ - ptyprocess=0.7.0=pyhd3deb0d_0
295
+ - pure_eval=0.2.2=pyhd8ed1ab_0
296
+ - pyarrow=11.0.0=py310h633f555_9_cpu
297
+ - pyasn1=0.5.1=pyhd8ed1ab_0
298
+ - pyasn1-modules=0.3.0=pyhd8ed1ab_0
299
+ - pybind11-abi=4=hd8ed1ab_3
300
+ - pycosat=0.6.6=py310h2372a71_0
301
+ - pygments=2.17.2=pyhd8ed1ab_0
302
+ - pylibraft=23.08.00=cuda12_py310_230809_ge588d7b5_0
303
+ - pynvml=11.4.1=pyhd8ed1ab_0
304
+ - pyopenssl=23.3.0=pyhd8ed1ab_0
305
+ - pyproj=3.6.1=py310hd5c30f3_5
306
+ - pyshp=2.3.1=pyhd8ed1ab_0
307
+ - pysocks=1.7.1=py310h06a4308_0
308
+ - python=3.10.13=hd12c33a_1_cpython
309
+ - python-fastjsonschema=2.19.1=pyhd8ed1ab_0
310
+ - python-json-logger=2.0.7=pyhd8ed1ab_0
311
+ - python_abi=3.10=4_cp310
312
+ - pyu2f=0.1.5=pyhd8ed1ab_0
313
+ - pyyaml=6.0.1=py310h2372a71_1
314
+ - raft-dask=23.08.00=cuda12_py310_230809_ge588d7b5_0
315
+ - rdma-core=28.9=h59595ed_1
316
+ - re2=2023.02.02=hcb278e6_0
317
+ - readline=8.2=h8228510_1
318
+ - referencing=0.32.1=pyhd8ed1ab_0
319
+ - reproc=14.2.4.post0=hd590300_1
320
+ - reproc-cpp=14.2.4.post0=h59595ed_1
321
+ - requests=2.31.0=pyhd8ed1ab_0
322
+ - rfc3339-validator=0.1.4=pyhd8ed1ab_0
323
+ - rfc3986-validator=0.1.1=pyh9f0ad1d_0
324
+ - rmm=23.08.00=cuda12_py310_230809_gf3af0e8d_0
325
+ - rpds-py=0.16.2=py310hcb5633a_0
326
+ - rsa=4.9=pyhd8ed1ab_0
327
+ - ruamel.yaml=0.17.40=py310h2372a71_0
328
+ - ruamel.yaml.clib=0.2.7=py310h2372a71_2
329
+ - ruamel_yaml=0.15.100=py310h7f8727e_0
330
+ - s2n=1.3.41=h3358134_0
331
+ - send2trash=1.8.2=pyh41d4057_0
332
+ - setuptools=69.0.3=pyhd8ed1ab_0
333
+ - shellingham=1.5.4=pyhd8ed1ab_0
334
+ - smart_open=6.4.0=pyhd8ed1ab_0
335
+ - snappy=1.1.10=h9fff704_0
336
+ - sniffio=1.3.0=pyhd8ed1ab_0
337
+ - sortedcontainers=2.4.0=pyhd8ed1ab_0
338
+ - soupsieve=2.5=pyhd8ed1ab_1
339
+ - spacy=3.7.2=py310hcb52e73_0
340
+ - spacy-legacy=3.0.12=pyhd8ed1ab_0
341
+ - spacy-loggers=1.0.5=pyhd8ed1ab_0
342
+ - spdlog=1.11.0=h9b3ece8_1
343
+ - sqlite=3.38.2=hc218d9a_0
344
+ - srsly=2.4.8=py310hc6cd4ac_1
345
+ - stack_data=0.6.2=pyhd8ed1ab_0
346
+ - tblib=3.0.0=pyhd8ed1ab_0
347
+ - terminado=0.18.0=pyh0d859eb_0
348
+ - thinc=8.2.2=py310hcb52e73_0
349
+ - tinycss2=1.2.1=pyhd8ed1ab_0
350
+ - tk=8.6.13=noxft_h4845f30_101
351
+ - toolz=0.12.1=pyhd8ed1ab_0
352
+ - tornado=6.3.3=py310h2372a71_1
353
+ - tqdm=4.66.1=pyhd8ed1ab_0
354
+ - traitlets=5.9.0=pyhd8ed1ab_0
355
+ - treelite=3.2.0=py310h1be96d9_0
356
+ - truststore=0.8.0=pyhd8ed1ab_0
357
+ - typer=0.9.0=pyhd8ed1ab_0
358
+ - types-python-dateutil=2.8.19.20240106=pyhd8ed1ab_0
359
+ - typing-extensions=4.9.0=hd8ed1ab_0
360
+ - typing_extensions=4.9.0=pyha770c72_0
361
+ - typing_utils=0.1.0=pyhd8ed1ab_0
362
+ - ucx=1.14.1=h195a15c_5
363
+ - ucx-proc=1.0.0=gpu
364
+ - ucx-py=0.33.00=py310_230809_gea1eb8f_0
365
+ - unicodedata2=15.1.0=py310h2372a71_0
366
+ - uri-template=1.3.0=pyhd8ed1ab_0
367
+ - wasabi=1.1.2=py310hff52083_0
368
+ - wcwidth=0.2.13=pyhd8ed1ab_0
369
+ - weasel=0.3.4=pyhd8ed1ab_0
370
+ - webcolors=1.13=pyhd8ed1ab_0
371
+ - webencodings=0.5.1=pyhd8ed1ab_2
372
+ - websocket-client=1.7.0=pyhd8ed1ab_0
373
+ - wheel=0.42.0=pyhd8ed1ab_0
374
+ - xorg-kbproto=1.0.7=h7f98852_1002
375
+ - xorg-libice=1.1.1=hd590300_0
376
+ - xorg-libsm=1.2.4=h7391055_0
377
+ - xorg-libx11=1.8.7=h8ee46fc_0
378
+ - xorg-libxau=1.0.11=hd590300_0
379
+ - xorg-libxdmcp=1.1.3=h7f98852_0
380
+ - xorg-libxext=1.3.4=h0b41bf4_2
381
+ - xorg-libxrender=0.9.11=hd590300_0
382
+ - xorg-libxt=1.3.0=hd590300_1
383
+ - xorg-renderproto=0.11.1=h7f98852_1002
384
+ - xorg-xextproto=7.3.0=h0b41bf4_1003
385
+ - xorg-xproto=7.0.31=h7f98852_1007
386
+ - xyzservices=2023.10.1=pyhd8ed1ab_0
387
+ - xz=5.2.6=h166bdaf_0
388
+ - yaml=0.2.5=h7b6447c_0
389
+ - yaml-cpp=0.7.0=h59595ed_3
390
+ - zeromq=4.3.5=h59595ed_0
391
+ - zict=3.0.0=pyhd8ed1ab_0
392
+ - zipp=3.17.0=pyhd8ed1ab_0
393
+ - zlib=1.2.13=hd590300_5
394
+ - zstandard=0.22.0=py310h1275a96_0
395
+ - zstd=1.5.5=hfc55251_0
396
+ - pip:
397
+ - absl-py==1.4.0
398
+ - accelerate==0.28.0
399
+ - access==1.1.9
400
+ - affine==2.4.0
401
+ - aiobotocore==2.12.1
402
+ - aiofiles==22.1.0
403
+ - aiohttp-cors==0.7.0
404
+ - aioitertools==0.11.0
405
+ - aiorwlock==1.3.0
406
+ - aiosqlite==0.19.0
407
+ - albumentations==1.4.0
408
+ - alembic==1.13.1
409
+ - altair==5.2.0
410
+ - annoy==1.17.3
411
+ - apache-beam==2.46.0
412
+ - aplus==0.11.0
413
+ - appdirs==1.4.4
414
+ - array-record==0.5.0
415
+ - arviz==0.17.1
416
+ - astroid==3.0.3
417
+ - astropy==6.0.0
418
+ - astropy-iers-data==0.2024.3.18.0.29.47
419
+ - astunparse==1.6.3
420
+ - async-lru==2.0.4
421
+ - audioread==3.0.1
422
+ - auto-gptq==0.7.1
423
+ - autopep8==2.0.4
424
+ - babel==2.14.0
425
+ - backoff==2.2.1
426
+ - bayesian-optimization==1.4.3
427
+ - beatrix-jupyterlab==2023.128.151533
428
+ - bidict==0.23.1
429
+ - bitsandbytes==0.43.0
430
+ - blake3==0.2.1
431
+ - blessed==1.20.0
432
+ - blinker==1.7.0
433
+ - blosc2==2.5.1
434
+ - bokeh==3.3.4
435
+ - boruta==0.3
436
+ - boto3==1.26.100
437
+ - botocore==1.34.51
438
+ - bqplot==0.12.43
439
+ - branca==0.7.1
440
+ - brewer2mpl==1.4.1
441
+ - cachetools==4.2.4
442
+ - catalyst==22.4
443
+ - catboost==1.2.3
444
+ - category-encoders==2.6.3
445
+ - cesium==0.12.1
446
+ - chex==0.1.85
447
+ - cleverhans==4.0.0
448
+ - click-plugins==1.1.1
449
+ - cligj==0.7.2
450
+ - cloud-tpu-client==0.10
451
+ - cloud-tpu-profiler==2.4.0
452
+ - cloudpickle==2.2.1
453
+ - cmdstanpy==1.2.1
454
+ - cmudict==1.0.21
455
+ - colorcet==3.1.0
456
+ - coloredlogs==15.0.1
457
+ - colorful==0.5.6
458
+ - colorlog==6.8.2
459
+ - colorlover==0.3.0
460
+ - contextily==1.5.2
461
+ - convertdate==2.4.0
462
+ - cramjam==2.8.3
463
+ - crcmod==1.7
464
+ - cufflinks==0.17.3
465
+ - cvxcanon==0.1.2
466
+ - cython==3.0.8
467
+ - daal==2024.1.0
468
+ - daal4py==2024.1.0
469
+ - dacite==1.8.1
470
+ - dask==2024.3.1
471
+ - dask-expr==1.0.4
472
+ - dataclasses-json==0.6.4
473
+ - dataproc-jupyter-plugin==0.1.66
474
+ - datasets==2.1.0
475
+ - datashader==0.16.0
476
+ - datatile==1.0.3
477
+ - db-dtypes==1.2.0
478
+ - deap==1.4.1
479
+ - deepdiff==6.7.1
480
+ - deprecated==1.2.14
481
+ - deprecation==2.1.0
482
+ - descartes==1.1.0
483
+ - dill==0.3.8
484
+ - dipy==1.9.0
485
+ - distlib==0.3.8
486
+ - dm-tree==0.1.8
487
+ - docker==7.0.0
488
+ - docker-pycreds==0.4.0
489
+ - docopt==0.6.2
490
+ - docstring-parser==0.15
491
+ - docstring-to-markdown==0.15
492
+ - docutils==0.20.1
493
+ - earthengine-api==0.1.394
494
+ - easydict==1.13
495
+ - easyocr==1.7.1
496
+ - ecos==2.0.13
497
+ - eli5==0.13.0
498
+ - emoji==2.10.1
499
+ - en-core-web-lg==3.7.1
500
+ - en-core-web-sm==3.7.1
501
+ - ephem==4.1.5
502
+ - esda==2.5.1
503
+ - essentia==2.1b6.dev1110
504
+ - et-xmlfile==1.1.0
505
+ - etils==1.6.0
506
+ - explainable-ai-sdk==1.3.3
507
+ - farama-notifications==0.0.4
508
+ - fastai==2.7.14
509
+ - fastapi==0.108.0
510
+ - fastavro==1.9.3
511
+ - fastcore==1.5.29
512
+ - fastdownload==0.0.7
513
+ - fasteners==0.19
514
+ - fastparquet==2024.2.0
515
+ - fastprogress==1.0.3
516
+ - fasttext==0.9.2
517
+ - feather-format==0.4.1
518
+ - featuretools==1.30.0
519
+ - filelock==3.13.1
520
+ - fiona==1.9.6
521
+ - fitter==1.7.0
522
+ - flake8==7.0.0
523
+ - flashtext==2.7
524
+ - flask==3.0.2
525
+ - flatbuffers==23.5.26
526
+ - flax==0.8.2
527
+ - folium==0.16.0
528
+ - fonttools==4.47.0
529
+ - frozendict==2.4.0
530
+ - funcy==2.0
531
+ - fury==0.10.0
532
+ - future==1.0.0
533
+ - fuzzywuzzy==0.18.0
534
+ - gast==0.5.4
535
+ - gatspy==0.3
536
+ - gcsfs==2023.12.2.post1
537
+ - gekko==1.1.0
538
+ - gensim==4.3.2
539
+ - geographiclib==2.0
540
+ - geohash==1.0
541
+ - geojson==3.1.0
542
+ - geopandas==0.14.3
543
+ - geoplot==0.5.1
544
+ - geopy==2.4.1
545
+ - geoviews==1.11.1
546
+ - ggplot==0.11.5
547
+ - giddy==2.3.5
548
+ - gitdb==4.0.11
549
+ - gitpython==3.1.41
550
+ - google-ai-generativelanguage==0.4.0
551
+ - google-api-core==2.17.1
552
+ - google-api-python-client==2.122.0
553
+ - google-apitools==0.5.31
554
+ - google-auth-httplib2==0.1.1
555
+ - google-auth-oauthlib==1.2.0
556
+ - google-cloud-aiplatform==0.6.0a1
557
+ - google-cloud-artifact-registry==1.10.0
558
+ - google-cloud-automl==1.0.1
559
+ - google-cloud-bigquery==2.34.4
560
+ - google-cloud-bigtable==1.7.3
561
+ - google-cloud-dlp==3.14.0
562
+ - google-cloud-jupyter-config==0.0.5
563
+ - google-cloud-language==2.13.3
564
+ - google-cloud-monitoring==2.18.0
565
+ - google-cloud-pubsub==2.19.0
566
+ - google-cloud-pubsublite==1.9.0
567
+ - google-cloud-recommendations-ai==0.7.1
568
+ - google-cloud-resource-manager==1.11.0
569
+ - google-cloud-spanner==3.40.1
570
+ - google-cloud-storage==1.44.0
571
+ - google-cloud-translate==3.12.1
572
+ - google-cloud-videointelligence==2.13.3
573
+ - google-cloud-vision==2.8.0
574
+ - google-crc32c==1.5.0
575
+ - google-generativeai==0.4.1
576
+ - google-pasta==0.2.0
577
+ - google-resumable-media==2.7.0
578
+ - gplearn==0.4.2
579
+ - gpustat==1.0.0
580
+ - gpxpy==1.6.2
581
+ - greenlet==3.0.3
582
+ - grpc-google-iam-v1==0.12.7
583
+ - grpcio==1.60.0
584
+ - grpcio-status==1.48.2
585
+ - gviz-api==1.10.0
586
+ - gym==0.26.2
587
+ - gym-notices==0.0.8
588
+ - gymnasium==0.29.0
589
+ - h11==0.14.0
590
+ - h2o==3.46.0.1
591
+ - h5netcdf==1.3.0
592
+ - h5py==3.10.0
593
+ - haversine==2.8.1
594
+ - hdfs==2.7.3
595
+ - hep-ml==0.7.2
596
+ - hijri-converter==2.3.1
597
+ - hmmlearn==0.3.2
598
+ - holidays==0.24
599
+ - holoviews==1.18.3
600
+ - hpsklearn==0.1.0
601
+ - html5lib==1.1
602
+ - htmlmin==0.1.12
603
+ - httpcore==1.0.4
604
+ - httplib2==0.21.0
605
+ - httptools==0.6.1
606
+ - httpx==0.27.0
607
+ - huggingface-hub==0.21.4
608
+ - humanfriendly==10.0
609
+ - hunspell==0.5.5
610
+ - husl==4.0.3
611
+ - hydra-slayer==0.5.0
612
+ - hyperopt==0.2.7
613
+ - hypertools==0.8.0
614
+ - igraph==0.11.4
615
+ - imagecodecs==2024.1.1
616
+ - imagehash==4.3.1
617
+ - imageio==2.33.1
618
+ - imbalanced-learn==0.12.0
619
+ - imgaug==0.4.0
620
+ - importlib-metadata==6.11.0
621
+ - inequality==1.0.1
622
+ - iniconfig==2.0.0
623
+ - ipydatawidgets==4.3.5
624
+ - ipyleaflet==0.18.2
625
+ - ipympl==0.7.0
626
+ - ipython-genutils==0.2.0
627
+ - ipython-sql==0.5.0
628
+ - ipyvolume==0.6.3
629
+ - ipyvue==1.10.2
630
+ - ipyvuetify==1.9.2
631
+ - ipywebrtc==0.6.0
632
+ - ipywidgets==7.7.1
633
+ - isort==5.13.2
634
+ - isoweek==1.3.3
635
+ - itsdangerous==2.1.2
636
+ - janome==0.5.0
637
+ - jaraco-classes==3.3.0
638
+ - jax==0.4.23
639
+ - jax-jumpy==1.0.0
640
+ - jaxlib==0.4.23.dev20240116
641
+ - jeepney==0.8.0
642
+ - jieba==0.42.1
643
+ - jmespath==1.0.1
644
+ - json5==0.9.14
645
+ - jupyter-client==7.4.9
646
+ - jupyter-console==6.6.3
647
+ - jupyter-http-over-ws==0.0.8
648
+ - jupyter-lsp==1.5.1
649
+ - jupyter-server==2.13.0
650
+ - jupyter-server-fileid==0.9.1
651
+ - jupyter-server-mathjax==0.2.6
652
+ - jupyter-server-proxy==4.1.0
653
+ - jupyter-server-ydoc==0.8.0
654
+ - jupyter-ydoc==0.2.5
655
+ - jupyterlab==4.1.5
656
+ - jupyterlab-git==0.44.0
657
+ - jupyterlab-lsp==5.1.0
658
+ - jupyterlab-server==2.25.2
659
+ - jupyterlab-widgets==3.0.9
660
+ - jupytext==1.16.0
661
+ - kaggle==1.6.6
662
+ - kaggle-environments==1.14.3
663
+ - kagglehub==0.2.0
664
+ - keras==3.0.5
665
+ - keras-cv==0.8.2
666
+ - keras-nlp==0.8.2
667
+ - keras-tuner==1.4.6
668
+ - kernels-mixer==0.0.7
669
+ - keyring==24.3.0
670
+ - keyrings-google-artifactregistry-auth==1.1.2
671
+ - kfp==2.5.0
672
+ - kfp-pipeline-spec==0.2.2
673
+ - kfp-server-api==2.0.5
674
+ - kmapper==2.0.1
675
+ - kmodes==0.12.2
676
+ - korean-lunar-calendar==0.3.1
677
+ - kornia==0.7.2
678
+ - kornia-rs==0.1.2
679
+ - kt-legacy==1.0.5
680
+ - kubernetes==26.1.0
681
+ - langid==1.1.6
682
+ - lazy-loader==0.3
683
+ - learntools==0.3.4
684
+ - leven==1.0.4
685
+ - levenshtein==0.25.0
686
+ - libclang==16.0.6
687
+ - libpysal==4.9.2
688
+ - librosa==0.10.1
689
+ - lightgbm==4.2.0
690
+ - lightning-utilities==0.10.1
691
+ - lime==0.2.0.1
692
+ - line-profiler==4.1.2
693
+ - linkify-it-py==2.0.3
694
+ - llvmlite==0.41.1
695
+ - lml==0.1.0
696
+ - loguru==0.7.2
697
+ - lunarcalendar==0.0.9
698
+ - lxml==5.1.0
699
+ - mako==1.3.2
700
+ - mapclassify==2.6.1
701
+ - markdown==3.5.2
702
+ - markovify==0.9.4
703
+ - markupsafe==2.1.5
704
+ - marshmallow==3.21.1
705
+ - matplotlib==3.7.5
706
+ - matplotlib-venn==0.11.10
707
+ - mccabe==0.7.0
708
+ - mdit-py-plugins==0.4.0
709
+ - memory-profiler==0.61.0
710
+ - mercantile==1.2.1
711
+ - mgwr==2.2.1
712
+ - missingno==0.5.2
713
+ - mistune==0.8.4
714
+ - mizani==0.11.0
715
+ - ml-dtypes==0.2.0
716
+ - mlcrate==0.2.0
717
+ - mlens==0.2.3
718
+ - mlxtend==0.23.1
719
+ - mmh3==4.1.0
720
+ - mne==1.6.1
721
+ - mnist==0.2.2
722
+ - mock==5.1.0
723
+ - momepy==0.7.0
724
+ - more-itertools==10.2.0
725
+ - mpld3==0.5.10
726
+ - mpmath==1.3.0
727
+ - msgpack-numpy==0.4.8
728
+ - multimethod==1.10
729
+ - multipledispatch==1.0.0
730
+ - multiprocess==0.70.16
731
+ - mypy-extensions==1.0.0
732
+ - namex==0.0.7
733
+ - nbclient==0.5.13
734
+ - nbconvert==6.4.5
735
+ - nbdime==3.2.0
736
+ - ndindex==1.8
737
+ - networkx==3.2.1
738
+ - nibabel==5.2.1
739
+ - nilearn==0.10.3
740
+ - ninja==1.11.1.1
741
+ - nltk==3.2.4
742
+ - nose==1.3.7
743
+ - notebook==6.5.6
744
+ - notebook-executor==0.2
745
+ - numba==0.58.1
746
+ - numexpr==2.9.0
747
+ - nvidia-ml-py==11.495.46
748
+ - oauth2client==4.1.3
749
+ - oauthlib==3.2.2
750
+ - objsize==0.6.1
751
+ - odfpy==1.4.1
752
+ - olefile==0.47
753
+ - onnx==1.15.0
754
+ - opencensus==0.11.4
755
+ - opencensus-context==0.1.3
756
+ - opencv-contrib-python==4.9.0.80
757
+ - opencv-python==4.9.0.80
758
+ - opencv-python-headless==4.9.0.80
759
+ - openpyxl==3.1.2
760
+ - openslide-python==1.3.1
761
+ - opentelemetry-api==1.22.0
762
+ - opentelemetry-exporter-otlp==1.22.0
763
+ - opentelemetry-exporter-otlp-proto-common==1.22.0
764
+ - opentelemetry-exporter-otlp-proto-grpc==1.22.0
765
+ - opentelemetry-exporter-otlp-proto-http==1.22.0
766
+ - opentelemetry-proto==1.22.0
767
+ - opentelemetry-sdk==1.22.0
768
+ - opentelemetry-semantic-conventions==0.43b0
769
+ - opt-einsum==3.3.0
770
+ - optax==0.2.1
771
+ - optimum==1.18.0
772
+ - optuna==3.6.0
773
+ - orbax-checkpoint==0.5.6
774
+ - ordered-set==4.1.0
775
+ - orjson==3.9.10
776
+ - ortools==9.4.1874
777
+ - osmnx==1.9.1
778
+ - packaging==21.3
779
+ - pandas==2.2.1
780
+ - pandas-datareader==0.10.0
781
+ - pandas-profiling==3.6.6
782
+ - pandas-summary==0.2.0
783
+ - pandasql==0.7.3
784
+ - panel==1.3.8
785
+ - papermill==2.5.0
786
+ - param==2.0.2
787
+ - path==16.10.0
788
+ - path-py==12.5.0
789
+ - pathos==0.3.2
790
+ - patsy==0.5.6
791
+ - pdf2image==1.17.0
792
+ - peft==0.10.0
793
+ - pettingzoo==1.24.0
794
+ - pexpect==4.9.0
795
+ - phik==0.12.4
796
+ - pillow==9.5.0
797
+ - platformdirs==4.2.0
798
+ - plotly==5.18.0
799
+ - plotly-express==0.4.1
800
+ - plotnine==0.13.2
801
+ - pluggy==1.4.0
802
+ - pointpats==2.4.0
803
+ - polars==0.20.15
804
+ - polyglot==16.7.4
805
+ - pooch==1.8.1
806
+ - pox==0.3.4
807
+ - ppca==0.0.4
808
+ - ppft==1.7.6.8
809
+ - preprocessing==0.1.13
810
+ - prettytable==3.9.0
811
+ - progressbar2==4.4.2
812
+ - promise==2.3
813
+ - prompt-toolkit==3.0.43
814
+ - pronouncing==0.2.0
815
+ - prophet==1.1.1
816
+ - protobuf==3.20.3
817
+ - psutil==5.9.3
818
+ - pudb==2024.1
819
+ - pulp==2.8.0
820
+ - py-cpuinfo==9.0.0
821
+ - py-spy==0.3.14
822
+ - py4j==0.10.9.7
823
+ - pyaml==23.12.0
824
+ - pyarabic==0.6.15
825
+ - pyastronomy==0.21.0
826
+ - pybind11==2.11.1
827
+ - pyclipper==1.3.0.post5
828
+ - pycodestyle==2.11.1
829
+ - pycparser==2.21
830
+ - pycryptodome==3.20.0
831
+ - pyct==0.5.0
832
+ - pycuda==2024.1
833
+ - pydantic==2.5.3
834
+ - pydantic-core==2.14.6
835
+ - pydegensac==0.1.2
836
+ - pydicom==2.4.4
837
+ - pydocstyle==6.3.0
838
+ - pydot==1.4.2
839
+ - pydub==0.25.1
840
+ - pyemd==1.0.0
841
+ - pyerfa==2.0.1.1
842
+ - pyexcel-io==0.6.6
843
+ - pyexcel-ods==0.6.0
844
+ - pyflakes==3.2.0
845
+ - pygltflib==1.16.2
846
+ - pyjwt==2.8.0
847
+ - pykalman==0.9.5
848
+ - pyldavis==3.4.1
849
+ - pylint==3.0.4
850
+ - pymc3==3.11.4
851
+ - pymeeus==0.5.12
852
+ - pymongo==3.13.0
853
+ - pympler==1.0.1
854
+ - pynndescent==0.5.11
855
+ - pynvrtc==9.2
856
+ - pyocr==0.8.5
857
+ - pyparsing==3.1.1
858
+ - pypdf==4.1.0
859
+ - pysal==24.1
860
+ - pytesseract==0.3.10
861
+ - pytest==8.1.1
862
+ - python-bidi==0.4.2
863
+ - python-dateutil==2.9.0.post0
864
+ - python-dotenv==1.0.0
865
+ - python-graphviz==0.20.2
866
+ - python-levenshtein==0.25.0
867
+ - python-louvain==0.16
868
+ - python-lsp-jsonrpc==1.1.2
869
+ - python-lsp-server==1.10.1
870
+ - python-slugify==8.0.4
871
+ - python-utils==3.8.2
872
+ - pythreejs==2.4.2
873
+ - pytoolconfig==1.3.1
874
+ - pytools==2023.1.1
875
+ - pytorch-ignite==0.4.13
876
+ - pytorch-lightning==2.2.1
877
+ - pytz==2023.3.post1
878
+ - pyupset==0.1.1.post7
879
+ - pyviz-comms==3.0.1
880
+ - pywavelets==1.5.0
881
+ - pyzmq==24.0.1
882
+ - qgrid==1.3.1
883
+ - qtconsole==5.5.1
884
+ - qtpy==2.4.1
885
+ - quantecon==0.7.2
886
+ - quantities==0.15.0
887
+ - qudida==0.0.4
888
+ - rapidfuzz==3.6.2
889
+ - rasterio==1.3.9
890
+ - rasterstats==0.19.0
891
+ - ray==2.9.0
892
+ - ray-cpp==2.9.0
893
+ - regex==2023.12.25
894
+ - requests-oauthlib==1.3.1
895
+ - requests-toolbelt==0.10.1
896
+ - responses==0.18.0
897
+ - retrying==1.3.4
898
+ - rgf-python==3.12.0
899
+ - rich==13.7.0
900
+ - rich-click==1.7.4
901
+ - rope==1.12.0
902
+ - rouge==1.0.1
903
+ - rtree==1.2.0
904
+ - s2sphere==0.2.5
905
+ - s3fs==2024.3.0
906
+ - s3transfer==0.6.2
907
+ - safetensors==0.4.2
908
+ - scattertext==0.1.19
909
+ - scikit-image==0.22.0
910
+ - scikit-learn==1.2.2
911
+ - scikit-learn-intelex==2024.1.0
912
+ - scikit-multilearn==0.2.0
913
+ - scikit-optimize==0.10.1
914
+ - scikit-plot==0.3.7
915
+ - scikit-surprise==1.1.3
916
+ - scipy==1.11.4
917
+ - seaborn==0.12.2
918
+ - secretstorage==3.3.3
919
+ - segment-anything==1.0
920
+ - segregation==2.5
921
+ - semver==3.0.2
922
+ - sentencepiece==0.2.0
923
+ - sentry-sdk==1.42.0
924
+ - setproctitle==1.3.3
925
+ - setuptools-git==1.2
926
+ - setuptools-scm==8.0.4
927
+ - shap==0.44.1
928
+ - shapely==2.0.3
929
+ - shimmy==1.3.0
930
+ - simpervisor==1.0.0
931
+ - simpleitk==2.3.1
932
+ - simplejson==3.19.2
933
+ - six==1.16.0
934
+ - sklearn-pandas==2.2.0
935
+ - slicer==0.0.7
936
+ - smmap==5.0.1
937
+ - snowballstemmer==2.2.0
938
+ - snuggs==1.4.7
939
+ - soundfile==0.12.1
940
+ - soxr==0.3.7
941
+ - spaghetti==1.7.5.post1
942
+ - spectral==0.23.1
943
+ - spglm==1.1.0
944
+ - sphinx-rtd-theme==0.2.4
945
+ - spint==1.0.7
946
+ - splot==1.1.5.post1
947
+ - spopt==0.6.0
948
+ - spreg==1.4.2
949
+ - spvcm==0.3.0
950
+ - sqlalchemy==2.0.25
951
+ - sqlparse==0.4.4
952
+ - squarify==0.4.3
953
+ - stable-baselines3==2.1.0
954
+ - stack-data==0.6.3
955
+ - stanio==0.3.0
956
+ - starlette==0.32.0.post1
957
+ - statsmodels==0.14.1
958
+ - stemming==1.0.1
959
+ - stop-words==2018.7.23
960
+ - stopit==1.1.2
961
+ - stumpy==1.12.0
962
+ - sympy==1.12
963
+ - tables==3.9.2
964
+ - tabulate==0.9.0
965
+ - tangled-up-in-unicode==0.2.0
966
+ - tbb==2021.11.0
967
+ - tenacity==8.2.3
968
+ - tensorboard==2.15.1
969
+ - tensorboard-data-server==0.7.2
970
+ - tensorboard-plugin-profile==2.15.0
971
+ - tensorboardx==2.6.2.2
972
+ - tensorflow==2.15.0
973
+ - tensorflow-cloud==0.1.16
974
+ - tensorflow-datasets==4.9.4
975
+ - tensorflow-decision-forests==1.8.1
976
+ - tensorflow-estimator==2.15.0
977
+ - tensorflow-hub==0.16.1
978
+ - tensorflow-io==0.35.0
979
+ - tensorflow-io-gcs-filesystem==0.35.0
980
+ - tensorflow-metadata==0.14.0
981
+ - tensorflow-probability==0.23.0
982
+ - tensorflow-serving-api==2.14.1
983
+ - tensorflow-text==2.15.0
984
+ - tensorflow-transform==0.14.0
985
+ - tensorpack==0.11
986
+ - tensorstore==0.1.56
987
+ - termcolor==2.4.0
988
+ - testpath==0.6.0
989
+ - text-unidecode==1.3
990
+ - textblob==0.18.0.post0
991
+ - texttable==1.7.0
992
+ - tf-keras==2.15.1
993
+ - tfp-nightly==0.24.0.dev0
994
+ - theano==1.0.5
995
+ - theano-pymc==1.1.2
996
+ - threadpoolctl==3.2.0
997
+ - tifffile==2023.12.9
998
+ - timm==0.9.16
999
+ - tobler==0.11.2
1000
+ - tokenizers==0.15.2
1001
+ - toml==0.10.2
1002
+ - tomli==2.0.1
1003
+ - tomlkit==0.12.4
1004
+ - torch==2.1.2
1005
+ - torchaudio==2.1.2
1006
+ - torchdata==0.7.1
1007
+ - torchinfo==1.8.0
1008
+ - torchmetrics==1.3.2
1009
+ - torchtext==0.16.2
1010
+ - torchvision==0.16.2
1011
+ - tpot==0.12.1
1012
+ - traceml==1.0.8
1013
+ - traittypes==0.2.1
1014
+ - transformers==4.38.2
1015
+ - treelite-runtime==3.2.0
1016
+ - trueskill==0.4.5
1017
+ - trx-python==0.2.9
1018
+ - tsfresh==0.20.2
1019
+ - typeguard==4.1.5
1020
+ - typing-inspect==0.9.0
1021
+ - tzdata==2023.4
1022
+ - uc-micro-py==1.0.3
1023
+ - ujson==5.9.0
1024
+ - umap-learn==0.5.5
1025
+ - unidecode==1.3.8
1026
+ - update-checker==0.18.0
1027
+ - uritemplate==3.0.1
1028
+ - urllib3==1.26.18
1029
+ - urwid==2.6.9
1030
+ - urwid-readline==0.14
1031
+ - uvicorn==0.25.0
1032
+ - uvloop==0.19.0
1033
+ - vaex==4.17.0
1034
+ - vaex-astro==0.9.3
1035
+ - vaex-core==4.17.1
1036
+ - vaex-hdf5==0.14.1
1037
+ - vaex-jupyter==0.8.2
1038
+ - vaex-ml==0.18.3
1039
+ - vaex-server==0.9.0
1040
+ - vaex-viz==0.5.4
1041
+ - vec-noise==1.1.4
1042
+ - vecstack==0.4.0
1043
+ - virtualenv==20.21.0
1044
+ - visions==0.7.5
1045
+ - vowpalwabbit==9.9.0
1046
+ - vtk==9.3.0
1047
+ - wand==0.6.13
1048
+ - wandb==0.16.4
1049
+ - watchfiles==0.21.0
1050
+ - wavio==0.0.8
1051
+ - websockets==12.0
1052
+ - werkzeug==3.0.1
1053
+ - wfdb==4.1.2
1054
+ - whatthepatch==1.0.5
1055
+ - widgetsnbextension==3.6.6
1056
+ - witwidget==1.8.1
1057
+ - woodwork==0.29.0
1058
+ - wordcloud==1.9.3
1059
+ - wordsegment==1.3.1
1060
+ - wrapt==1.14.1
1061
+ - xarray==2024.2.0
1062
+ - xarray-einstats==0.7.0
1063
+ - xgboost==2.0.3
1064
+ - xvfbwrapper==0.2.9
1065
+ - xxhash==3.4.1
1066
+ - y-py==0.6.2
1067
+ - yapf==0.40.2
1068
+ - yarl==1.9.4
1069
+ - ydata-profiling==4.6.4
1070
+ - yellowbrick==1.5
1071
+ - ypy-websocket==0.8.4
1072
+ prefix: /opt/conda
wandb/run-20240401_093743-ie4jv7u5/files/config.yaml ADDED
@@ -0,0 +1,704 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.4
8
+ framework: huggingface
9
+ huggingface_version: 4.38.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1711964263.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 99
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 98
42
+ - 99
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.4
49
+ 6: 4.38.2
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ - 1: eval/loss
78
+ 5: 1
79
+ 6:
80
+ - 1
81
+ - 1: eval/runtime
82
+ 5: 1
83
+ 6:
84
+ - 1
85
+ - 1: eval/samples_per_second
86
+ 5: 1
87
+ 6:
88
+ - 1
89
+ - 1: eval/steps_per_second
90
+ 5: 1
91
+ 6:
92
+ - 1
93
+ vocab_size:
94
+ desc: null
95
+ value: 32000
96
+ max_position_embeddings:
97
+ desc: null
98
+ value: 32768
99
+ hidden_size:
100
+ desc: null
101
+ value: 4096
102
+ intermediate_size:
103
+ desc: null
104
+ value: 14336
105
+ num_hidden_layers:
106
+ desc: null
107
+ value: 32
108
+ num_attention_heads:
109
+ desc: null
110
+ value: 32
111
+ sliding_window:
112
+ desc: null
113
+ value: null
114
+ num_key_value_heads:
115
+ desc: null
116
+ value: 8
117
+ hidden_act:
118
+ desc: null
119
+ value: silu
120
+ initializer_range:
121
+ desc: null
122
+ value: 0.02
123
+ rms_norm_eps:
124
+ desc: null
125
+ value: 1.0e-05
126
+ use_cache:
127
+ desc: null
128
+ value: false
129
+ rope_theta:
130
+ desc: null
131
+ value: 1000000.0
132
+ attention_dropout:
133
+ desc: null
134
+ value: 0.0
135
+ return_dict:
136
+ desc: null
137
+ value: true
138
+ output_hidden_states:
139
+ desc: null
140
+ value: false
141
+ output_attentions:
142
+ desc: null
143
+ value: false
144
+ torchscript:
145
+ desc: null
146
+ value: false
147
+ torch_dtype:
148
+ desc: null
149
+ value: bfloat16
150
+ use_bfloat16:
151
+ desc: null
152
+ value: false
153
+ tf_legacy_loss:
154
+ desc: null
155
+ value: false
156
+ pruned_heads:
157
+ desc: null
158
+ value: {}
159
+ tie_word_embeddings:
160
+ desc: null
161
+ value: false
162
+ chunk_size_feed_forward:
163
+ desc: null
164
+ value: 0
165
+ is_encoder_decoder:
166
+ desc: null
167
+ value: false
168
+ is_decoder:
169
+ desc: null
170
+ value: false
171
+ cross_attention_hidden_size:
172
+ desc: null
173
+ value: null
174
+ add_cross_attention:
175
+ desc: null
176
+ value: false
177
+ tie_encoder_decoder:
178
+ desc: null
179
+ value: false
180
+ max_length:
181
+ desc: null
182
+ value: 20
183
+ min_length:
184
+ desc: null
185
+ value: 0
186
+ do_sample:
187
+ desc: null
188
+ value: false
189
+ early_stopping:
190
+ desc: null
191
+ value: false
192
+ num_beams:
193
+ desc: null
194
+ value: 1
195
+ num_beam_groups:
196
+ desc: null
197
+ value: 1
198
+ diversity_penalty:
199
+ desc: null
200
+ value: 0.0
201
+ temperature:
202
+ desc: null
203
+ value: 1.0
204
+ top_k:
205
+ desc: null
206
+ value: 50
207
+ top_p:
208
+ desc: null
209
+ value: 1.0
210
+ typical_p:
211
+ desc: null
212
+ value: 1.0
213
+ repetition_penalty:
214
+ desc: null
215
+ value: 1.0
216
+ length_penalty:
217
+ desc: null
218
+ value: 1.0
219
+ no_repeat_ngram_size:
220
+ desc: null
221
+ value: 0
222
+ encoder_no_repeat_ngram_size:
223
+ desc: null
224
+ value: 0
225
+ bad_words_ids:
226
+ desc: null
227
+ value: null
228
+ num_return_sequences:
229
+ desc: null
230
+ value: 1
231
+ output_scores:
232
+ desc: null
233
+ value: false
234
+ return_dict_in_generate:
235
+ desc: null
236
+ value: false
237
+ forced_bos_token_id:
238
+ desc: null
239
+ value: null
240
+ forced_eos_token_id:
241
+ desc: null
242
+ value: null
243
+ remove_invalid_values:
244
+ desc: null
245
+ value: false
246
+ exponential_decay_length_penalty:
247
+ desc: null
248
+ value: null
249
+ suppress_tokens:
250
+ desc: null
251
+ value: null
252
+ begin_suppress_tokens:
253
+ desc: null
254
+ value: null
255
+ architectures:
256
+ desc: null
257
+ value:
258
+ - MistralForCausalLM
259
+ finetuning_task:
260
+ desc: null
261
+ value: null
262
+ id2label:
263
+ desc: null
264
+ value:
265
+ '0': LABEL_0
266
+ '1': LABEL_1
267
+ label2id:
268
+ desc: null
269
+ value:
270
+ LABEL_0: 0
271
+ LABEL_1: 1
272
+ tokenizer_class:
273
+ desc: null
274
+ value: null
275
+ prefix:
276
+ desc: null
277
+ value: null
278
+ bos_token_id:
279
+ desc: null
280
+ value: 1
281
+ pad_token_id:
282
+ desc: null
283
+ value: 0
284
+ eos_token_id:
285
+ desc: null
286
+ value: 2
287
+ sep_token_id:
288
+ desc: null
289
+ value: null
290
+ decoder_start_token_id:
291
+ desc: null
292
+ value: null
293
+ task_specific_params:
294
+ desc: null
295
+ value: null
296
+ problem_type:
297
+ desc: null
298
+ value: null
299
+ _name_or_path:
300
+ desc: null
301
+ value: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
302
+ transformers_version:
303
+ desc: null
304
+ value: 4.38.2
305
+ model_type:
306
+ desc: null
307
+ value: mistral
308
+ pretraining_tp:
309
+ desc: null
310
+ value: 1
311
+ quantization_config:
312
+ desc: null
313
+ value:
314
+ quant_method: QuantizationMethod.GPTQ
315
+ bits: 4
316
+ tokenizer: null
317
+ dataset: null
318
+ group_size: 128
319
+ damp_percent: 0.1
320
+ desc_act: true
321
+ sym: true
322
+ true_sequential: true
323
+ use_cuda_fp16: false
324
+ model_seqlen: null
325
+ block_name_to_quantize: null
326
+ module_name_preceding_first_block: null
327
+ batch_size: 1
328
+ pad_token_id: null
329
+ use_exllama: true
330
+ max_input_length: null
331
+ exllama_config:
332
+ version: ExllamaVersion.ONE
333
+ cache_block_outputs: true
334
+ modules_in_block_to_quantize: null
335
+ output_dir:
336
+ desc: null
337
+ value: /kaggle/working/
338
+ overwrite_output_dir:
339
+ desc: null
340
+ value: false
341
+ do_train:
342
+ desc: null
343
+ value: false
344
+ do_eval:
345
+ desc: null
346
+ value: true
347
+ do_predict:
348
+ desc: null
349
+ value: false
350
+ evaluation_strategy:
351
+ desc: null
352
+ value: epoch
353
+ prediction_loss_only:
354
+ desc: null
355
+ value: false
356
+ per_device_train_batch_size:
357
+ desc: null
358
+ value: 5
359
+ per_device_eval_batch_size:
360
+ desc: null
361
+ value: 5
362
+ per_gpu_train_batch_size:
363
+ desc: null
364
+ value: null
365
+ per_gpu_eval_batch_size:
366
+ desc: null
367
+ value: null
368
+ gradient_accumulation_steps:
369
+ desc: null
370
+ value: 3
371
+ eval_accumulation_steps:
372
+ desc: null
373
+ value: null
374
+ eval_delay:
375
+ desc: null
376
+ value: 0
377
+ learning_rate:
378
+ desc: null
379
+ value: 0.0002
380
+ weight_decay:
381
+ desc: null
382
+ value: 0.01
383
+ adam_beta1:
384
+ desc: null
385
+ value: 0.9
386
+ adam_beta2:
387
+ desc: null
388
+ value: 0.999
389
+ adam_epsilon:
390
+ desc: null
391
+ value: 1.0e-08
392
+ max_grad_norm:
393
+ desc: null
394
+ value: 1.0
395
+ num_train_epochs:
396
+ desc: null
397
+ value: 50
398
+ max_steps:
399
+ desc: null
400
+ value: -1
401
+ lr_scheduler_type:
402
+ desc: null
403
+ value: linear
404
+ lr_scheduler_kwargs:
405
+ desc: null
406
+ value: {}
407
+ warmup_ratio:
408
+ desc: null
409
+ value: 0.0
410
+ warmup_steps:
411
+ desc: null
412
+ value: 1
413
+ log_level:
414
+ desc: null
415
+ value: passive
416
+ log_level_replica:
417
+ desc: null
418
+ value: warning
419
+ log_on_each_node:
420
+ desc: null
421
+ value: true
422
+ logging_dir:
423
+ desc: null
424
+ value: /kaggle/working/runs/Apr01_09-41-45_148e249fd8ed
425
+ logging_strategy:
426
+ desc: null
427
+ value: epoch
428
+ logging_first_step:
429
+ desc: null
430
+ value: false
431
+ logging_steps:
432
+ desc: null
433
+ value: 500
434
+ logging_nan_inf_filter:
435
+ desc: null
436
+ value: true
437
+ save_strategy:
438
+ desc: null
439
+ value: epoch
440
+ save_steps:
441
+ desc: null
442
+ value: 500
443
+ save_total_limit:
444
+ desc: null
445
+ value: null
446
+ save_safetensors:
447
+ desc: null
448
+ value: true
449
+ save_on_each_node:
450
+ desc: null
451
+ value: false
452
+ save_only_model:
453
+ desc: null
454
+ value: false
455
+ no_cuda:
456
+ desc: null
457
+ value: false
458
+ use_cpu:
459
+ desc: null
460
+ value: false
461
+ use_mps_device:
462
+ desc: null
463
+ value: false
464
+ seed:
465
+ desc: null
466
+ value: 42
467
+ data_seed:
468
+ desc: null
469
+ value: null
470
+ jit_mode_eval:
471
+ desc: null
472
+ value: false
473
+ use_ipex:
474
+ desc: null
475
+ value: false
476
+ bf16:
477
+ desc: null
478
+ value: false
479
+ fp16:
480
+ desc: null
481
+ value: true
482
+ fp16_opt_level:
483
+ desc: null
484
+ value: O1
485
+ half_precision_backend:
486
+ desc: null
487
+ value: auto
488
+ bf16_full_eval:
489
+ desc: null
490
+ value: false
491
+ fp16_full_eval:
492
+ desc: null
493
+ value: false
494
+ tf32:
495
+ desc: null
496
+ value: null
497
+ local_rank:
498
+ desc: null
499
+ value: 0
500
+ ddp_backend:
501
+ desc: null
502
+ value: null
503
+ tpu_num_cores:
504
+ desc: null
505
+ value: null
506
+ tpu_metrics_debug:
507
+ desc: null
508
+ value: false
509
+ debug:
510
+ desc: null
511
+ value: []
512
+ dataloader_drop_last:
513
+ desc: null
514
+ value: false
515
+ eval_steps:
516
+ desc: null
517
+ value: null
518
+ dataloader_num_workers:
519
+ desc: null
520
+ value: 0
521
+ dataloader_prefetch_factor:
522
+ desc: null
523
+ value: null
524
+ past_index:
525
+ desc: null
526
+ value: -1
527
+ run_name:
528
+ desc: null
529
+ value: /kaggle/working/
530
+ disable_tqdm:
531
+ desc: null
532
+ value: false
533
+ remove_unused_columns:
534
+ desc: null
535
+ value: true
536
+ label_names:
537
+ desc: null
538
+ value: null
539
+ load_best_model_at_end:
540
+ desc: null
541
+ value: true
542
+ metric_for_best_model:
543
+ desc: null
544
+ value: loss
545
+ greater_is_better:
546
+ desc: null
547
+ value: false
548
+ ignore_data_skip:
549
+ desc: null
550
+ value: false
551
+ fsdp:
552
+ desc: null
553
+ value: []
554
+ fsdp_min_num_params:
555
+ desc: null
556
+ value: 0
557
+ fsdp_config:
558
+ desc: null
559
+ value:
560
+ min_num_params: 0
561
+ xla: false
562
+ xla_fsdp_v2: false
563
+ xla_fsdp_grad_ckpt: false
564
+ fsdp_transformer_layer_cls_to_wrap:
565
+ desc: null
566
+ value: null
567
+ accelerator_config:
568
+ desc: null
569
+ value:
570
+ split_batches: false
571
+ dispatch_batches: null
572
+ even_batches: true
573
+ use_seedable_sampler: true
574
+ deepspeed:
575
+ desc: null
576
+ value: null
577
+ label_smoothing_factor:
578
+ desc: null
579
+ value: 0.0
580
+ optim:
581
+ desc: null
582
+ value: paged_adamw_8bit
583
+ optim_args:
584
+ desc: null
585
+ value: null
586
+ adafactor:
587
+ desc: null
588
+ value: false
589
+ group_by_length:
590
+ desc: null
591
+ value: false
592
+ length_column_name:
593
+ desc: null
594
+ value: length
595
+ report_to:
596
+ desc: null
597
+ value:
598
+ - tensorboard
599
+ - wandb
600
+ ddp_find_unused_parameters:
601
+ desc: null
602
+ value: null
603
+ ddp_bucket_cap_mb:
604
+ desc: null
605
+ value: null
606
+ ddp_broadcast_buffers:
607
+ desc: null
608
+ value: null
609
+ dataloader_pin_memory:
610
+ desc: null
611
+ value: true
612
+ dataloader_persistent_workers:
613
+ desc: null
614
+ value: false
615
+ skip_memory_metrics:
616
+ desc: null
617
+ value: true
618
+ use_legacy_prediction_loop:
619
+ desc: null
620
+ value: false
621
+ push_to_hub:
622
+ desc: null
623
+ value: false
624
+ resume_from_checkpoint:
625
+ desc: null
626
+ value: null
627
+ hub_model_id:
628
+ desc: null
629
+ value: null
630
+ hub_strategy:
631
+ desc: null
632
+ value: every_save
633
+ hub_token:
634
+ desc: null
635
+ value: <HUB_TOKEN>
636
+ hub_private_repo:
637
+ desc: null
638
+ value: false
639
+ hub_always_push:
640
+ desc: null
641
+ value: false
642
+ gradient_checkpointing:
643
+ desc: null
644
+ value: false
645
+ gradient_checkpointing_kwargs:
646
+ desc: null
647
+ value: null
648
+ include_inputs_for_metrics:
649
+ desc: null
650
+ value: false
651
+ fp16_backend:
652
+ desc: null
653
+ value: auto
654
+ push_to_hub_model_id:
655
+ desc: null
656
+ value: null
657
+ push_to_hub_organization:
658
+ desc: null
659
+ value: null
660
+ push_to_hub_token:
661
+ desc: null
662
+ value: <PUSH_TO_HUB_TOKEN>
663
+ mp_parameters:
664
+ desc: null
665
+ value: ''
666
+ auto_find_batch_size:
667
+ desc: null
668
+ value: false
669
+ full_determinism:
670
+ desc: null
671
+ value: false
672
+ torchdynamo:
673
+ desc: null
674
+ value: null
675
+ ray_scope:
676
+ desc: null
677
+ value: last
678
+ ddp_timeout:
679
+ desc: null
680
+ value: 1800
681
+ torch_compile:
682
+ desc: null
683
+ value: false
684
+ torch_compile_backend:
685
+ desc: null
686
+ value: null
687
+ torch_compile_mode:
688
+ desc: null
689
+ value: null
690
+ dispatch_batches:
691
+ desc: null
692
+ value: null
693
+ split_batches:
694
+ desc: null
695
+ value: null
696
+ include_tokens_per_second:
697
+ desc: null
698
+ value: false
699
+ include_num_input_tokens_seen:
700
+ desc: null
701
+ value: false
702
+ neftune_noise_alpha:
703
+ desc: null
704
+ value: null
wandb/run-20240401_093743-ie4jv7u5/files/output.log ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
2
+ warnings.warn(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
6
+ warnings.warn(
7
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
8
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
9
+ warnings.warn(
10
+ Checkpoint destination directory /kaggle/working/checkpoint-1 already exists and is non-empty. Saving will proceed but saved results may be invalid.
11
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
12
+ warnings.warn(
13
+ Checkpoint destination directory /kaggle/working/checkpoint-2 already exists and is non-empty. Saving will proceed but saved results may be invalid.
14
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
15
+ warnings.warn(
16
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
17
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
18
+ warnings.warn(
19
+ Checkpoint destination directory /kaggle/working/checkpoint-1 already exists and is non-empty. Saving will proceed but saved results may be invalid.
20
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
21
+ warnings.warn(
22
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
23
+ warnings.warn(
24
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
25
+ warnings.warn(
26
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
27
+ warnings.warn(
28
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
29
+ warnings.warn(
30
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
31
+ warnings.warn(
32
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
33
+ warnings.warn(
34
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
35
+ warnings.warn(
36
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
37
+ warnings.warn(
38
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
39
+ warnings.warn(
40
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
41
+ warnings.warn(
42
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
wandb/run-20240401_093743-ie4jv7u5/files/requirements.txt ADDED
@@ -0,0 +1,882 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ Levenshtein==0.25.0
16
+ LunarCalendar==0.0.9
17
+ Mako==1.3.2
18
+ Markdown==3.5.2
19
+ MarkupSafe==2.1.3
20
+ MarkupSafe==2.1.5
21
+ Pillow==9.5.0
22
+ PuLP==2.8.0
23
+ PyArabic==0.6.15
24
+ PyAstronomy==0.21.0
25
+ PyJWT==2.8.0
26
+ PyMeeus==0.5.12
27
+ PySocks==1.7.1
28
+ PyUpSet==0.1.1.post7
29
+ PyWavelets==1.5.0
30
+ PyYAML==6.0.1
31
+ Pygments==2.17.2
32
+ Pympler==1.0.1
33
+ QtPy==2.4.1
34
+ Rtree==1.2.0
35
+ SQLAlchemy==2.0.25
36
+ SecretStorage==3.3.3
37
+ Send2Trash==1.8.2
38
+ Shapely==1.8.5.post1
39
+ Shimmy==1.3.0
40
+ SimpleITK==2.3.1
41
+ TPOT==0.12.1
42
+ Theano-PyMC==1.1.2
43
+ Theano==1.0.5
44
+ Unidecode==1.3.8
45
+ Wand==0.6.13
46
+ Werkzeug==3.0.1
47
+ absl-py==1.4.0
48
+ accelerate==0.28.0
49
+ access==1.1.9
50
+ affine==2.4.0
51
+ aiobotocore==2.12.1
52
+ aiofiles==22.1.0
53
+ aiohttp-cors==0.7.0
54
+ aiohttp==3.9.1
55
+ aioitertools==0.11.0
56
+ aiorwlock==1.3.0
57
+ aiosignal==1.3.1
58
+ aiosqlite==0.19.0
59
+ albumentations==1.4.0
60
+ alembic==1.13.1
61
+ altair==5.2.0
62
+ annotated-types==0.6.0
63
+ annoy==1.17.3
64
+ anyio==4.2.0
65
+ apache-beam==2.46.0
66
+ aplus==0.11.0
67
+ appdirs==1.4.4
68
+ archspec==0.2.2
69
+ argon2-cffi-bindings==21.2.0
70
+ argon2-cffi==23.1.0
71
+ array-record==0.5.0
72
+ arrow==1.3.0
73
+ arviz==0.17.1
74
+ astroid==3.0.3
75
+ astropy-iers-data==0.2024.3.18.0.29.47
76
+ astropy==6.0.0
77
+ asttokens==2.4.1
78
+ astunparse==1.6.3
79
+ async-lru==2.0.4
80
+ async-timeout==4.0.3
81
+ attrs==23.2.0
82
+ audioread==3.0.1
83
+ auto_gptq==0.7.1
84
+ autopep8==2.0.4
85
+ backoff==2.2.1
86
+ bayesian-optimization==1.4.3
87
+ beatrix_jupyterlab==2023.128.151533
88
+ beautifulsoup4==4.12.2
89
+ bidict==0.23.1
90
+ bitsandbytes==0.43.0
91
+ blake3==0.2.1
92
+ bleach==6.1.0
93
+ blessed==1.20.0
94
+ blinker==1.7.0
95
+ blis==0.7.10
96
+ blosc2==2.5.1
97
+ bokeh==3.3.4
98
+ boltons==23.1.1
99
+ boto3==1.26.100
100
+ botocore==1.34.51
101
+ bq_helper==0.4.1
102
+ bqplot==0.12.43
103
+ branca==0.7.1
104
+ brewer2mpl==1.4.1
105
+ brotlipy==0.7.0
106
+ cached-property==1.5.2
107
+ cachetools==4.2.4
108
+ cachetools==5.3.2
109
+ catalogue==2.0.10
110
+ catalyst==22.4
111
+ catboost==1.2.3
112
+ category-encoders==2.6.3
113
+ certifi==2024.2.2
114
+ cesium==0.12.1
115
+ cffi==1.16.0
116
+ charset-normalizer==3.3.2
117
+ chex==0.1.85
118
+ cleverhans==4.0.0
119
+ click-plugins==1.1.1
120
+ click==8.1.7
121
+ cligj==0.7.2
122
+ cloud-tpu-client==0.10
123
+ cloud-tpu-profiler==2.4.0
124
+ cloudpathlib==0.16.0
125
+ cloudpickle==2.2.1
126
+ cloudpickle==3.0.0
127
+ cmdstanpy==1.2.1
128
+ cmudict==1.0.21
129
+ colorama==0.4.6
130
+ colorcet==3.1.0
131
+ coloredlogs==15.0.1
132
+ colorful==0.5.6
133
+ colorlog==6.8.2
134
+ colorlover==0.3.0
135
+ comm==0.2.1
136
+ conda-libmamba-solver==23.7.0
137
+ conda-package-handling==2.2.0
138
+ conda==23.7.4
139
+ conda_package_streaming==0.9.0
140
+ confection==0.1.4
141
+ contextily==1.5.2
142
+ contourpy==1.2.0
143
+ convertdate==2.4.0
144
+ cramjam==2.8.3
145
+ crcmod==1.7
146
+ cryptography==41.0.7
147
+ cuda-python==12.4.0
148
+ cudf==23.8.0
149
+ cufflinks==0.17.3
150
+ cuml==23.8.0
151
+ cupy==13.0.0
152
+ cycler==0.12.1
153
+ cymem==2.0.8
154
+ cytoolz==0.12.3
155
+ daal4py==2024.1.0
156
+ daal==2024.1.0
157
+ dacite==1.8.1
158
+ dask-cuda==23.8.0
159
+ dask-cudf==23.8.0
160
+ dask-expr==1.0.4
161
+ dask==2024.3.1
162
+ dataclasses-json==0.6.4
163
+ dataproc_jupyter_plugin==0.1.66
164
+ datasets==2.1.0
165
+ datashader==0.16.0
166
+ datatile==1.0.3
167
+ db-dtypes==1.2.0
168
+ deap==1.4.1
169
+ debugpy==1.8.0
170
+ decorator==5.1.1
171
+ deepdiff==6.7.1
172
+ defusedxml==0.7.1
173
+ deprecation==2.1.0
174
+ descartes==1.1.0
175
+ dill==0.3.8
176
+ dipy==1.9.0
177
+ distlib==0.3.8
178
+ distributed==2023.7.1
179
+ distro==1.9.0
180
+ dm-tree==0.1.8
181
+ docker-pycreds==0.4.0
182
+ docker==7.0.0
183
+ docopt==0.6.2
184
+ docstring-parser==0.15
185
+ docstring-to-markdown==0.15
186
+ docutils==0.20.1
187
+ earthengine-api==0.1.394
188
+ easydict==1.13
189
+ easyocr==1.7.1
190
+ ecos==2.0.13
191
+ eli5==0.13.0
192
+ emoji==2.10.1
193
+ en-core-web-lg==3.7.1
194
+ en-core-web-sm==3.7.1
195
+ entrypoints==0.4
196
+ ephem==4.1.5
197
+ esda==2.5.1
198
+ essentia==2.1b6.dev1110
199
+ et-xmlfile==1.1.0
200
+ etils==1.6.0
201
+ exceptiongroup==1.2.0
202
+ executing==2.0.1
203
+ explainable-ai-sdk==1.3.3
204
+ fastai==2.7.14
205
+ fastapi==0.108.0
206
+ fastavro==1.9.3
207
+ fastcore==1.5.29
208
+ fastdownload==0.0.7
209
+ fasteners==0.19
210
+ fastjsonschema==2.19.1
211
+ fastparquet==2024.2.0
212
+ fastprogress==1.0.3
213
+ fastrlock==0.8.2
214
+ fasttext==0.9.2
215
+ feather-format==0.4.1
216
+ featuretools==1.30.0
217
+ filelock==3.13.1
218
+ fiona==1.9.6
219
+ fitter==1.7.0
220
+ flake8==7.0.0
221
+ flashtext==2.7
222
+ flatbuffers==23.5.26
223
+ flax==0.8.2
224
+ folium==0.16.0
225
+ fonttools==4.47.0
226
+ fonttools==4.49.0
227
+ fqdn==1.5.1
228
+ frozendict==2.4.0
229
+ frozenlist==1.4.1
230
+ fsspec==2024.3.0
231
+ funcy==2.0
232
+ fury==0.10.0
233
+ future==1.0.0
234
+ fuzzywuzzy==0.18.0
235
+ gast==0.5.4
236
+ gatspy==0.3
237
+ gcsfs==2023.12.2.post1
238
+ gekko==1.1.0
239
+ gensim==4.3.2
240
+ geographiclib==2.0
241
+ geojson==3.1.0
242
+ geopandas==0.14.3
243
+ geoplot==0.5.1
244
+ geopy==2.4.1
245
+ geoviews==1.11.1
246
+ ggplot==0.11.5
247
+ giddy==2.3.5
248
+ gitdb==4.0.11
249
+ google-ai-generativelanguage==0.4.0
250
+ google-api-core==2.11.1
251
+ google-api-core==2.17.1
252
+ google-api-python-client==2.122.0
253
+ google-apitools==0.5.31
254
+ google-auth-httplib2==0.1.1
255
+ google-auth-oauthlib==1.2.0
256
+ google-auth==2.26.1
257
+ google-cloud-aiplatform==0.6.0a1
258
+ google-cloud-artifact-registry==1.10.0
259
+ google-cloud-automl==1.0.1
260
+ google-cloud-bigquery==2.34.4
261
+ google-cloud-bigtable==1.7.3
262
+ google-cloud-core==2.4.1
263
+ google-cloud-datastore==2.19.0
264
+ google-cloud-dlp==3.14.0
265
+ google-cloud-jupyter-config==0.0.5
266
+ google-cloud-language==2.13.3
267
+ google-cloud-monitoring==2.18.0
268
+ google-cloud-pubsub==2.19.0
269
+ google-cloud-pubsublite==1.9.0
270
+ google-cloud-recommendations-ai==0.7.1
271
+ google-cloud-resource-manager==1.11.0
272
+ google-cloud-spanner==3.40.1
273
+ google-cloud-storage==1.44.0
274
+ google-cloud-translate==3.12.1
275
+ google-cloud-videointelligence==2.13.3
276
+ google-cloud-vision==2.8.0
277
+ google-crc32c==1.5.0
278
+ google-generativeai==0.4.1
279
+ google-pasta==0.2.0
280
+ google-resumable-media==2.7.0
281
+ googleapis-common-protos==1.62.0
282
+ gplearn==0.4.2
283
+ gpustat==1.0.0
284
+ gpxpy==1.6.2
285
+ graphviz==0.20.2
286
+ greenlet==3.0.3
287
+ grpc-google-iam-v1==0.12.7
288
+ grpcio-status==1.48.1
289
+ grpcio-status==1.48.2
290
+ grpcio==1.51.1
291
+ grpcio==1.60.0
292
+ gviz-api==1.10.0
293
+ gym-notices==0.0.8
294
+ gym==0.26.2
295
+ gymnasium==0.29.0
296
+ h11==0.14.0
297
+ h2o==3.46.0.1
298
+ h5netcdf==1.3.0
299
+ h5py==3.10.0
300
+ haversine==2.8.1
301
+ hdfs==2.7.3
302
+ hep-ml==0.7.2
303
+ hijri-converter==2.3.1
304
+ hmmlearn==0.3.2
305
+ holidays==0.24
306
+ holoviews==1.18.3
307
+ hpsklearn==0.1.0
308
+ html5lib==1.1
309
+ htmlmin==0.1.12
310
+ httpcore==1.0.4
311
+ httplib2==0.21.0
312
+ httptools==0.6.1
313
+ httpx==0.27.0
314
+ huggingface-hub==0.21.4
315
+ humanfriendly==10.0
316
+ hunspell==0.5.5
317
+ husl==4.0.3
318
+ hydra-slayer==0.5.0
319
+ hyperopt==0.2.7
320
+ hypertools==0.8.0
321
+ idna==3.6
322
+ igraph==0.11.4
323
+ imagecodecs==2024.1.1
324
+ imageio==2.33.1
325
+ imbalanced-learn==0.12.0
326
+ imgaug==0.4.0
327
+ importlib-metadata==6.11.0
328
+ importlib-metadata==7.0.1
329
+ importlib-resources==6.1.1
330
+ inequality==1.0.1
331
+ iniconfig==2.0.0
332
+ ipydatawidgets==4.3.5
333
+ ipykernel==6.28.0
334
+ ipyleaflet==0.18.2
335
+ ipympl==0.7.0
336
+ ipython-genutils==0.2.0
337
+ ipython-genutils==0.2.0
338
+ ipython-sql==0.5.0
339
+ ipython==8.20.0
340
+ ipyvolume==0.6.3
341
+ ipyvue==1.10.2
342
+ ipyvuetify==1.9.2
343
+ ipywebrtc==0.6.0
344
+ ipywidgets==7.7.1
345
+ isoduration==20.11.0
346
+ isort==5.13.2
347
+ isoweek==1.3.3
348
+ itsdangerous==2.1.2
349
+ jaraco.classes==3.3.0
350
+ jax-jumpy==1.0.0
351
+ jax==0.4.23
352
+ jaxlib==0.4.23.dev20240116
353
+ jedi==0.19.1
354
+ jeepney==0.8.0
355
+ jieba==0.42.1
356
+ jmespath==1.0.1
357
+ joblib==1.3.2
358
+ json5==0.9.14
359
+ jsonpatch==1.33
360
+ jsonpointer==2.4
361
+ jsonschema-specifications==2023.12.1
362
+ jsonschema==4.20.0
363
+ jupyter-console==6.6.3
364
+ jupyter-events==0.9.0
365
+ jupyter-http-over-ws==0.0.8
366
+ jupyter-lsp==1.5.1
367
+ jupyter-server-mathjax==0.2.6
368
+ jupyter-ydoc==0.2.5
369
+ jupyter_client==7.4.9
370
+ jupyter_client==8.6.0
371
+ jupyter_core==5.7.1
372
+ jupyter_server==2.13.0
373
+ jupyter_server_fileid==0.9.1
374
+ jupyter_server_proxy==4.1.0
375
+ jupyter_server_terminals==0.5.1
376
+ jupyter_server_ydoc==0.8.0
377
+ jupyterlab-lsp==5.1.0
378
+ jupyterlab-widgets==3.0.9
379
+ jupyterlab==4.1.5
380
+ jupyterlab_git==0.44.0
381
+ jupyterlab_pygments==0.3.0
382
+ jupyterlab_server==2.25.2
383
+ jupytext==1.16.0
384
+ kaggle-environments==1.14.3
385
+ kaggle==1.6.6
386
+ kagglehub==0.2.0
387
+ keras-cv==0.8.2
388
+ keras-nlp==0.8.2
389
+ keras-tuner==1.4.6
390
+ keras==3.0.5
391
+ kernels-mixer==0.0.7
392
+ keyring==24.3.0
393
+ keyrings.google-artifactregistry-auth==1.1.2
394
+ kfp-pipeline-spec==0.2.2
395
+ kfp-server-api==2.0.5
396
+ kfp==2.5.0
397
+ kiwisolver==1.4.5
398
+ kmapper==2.0.1
399
+ kmodes==0.12.2
400
+ korean-lunar-calendar==0.3.1
401
+ kornia==0.7.2
402
+ kornia_rs==0.1.2
403
+ kt-legacy==1.0.5
404
+ kubernetes==26.1.0
405
+ langcodes==3.3.0
406
+ langid==1.1.6
407
+ lazy_loader==0.3
408
+ learntools==0.3.4
409
+ leven==1.0.4
410
+ libclang==16.0.6
411
+ libmambapy==1.5.0
412
+ libpysal==4.9.2
413
+ librosa==0.10.1
414
+ lightgbm==4.2.0
415
+ lightning-utilities==0.10.1
416
+ lime==0.2.0.1
417
+ line-profiler==4.1.2
418
+ linkify-it-py==2.0.3
419
+ llvmlite==0.41.1
420
+ llvmlite==0.42.0
421
+ lml==0.1.0
422
+ locket==1.0.0
423
+ loguru==0.7.2
424
+ lxml==5.1.0
425
+ lz4==4.3.3
426
+ mamba==1.5.0
427
+ mapclassify==2.6.1
428
+ markdown-it-py==3.0.0
429
+ markovify==0.9.4
430
+ marshmallow==3.21.1
431
+ matplotlib-inline==0.1.6
432
+ matplotlib-venn==0.11.10
433
+ matplotlib==3.7.5
434
+ matplotlib==3.8.3
435
+ mccabe==0.7.0
436
+ mdit-py-plugins==0.4.0
437
+ mdurl==0.1.2
438
+ memory-profiler==0.61.0
439
+ menuinst==2.0.1
440
+ mercantile==1.2.1
441
+ mgwr==2.2.1
442
+ missingno==0.5.2
443
+ mistune==0.8.4
444
+ mizani==0.11.0
445
+ ml-dtypes==0.2.0
446
+ mlcrate==0.2.0
447
+ mlens==0.2.3
448
+ mlxtend==0.23.1
449
+ mmh3==4.1.0
450
+ mne==1.6.1
451
+ mnist==0.2.2
452
+ mock==5.1.0
453
+ momepy==0.7.0
454
+ more-itertools==10.2.0
455
+ mpld3==0.5.10
456
+ mpmath==1.3.0
457
+ msgpack-numpy==0.4.8
458
+ msgpack==1.0.7
459
+ multidict==6.0.4
460
+ multimethod==1.10
461
+ multipledispatch==1.0.0
462
+ multiprocess==0.70.16
463
+ munkres==1.1.4
464
+ murmurhash==1.0.10
465
+ mypy-extensions==1.0.0
466
+ namex==0.0.7
467
+ nb-conda-kernels==2.3.1
468
+ nb_conda==2.2.1
469
+ nbclassic==1.0.0
470
+ nbclient==0.5.13
471
+ nbconvert==6.4.5
472
+ nbdime==3.2.0
473
+ nbformat==5.9.2
474
+ ndindex==1.8
475
+ nest-asyncio==1.5.8
476
+ networkx==3.2.1
477
+ nibabel==5.2.1
478
+ nilearn==0.10.3
479
+ ninja==1.11.1.1
480
+ nltk==3.2.4
481
+ nose==1.3.7
482
+ notebook==6.5.4
483
+ notebook==6.5.6
484
+ notebook_executor==0.2
485
+ notebook_shim==0.2.3
486
+ numba==0.58.1
487
+ numba==0.59.0
488
+ numexpr==2.9.0
489
+ numpy==1.26.4
490
+ nvidia-ml-py==11.495.46
491
+ nvtx==0.2.10
492
+ oauth2client==4.1.3
493
+ oauthlib==3.2.2
494
+ objsize==0.6.1
495
+ odfpy==1.4.1
496
+ olefile==0.47
497
+ onnx==1.15.0
498
+ opencensus-context==0.1.3
499
+ opencensus==0.11.4
500
+ opencv-contrib-python==4.9.0.80
501
+ opencv-python-headless==4.9.0.80
502
+ opencv-python==4.9.0.80
503
+ openpyxl==3.1.2
504
+ openslide-python==1.3.1
505
+ opentelemetry-api==1.22.0
506
+ opentelemetry-exporter-otlp-proto-common==1.22.0
507
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
508
+ opentelemetry-exporter-otlp-proto-http==1.22.0
509
+ opentelemetry-exporter-otlp==1.22.0
510
+ opentelemetry-proto==1.22.0
511
+ opentelemetry-sdk==1.22.0
512
+ opentelemetry-semantic-conventions==0.43b0
513
+ opt-einsum==3.3.0
514
+ optax==0.2.1
515
+ optimum==1.18.0
516
+ optuna==3.6.0
517
+ orbax-checkpoint==0.5.6
518
+ ordered-set==4.1.0
519
+ orjson==3.9.10
520
+ ortools==9.4.1874
521
+ osmnx==1.9.1
522
+ overrides==7.4.0
523
+ packaging==21.3
524
+ pandas-datareader==0.10.0
525
+ pandas-profiling==3.6.6
526
+ pandas-summary==0.2.0
527
+ pandas==2.1.4
528
+ pandas==2.2.1
529
+ pandasql==0.7.3
530
+ pandocfilters==1.5.0
531
+ panel==1.3.8
532
+ papermill==2.5.0
533
+ param==2.0.2
534
+ parso==0.8.3
535
+ partd==1.4.1
536
+ path.py==12.5.0
537
+ path==16.10.0
538
+ pathos==0.3.2
539
+ pathy==0.10.3
540
+ patsy==0.5.6
541
+ pdf2image==1.17.0
542
+ peft==0.10.0
543
+ pettingzoo==1.24.0
544
+ pexpect==4.8.0
545
+ pexpect==4.9.0
546
+ phik==0.12.4
547
+ pickleshare==0.7.5
548
+ pip==23.3.2
549
+ pkgutil_resolve_name==1.3.10
550
+ platformdirs==4.2.0
551
+ plotly-express==0.4.1
552
+ plotly==5.18.0
553
+ plotnine==0.13.2
554
+ pluggy==1.4.0
555
+ pointpats==2.4.0
556
+ polars==0.20.15
557
+ polyglot==16.7.4
558
+ pooch==1.8.1
559
+ pox==0.3.4
560
+ ppca==0.0.4
561
+ ppft==1.7.6.8
562
+ preprocessing==0.1.13
563
+ preshed==3.0.9
564
+ prettytable==3.9.0
565
+ progressbar2==4.4.2
566
+ prometheus-client==0.19.0
567
+ promise==2.3
568
+ prompt-toolkit==3.0.42
569
+ prompt-toolkit==3.0.43
570
+ pronouncing==0.2.0
571
+ prophet==1.1.1
572
+ proto-plus==1.23.0
573
+ protobuf==3.20.3
574
+ protobuf==4.21.12
575
+ psutil==5.9.3
576
+ psutil==5.9.7
577
+ ptyprocess==0.7.0
578
+ pudb==2024.1
579
+ pure-eval==0.2.2
580
+ py-cpuinfo==9.0.0
581
+ py-spy==0.3.14
582
+ py4j==0.10.9.7
583
+ pyLDAvis==3.4.1
584
+ pyOpenSSL==23.3.0
585
+ pyaml==23.12.0
586
+ pyarrow==11.0.0
587
+ pyasn1-modules==0.3.0
588
+ pyasn1==0.5.1
589
+ pybind11==2.11.1
590
+ pyclipper==1.3.0.post5
591
+ pycodestyle==2.11.1
592
+ pycosat==0.6.6
593
+ pycparser==2.21
594
+ pycryptodome==3.20.0
595
+ pyct==0.5.0
596
+ pycuda==2024.1
597
+ pydantic==2.5.3
598
+ pydantic==2.6.4
599
+ pydantic_core==2.14.6
600
+ pydantic_core==2.16.3
601
+ pydegensac==0.1.2
602
+ pydicom==2.4.4
603
+ pydocstyle==6.3.0
604
+ pydot==1.4.2
605
+ pydub==0.25.1
606
+ pyemd==1.0.0
607
+ pyerfa==2.0.1.1
608
+ pyexcel-io==0.6.6
609
+ pyexcel-ods==0.6.0
610
+ pyflakes==3.2.0
611
+ pygltflib==1.16.2
612
+ pykalman==0.9.5
613
+ pylibraft==23.8.0
614
+ pylint==3.0.4
615
+ pymc3==3.11.4
616
+ pymongo==3.13.0
617
+ pynndescent==0.5.11
618
+ pynvml==11.4.1
619
+ pynvrtc==9.2
620
+ pyocr==0.8.5
621
+ pyparsing==3.1.1
622
+ pyparsing==3.1.2
623
+ pypdf==4.1.0
624
+ pyproj==3.6.1
625
+ pysal==24.1
626
+ pyshp==2.3.1
627
+ pytesseract==0.3.10
628
+ pytest==8.1.1
629
+ python-Levenshtein==0.25.0
630
+ python-bidi==0.4.2
631
+ python-dateutil==2.9.0.post0
632
+ python-dotenv==1.0.0
633
+ python-json-logger==2.0.7
634
+ python-louvain==0.16
635
+ python-lsp-jsonrpc==1.1.2
636
+ python-lsp-server==1.10.1
637
+ python-slugify==8.0.4
638
+ python-utils==3.8.2
639
+ pythreejs==2.4.2
640
+ pytoolconfig==1.3.1
641
+ pytools==2023.1.1
642
+ pytorch-ignite==0.4.13
643
+ pytorch-lightning==2.2.1
644
+ pytz==2023.3.post1
645
+ pytz==2024.1
646
+ pyu2f==0.1.5
647
+ pyviz_comms==3.0.1
648
+ pyzmq==24.0.1
649
+ pyzmq==25.1.2
650
+ qgrid==1.3.1
651
+ qtconsole==5.5.1
652
+ quantecon==0.7.2
653
+ quantities==0.15.0
654
+ qudida==0.0.4
655
+ raft-dask==23.8.0
656
+ rapidfuzz==3.6.2
657
+ rasterio==1.3.9
658
+ rasterstats==0.19.0
659
+ ray-cpp==2.9.0
660
+ ray==2.9.0
661
+ referencing==0.32.1
662
+ regex==2023.12.25
663
+ requests-oauthlib==1.3.1
664
+ requests-toolbelt==0.10.1
665
+ requests==2.31.0
666
+ responses==0.18.0
667
+ retrying==1.3.3
668
+ retrying==1.3.4
669
+ rfc3339-validator==0.1.4
670
+ rfc3986-validator==0.1.1
671
+ rgf-python==3.12.0
672
+ rich-click==1.7.4
673
+ rich==13.7.0
674
+ rich==13.7.1
675
+ rmm==23.8.0
676
+ rope==1.12.0
677
+ rouge==1.0.1
678
+ rpds-py==0.16.2
679
+ rsa==4.9
680
+ ruamel-yaml-conda==0.15.100
681
+ ruamel.yaml.clib==0.2.7
682
+ ruamel.yaml==0.17.40
683
+ s2sphere==0.2.5
684
+ s3fs==2024.3.0
685
+ s3transfer==0.6.2
686
+ safetensors==0.4.2
687
+ scattertext==0.1.19
688
+ scikit-image==0.22.0
689
+ scikit-learn-intelex==2024.1.0
690
+ scikit-learn==1.2.2
691
+ scikit-multilearn==0.2.0
692
+ scikit-optimize==0.10.1
693
+ scikit-plot==0.3.7
694
+ scikit-surprise==1.1.3
695
+ scipy==1.11.4
696
+ scipy==1.12.0
697
+ seaborn==0.12.2
698
+ segment_anything==1.0
699
+ segregation==2.5
700
+ semver==3.0.2
701
+ sentencepiece==0.2.0
702
+ sentry-sdk==1.42.0
703
+ setproctitle==1.3.3
704
+ setuptools-git==1.2
705
+ setuptools-scm==8.0.4
706
+ setuptools==69.0.3
707
+ shap==0.44.1
708
+ shapely==2.0.3
709
+ shellingham==1.5.4
710
+ simpervisor==1.0.0
711
+ simplejson==3.19.2
712
+ six==1.16.0
713
+ sklearn-pandas==2.2.0
714
+ slicer==0.0.7
715
+ smart-open==6.4.0
716
+ smmap==5.0.1
717
+ sniffio==1.3.0
718
+ snowballstemmer==2.2.0
719
+ snuggs==1.4.7
720
+ sortedcontainers==2.4.0
721
+ soundfile==0.12.1
722
+ soupsieve==2.5
723
+ soxr==0.3.7
724
+ spacy-legacy==3.0.12
725
+ spacy-loggers==1.0.5
726
+ spacy==3.7.2
727
+ spaghetti==1.7.5.post1
728
+ spectral==0.23.1
729
+ spglm==1.1.0
730
+ sphinx-rtd-theme==0.2.4
731
+ spint==1.0.7
732
+ splot==1.1.5.post1
733
+ spopt==0.6.0
734
+ spreg==1.4.2
735
+ spvcm==0.3.0
736
+ sqlparse==0.4.4
737
+ squarify==0.4.3
738
+ srsly==2.4.8
739
+ stable-baselines3==2.1.0
740
+ stack-data==0.6.2
741
+ stack-data==0.6.3
742
+ stanio==0.3.0
743
+ starlette==0.32.0.post1
744
+ statsmodels==0.14.1
745
+ stemming==1.0.1
746
+ stop-words==2018.7.23
747
+ stopit==1.1.2
748
+ stumpy==1.12.0
749
+ sympy==1.12
750
+ tables==3.9.2
751
+ tabulate==0.9.0
752
+ tangled-up-in-unicode==0.2.0
753
+ tbb==2021.11.0
754
+ tblib==3.0.0
755
+ tenacity==8.2.3
756
+ tensorboard-data-server==0.7.2
757
+ tensorboard-plugin-profile==2.15.0
758
+ tensorboard==2.15.1
759
+ tensorboardX==2.6.2.2
760
+ tensorflow-cloud==0.1.16
761
+ tensorflow-datasets==4.9.4
762
+ tensorflow-decision-forests==1.8.1
763
+ tensorflow-estimator==2.15.0
764
+ tensorflow-hub==0.16.1
765
+ tensorflow-io-gcs-filesystem==0.35.0
766
+ tensorflow-io==0.35.0
767
+ tensorflow-metadata==0.14.0
768
+ tensorflow-probability==0.23.0
769
+ tensorflow-serving-api==2.14.1
770
+ tensorflow-text==2.15.0
771
+ tensorflow-transform==0.14.0
772
+ tensorflow==2.15.0
773
+ tensorpack==0.11
774
+ tensorstore==0.1.56
775
+ termcolor==2.4.0
776
+ terminado==0.18.0
777
+ testpath==0.6.0
778
+ text-unidecode==1.3
779
+ textblob==0.18.0.post0
780
+ texttable==1.7.0
781
+ tf_keras==2.15.1
782
+ tfp-nightly==0.24.0.dev0
783
+ thinc==8.2.2
784
+ threadpoolctl==3.2.0
785
+ tifffile==2023.12.9
786
+ timm==0.9.16
787
+ tinycss2==1.2.1
788
+ tobler==0.11.2
789
+ tokenizers==0.15.2
790
+ toml==0.10.2
791
+ tomli==2.0.1
792
+ tomlkit==0.12.4
793
+ toolz==0.12.1
794
+ torch==2.1.2
795
+ torchaudio==2.1.2
796
+ torchdata==0.7.1
797
+ torchinfo==1.8.0
798
+ torchmetrics==1.3.2
799
+ torchtext==0.16.2
800
+ torchvision==0.16.2
801
+ tornado==6.3.3
802
+ tqdm==4.66.1
803
+ traceml==1.0.8
804
+ traitlets==5.9.0
805
+ traittypes==0.2.1
806
+ transformers==4.38.2
807
+ treelite-runtime==3.2.0
808
+ treelite==3.2.0
809
+ trueskill==0.4.5
810
+ truststore==0.8.0
811
+ trx-python==0.2.9
812
+ tsfresh==0.20.2
813
+ typeguard==4.1.5
814
+ typer==0.9.0
815
+ types-python-dateutil==2.8.19.20240106
816
+ typing-inspect==0.9.0
817
+ typing-utils==0.1.0
818
+ typing_extensions==4.9.0
819
+ tzdata==2023.4
820
+ uc-micro-py==1.0.3
821
+ ucx-py==0.33.0
822
+ ujson==5.9.0
823
+ umap-learn==0.5.5
824
+ unicodedata2==15.1.0
825
+ update-checker==0.18.0
826
+ uri-template==1.3.0
827
+ uritemplate==3.0.1
828
+ urllib3==1.26.18
829
+ urllib3==2.1.0
830
+ urwid==2.6.9
831
+ urwid_readline==0.14
832
+ uvicorn==0.25.0
833
+ uvloop==0.19.0
834
+ vaex-astro==0.9.3
835
+ vaex-core==4.17.1
836
+ vaex-hdf5==0.14.1
837
+ vaex-jupyter==0.8.2
838
+ vaex-ml==0.18.3
839
+ vaex-server==0.9.0
840
+ vaex-viz==0.5.4
841
+ vaex==4.17.0
842
+ vec_noise==1.1.4
843
+ vecstack==0.4.0
844
+ virtualenv==20.21.0
845
+ visions==0.7.5
846
+ vowpalwabbit==9.9.0
847
+ vtk==9.3.0
848
+ wandb==0.16.4
849
+ wasabi==1.1.2
850
+ watchfiles==0.21.0
851
+ wavio==0.0.8
852
+ wcwidth==0.2.13
853
+ weasel==0.3.4
854
+ webcolors==1.13
855
+ webencodings==0.5.1
856
+ websocket-client==1.7.0
857
+ websockets==12.0
858
+ wfdb==4.1.2
859
+ whatthepatch==1.0.5
860
+ wheel==0.42.0
861
+ widgetsnbextension==3.6.6
862
+ witwidget==1.8.1
863
+ woodwork==0.29.0
864
+ wordcloud==1.9.3
865
+ wordsegment==1.3.1
866
+ wrapt==1.14.1
867
+ xarray-einstats==0.7.0
868
+ xarray==2024.2.0
869
+ xgboost==2.0.3
870
+ xvfbwrapper==0.2.9
871
+ xxhash==3.4.1
872
+ xyzservices==2023.10.1
873
+ y-py==0.6.2
874
+ yapf==0.40.2
875
+ yarl==1.9.3
876
+ yarl==1.9.4
877
+ ydata-profiling==4.6.4
878
+ yellowbrick==1.5
879
+ ypy-websocket==0.8.4
880
+ zict==3.0.0
881
+ zipp==3.17.0
882
+ zstandard==0.22.0
wandb/run-20240401_093743-ie4jv7u5/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-01T09:37:43.894456",
5
+ "startedAt": "2024-04-01T09:37:43.076491",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "148e249fd8ed",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5524.099212646484
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.788, "train/grad_norm": 0.9128203988075256, "train/learning_rate": 0.00013877551020408165, "train/epoch": 12.0, "train/global_step": 16, "_timestamp": 1711964681.5606613, "_runtime": 418.4748373031616, "_step": 31, "eval/loss": 1.0970141887664795, "eval/runtime": 1.8296, "eval/samples_per_second": 1.093, "eval/steps_per_second": 0.547}
wandb/run-20240401_093743-ie4jv7u5/logs/debug-internal.log ADDED
@@ -0,0 +1,687 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-01 09:37:43,083 INFO StreamThr :119 [internal.py:wandb_internal():86] W&B internal server running at pid: 119, started at: 2024-04-01 09:37:43.082780
2
+ 2024-04-01 09:37:43,084 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-01 09:37:43,596 INFO WriterThread:119 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/run-ie4jv7u5.wandb
4
+ 2024-04-01 09:37:43,597 DEBUG SenderThread:119 [sender.py:send():378] send: header
5
+ 2024-04-01 09:37:43,600 DEBUG SenderThread:119 [sender.py:send():378] send: run
6
+ 2024-04-01 09:37:43,758 INFO SenderThread:119 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files
7
+ 2024-04-01 09:37:43,758 INFO SenderThread:119 [sender.py:_start_run_threads():1099] run started: ie4jv7u5 with start time 1711964263.085824
8
+ 2024-04-01 09:37:43,766 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-01 09:37:43,766 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: check_version
10
+ 2024-04-01 09:37:43,860 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-01 09:37:43,872 DEBUG HandlerThread:119 [system_info.py:__init__():26] System info init
12
+ 2024-04-01 09:37:43,872 DEBUG HandlerThread:119 [system_info.py:__init__():41] System info init done
13
+ 2024-04-01 09:37:43,872 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-01 09:37:43,872 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-01 09:37:43,872 INFO HandlerThread:119 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-01 09:37:43,873 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-01 09:37:43,873 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-01 09:37:43,875 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-01 09:37:43,876 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-01 09:37:43,877 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-01 09:37:43,894 DEBUG HandlerThread:119 [system_info.py:probe():150] Probing system
22
+ 2024-04-01 09:37:43,896 DEBUG HandlerThread:119 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-01 09:37:43,896 DEBUG HandlerThread:119 [system_info.py:probe():198] Probing system done
24
+ 2024-04-01 09:37:43,896 DEBUG HandlerThread:119 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-01T09:37:43.894456', 'startedAt': '2024-04-01T09:37:43.076491', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '148e249fd8ed', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5524.099212646484}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-01 09:37:43,896 INFO HandlerThread:119 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-01 09:37:43,896 INFO HandlerThread:119 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-01 09:37:43,896 DEBUG HandlerThread:119 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-01 09:37:44,760 INFO Thread-12 :119 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/conda-environment.yaml
29
+ 2024-04-01 09:38:42,776 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/conda-environment.yaml
30
+ 2024-04-01 09:38:42,830 DEBUG HandlerThread:119 [system_info.py:_save_conda():219] Saving conda packages done
31
+ 2024-04-01 09:38:42,831 INFO HandlerThread:119 [system_monitor.py:probe():229] Finished publishing system info
32
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
33
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
34
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
35
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
36
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
37
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
38
+ 2024-04-01 09:38:42,845 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
39
+ 2024-04-01 09:38:42,846 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
40
+ 2024-04-01 09:38:42,846 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
41
+ 2024-04-01 09:38:42,846 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
42
+ 2024-04-01 09:38:42,846 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-01 09:38:42,847 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: python_packages
44
+ 2024-04-01 09:38:42,847 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: stop_status
45
+ 2024-04-01 09:38:42,847 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: internal_messages
46
+ 2024-04-01 09:38:42,848 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: python_packages
47
+ 2024-04-01 09:38:42,849 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
48
+ 2024-04-01 09:38:42,852 DEBUG SenderThread:119 [sender.py:send():378] send: telemetry
49
+ 2024-04-01 09:38:42,852 DEBUG SenderThread:119 [sender.py:send():378] send: config
50
+ 2024-04-01 09:38:42,854 DEBUG SenderThread:119 [sender.py:send():378] send: metric
51
+ 2024-04-01 09:38:42,854 DEBUG SenderThread:119 [sender.py:send():378] send: telemetry
52
+ 2024-04-01 09:38:42,854 DEBUG SenderThread:119 [sender.py:send():378] send: metric
53
+ 2024-04-01 09:38:42,854 WARNING SenderThread:119 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
54
+ 2024-04-01 09:38:42,854 DEBUG SenderThread:119 [sender.py:send():378] send: telemetry
55
+ 2024-04-01 09:38:42,854 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: stop_status
56
+ 2024-04-01 09:38:42,852 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
57
+ 2024-04-01 09:38:42,857 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
58
+ 2024-04-01 09:38:42,857 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
59
+ 2024-04-01 09:38:42,858 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
60
+ 2024-04-01 09:38:42,859 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
61
+ 2024-04-01 09:38:42,859 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
62
+ 2024-04-01 09:38:42,860 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
63
+ 2024-04-01 09:38:42,860 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
64
+ 2024-04-01 09:38:42,861 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
65
+ 2024-04-01 09:38:42,861 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
66
+ 2024-04-01 09:38:42,861 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
67
+ 2024-04-01 09:38:42,861 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
68
+ 2024-04-01 09:38:42,861 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
69
+ 2024-04-01 09:38:42,891 DEBUG SenderThread:119 [sender.py:send():378] send: metric
70
+ 2024-04-01 09:38:42,891 DEBUG SenderThread:119 [sender.py:send():378] send: metric
71
+ 2024-04-01 09:38:42,891 DEBUG SenderThread:119 [sender.py:send():378] send: metric
72
+ 2024-04-01 09:38:42,891 DEBUG SenderThread:119 [sender.py:send():378] send: metric
73
+ 2024-04-01 09:38:42,892 DEBUG SenderThread:119 [sender.py:send():378] send: history
74
+ 2024-04-01 09:38:42,892 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
75
+ 2024-04-01 09:38:42,894 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
76
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send():378] send: metric
77
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send():378] send: metric
78
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send():378] send: metric
79
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send():378] send: metric
80
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send():378] send: history
81
+ 2024-04-01 09:38:42,895 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
82
+ 2024-04-01 09:38:42,896 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
83
+ 2024-04-01 09:38:42,897 DEBUG SenderThread:119 [sender.py:send():378] send: history
84
+ 2024-04-01 09:38:42,897 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
85
+ 2024-04-01 09:38:42,897 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
86
+ 2024-04-01 09:38:42,898 DEBUG SenderThread:119 [sender.py:send():378] send: history
87
+ 2024-04-01 09:38:42,898 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
88
+ 2024-04-01 09:38:42,898 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
89
+ 2024-04-01 09:38:42,899 DEBUG SenderThread:119 [sender.py:send():378] send: files
90
+ 2024-04-01 09:38:42,899 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-metadata.json with policy now
91
+ 2024-04-01 09:38:43,086 INFO wandb-upload_0:119 [upload_job.py:push():131] Uploaded file /tmp/tmpmz_jko_lwandb/4t9dn2ed-wandb-metadata.json
92
+ 2024-04-01 09:38:43,776 INFO Thread-12 :119 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
93
+ 2024-04-01 09:38:43,777 INFO Thread-12 :119 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-metadata.json
94
+ 2024-04-01 09:38:43,777 INFO Thread-12 :119 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/requirements.txt
95
+ 2024-04-01 09:38:43,777 INFO Thread-12 :119 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
96
+ 2024-04-01 09:38:43,877 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
97
+ 2024-04-01 09:38:43,878 DEBUG SenderThread:119 [sender.py:send():378] send: stats
98
+ 2024-04-01 09:38:43,884 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-01 09:38:44,777 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/config.yaml
100
+ 2024-04-01 09:38:44,908 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
101
+ 2024-04-01 09:38:45,777 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
102
+ 2024-04-01 09:38:49,013 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
103
+ 2024-04-01 09:38:49,909 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
104
+ 2024-04-01 09:38:54,014 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-04-01 09:38:54,920 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
106
+ 2024-04-01 09:38:55,054 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
107
+ 2024-04-01 09:38:55,055 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
108
+ 2024-04-01 09:38:55,055 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
109
+ 2024-04-01 09:38:55,055 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
110
+ 2024-04-01 09:38:55,055 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
111
+ 2024-04-01 09:38:55,056 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
112
+ 2024-04-01 09:38:55,056 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
113
+ 2024-04-01 09:38:55,067 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
114
+ 2024-04-01 09:38:55,067 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
115
+ 2024-04-01 09:38:55,067 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
116
+ 2024-04-01 09:38:55,067 DEBUG SenderThread:119 [sender.py:send():378] send: stats
117
+ 2024-04-01 09:38:59,068 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
118
+ 2024-04-01 09:38:59,927 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
119
+ 2024-04-01 09:39:04,069 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-04-01 09:39:04,927 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
121
+ 2024-04-01 09:39:09,070 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-01 09:39:09,928 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
123
+ 2024-04-01 09:39:14,072 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
124
+ 2024-04-01 09:39:14,929 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
125
+ 2024-04-01 09:39:19,073 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-04-01 09:39:19,930 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
127
+ 2024-04-01 09:39:21,056 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
128
+ 2024-04-01 09:39:21,056 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
129
+ 2024-04-01 09:39:21,056 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
130
+ 2024-04-01 09:39:21,056 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
131
+ 2024-04-01 09:39:21,057 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
132
+ 2024-04-01 09:39:21,057 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
133
+ 2024-04-01 09:39:21,058 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
134
+ 2024-04-01 09:39:21,059 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
135
+ 2024-04-01 09:39:21,061 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
136
+ 2024-04-01 09:39:21,062 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
137
+ 2024-04-01 09:39:21,062 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
138
+ 2024-04-01 09:39:21,062 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
139
+ 2024-04-01 09:39:21,062 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
140
+ 2024-04-01 09:39:21,063 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
141
+ 2024-04-01 09:39:21,063 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
142
+ 2024-04-01 09:39:21,063 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
143
+ 2024-04-01 09:39:21,065 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
144
+ 2024-04-01 09:39:21,082 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
145
+ 2024-04-01 09:39:21,082 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
146
+ 2024-04-01 09:39:21,082 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
147
+ 2024-04-01 09:39:21,083 DEBUG SenderThread:119 [sender.py:send():378] send: stats
148
+ 2024-04-01 09:39:21,458 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
149
+ 2024-04-01 09:39:21,458 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
150
+ 2024-04-01 09:39:21,458 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
151
+ 2024-04-01 09:39:21,458 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
152
+ 2024-04-01 09:39:21,459 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
153
+ 2024-04-01 09:39:21,460 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
154
+ 2024-04-01 09:39:21,462 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
155
+ 2024-04-01 09:39:21,463 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
156
+ 2024-04-01 09:39:21,463 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
157
+ 2024-04-01 09:39:22,013 DEBUG SenderThread:119 [sender.py:send():378] send: config
158
+ 2024-04-01 09:39:22,014 DEBUG SenderThread:119 [sender.py:send():378] send: metric
159
+ 2024-04-01 09:39:22,014 DEBUG SenderThread:119 [sender.py:send():378] send: metric
160
+ 2024-04-01 09:39:22,014 WARNING SenderThread:119 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
161
+ 2024-04-01 09:39:23,793 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
162
+ 2024-04-01 09:39:24,966 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
163
+ 2024-04-01 09:39:25,015 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
164
+ 2024-04-01 09:39:29,967 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
165
+ 2024-04-01 09:39:30,016 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-04-01 09:39:33,469 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
167
+ 2024-04-01 09:39:33,470 DEBUG SenderThread:119 [sender.py:send():378] send: history
168
+ 2024-04-01 09:39:33,470 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
169
+ 2024-04-01 09:39:33,471 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
170
+ 2024-04-01 09:39:33,797 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
171
+ 2024-04-01 09:39:34,968 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
172
+ 2024-04-01 09:39:35,168 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
173
+ 2024-04-01 09:39:35,169 DEBUG SenderThread:119 [sender.py:send():378] send: history
174
+ 2024-04-01 09:39:35,169 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
175
+ 2024-04-01 09:39:35,169 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
176
+ 2024-04-01 09:39:35,170 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
177
+ 2024-04-01 09:39:35,798 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
178
+ 2024-04-01 09:39:37,798 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
179
+ 2024-04-01 09:39:39,970 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
180
+ 2024-04-01 09:39:40,367 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
181
+ 2024-04-01 09:39:44,971 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
182
+ 2024-04-01 09:39:45,373 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
183
+ 2024-04-01 09:39:45,801 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/config.yaml
184
+ 2024-04-01 09:39:46,821 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
185
+ 2024-04-01 09:39:46,823 DEBUG SenderThread:119 [sender.py:send():378] send: history
186
+ 2024-04-01 09:39:46,823 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
187
+ 2024-04-01 09:39:46,825 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
188
+ 2024-04-01 09:39:47,802 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
189
+ 2024-04-01 09:39:48,560 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
190
+ 2024-04-01 09:39:48,562 DEBUG SenderThread:119 [sender.py:send():378] send: history
191
+ 2024-04-01 09:39:48,562 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
192
+ 2024-04-01 09:39:48,562 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
193
+ 2024-04-01 09:39:48,803 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
194
+ 2024-04-01 09:39:49,972 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
195
+ 2024-04-01 09:39:50,748 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
196
+ 2024-04-01 09:39:51,804 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
197
+ 2024-04-01 09:39:54,973 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
198
+ 2024-04-01 09:39:55,749 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-04-01 09:39:59,406 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
200
+ 2024-04-01 09:39:59,406 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
201
+ 2024-04-01 09:39:59,406 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
202
+ 2024-04-01 09:39:59,406 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
203
+ 2024-04-01 09:39:59,407 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
204
+ 2024-04-01 09:39:59,407 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
205
+ 2024-04-01 09:39:59,407 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
206
+ 2024-04-01 09:39:59,409 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
207
+ 2024-04-01 09:39:59,419 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
208
+ 2024-04-01 09:39:59,419 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
209
+ 2024-04-01 09:39:59,419 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
210
+ 2024-04-01 09:39:59,420 DEBUG SenderThread:119 [sender.py:send():378] send: stats
211
+ 2024-04-01 09:39:59,974 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
212
+ 2024-04-01 09:40:01,420 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
213
+ 2024-04-01 09:40:04,975 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
214
+ 2024-04-01 09:40:06,421 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
215
+ 2024-04-01 09:40:09,976 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
216
+ 2024-04-01 09:40:11,422 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-04-01 09:40:14,977 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
218
+ 2024-04-01 09:40:16,423 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-04-01 09:40:19,979 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
220
+ 2024-04-01 09:40:21,424 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
221
+ 2024-04-01 09:40:24,979 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
222
+ 2024-04-01 09:40:26,425 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
223
+ 2024-04-01 09:40:29,980 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
224
+ 2024-04-01 09:40:31,426 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
225
+ 2024-04-01 09:40:34,981 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
226
+ 2024-04-01 09:40:36,426 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
227
+ 2024-04-01 09:40:39,982 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
228
+ 2024-04-01 09:40:41,427 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
229
+ 2024-04-01 09:40:44,983 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
230
+ 2024-04-01 09:40:46,428 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
231
+ 2024-04-01 09:40:49,984 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
232
+ 2024-04-01 09:40:51,429 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
233
+ 2024-04-01 09:40:54,985 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
234
+ 2024-04-01 09:40:56,430 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
235
+ 2024-04-01 09:40:59,986 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
236
+ 2024-04-01 09:41:00,164 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
237
+ 2024-04-01 09:41:00,164 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
238
+ 2024-04-01 09:41:00,164 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
239
+ 2024-04-01 09:41:00,164 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
240
+ 2024-04-01 09:41:00,165 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
241
+ 2024-04-01 09:41:00,165 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
242
+ 2024-04-01 09:41:00,166 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
243
+ 2024-04-01 09:41:00,166 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
244
+ 2024-04-01 09:41:00,168 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
245
+ 2024-04-01 09:41:00,170 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
246
+ 2024-04-01 09:41:00,170 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
247
+ 2024-04-01 09:41:00,170 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
248
+ 2024-04-01 09:41:00,171 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
249
+ 2024-04-01 09:41:00,171 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
250
+ 2024-04-01 09:41:00,171 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
251
+ 2024-04-01 09:41:00,172 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
252
+ 2024-04-01 09:41:00,174 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
253
+ 2024-04-01 09:41:00,192 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
254
+ 2024-04-01 09:41:00,192 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
255
+ 2024-04-01 09:41:00,192 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
256
+ 2024-04-01 09:41:00,193 DEBUG SenderThread:119 [sender.py:send():378] send: stats
257
+ 2024-04-01 09:41:02,194 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
258
+ 2024-04-01 09:41:04,987 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
259
+ 2024-04-01 09:41:07,195 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
260
+ 2024-04-01 09:41:07,655 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
261
+ 2024-04-01 09:41:07,655 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
262
+ 2024-04-01 09:41:07,655 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
263
+ 2024-04-01 09:41:07,655 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
264
+ 2024-04-01 09:41:07,656 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
265
+ 2024-04-01 09:41:07,656 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
266
+ 2024-04-01 09:41:07,657 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
267
+ 2024-04-01 09:41:07,658 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
268
+ 2024-04-01 09:41:07,658 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
269
+ 2024-04-01 09:41:08,203 DEBUG SenderThread:119 [sender.py:send():378] send: config
270
+ 2024-04-01 09:41:08,205 DEBUG SenderThread:119 [sender.py:send():378] send: metric
271
+ 2024-04-01 09:41:08,205 DEBUG SenderThread:119 [sender.py:send():378] send: metric
272
+ 2024-04-01 09:41:08,205 WARNING SenderThread:119 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
273
+ 2024-04-01 09:41:09,833 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
274
+ 2024-04-01 09:41:10,104 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
275
+ 2024-04-01 09:41:12,205 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
276
+ 2024-04-01 09:41:12,441 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
277
+ 2024-04-01 09:41:12,442 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
278
+ 2024-04-01 09:41:12,442 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
279
+ 2024-04-01 09:41:12,442 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
280
+ 2024-04-01 09:41:12,442 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
281
+ 2024-04-01 09:41:12,443 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
282
+ 2024-04-01 09:41:12,443 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
283
+ 2024-04-01 09:41:12,443 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
284
+ 2024-04-01 09:41:12,454 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
285
+ 2024-04-01 09:41:12,455 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
286
+ 2024-04-01 09:41:12,455 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
287
+ 2024-04-01 09:41:12,455 DEBUG SenderThread:119 [sender.py:send():378] send: stats
288
+ 2024-04-01 09:41:15,105 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
289
+ 2024-04-01 09:41:17,461 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
290
+ 2024-04-01 09:41:17,836 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/config.yaml
291
+ 2024-04-01 09:41:20,106 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
292
+ 2024-04-01 09:41:22,548 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
293
+ 2024-04-01 09:41:25,108 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
294
+ 2024-04-01 09:41:27,549 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
295
+ 2024-04-01 09:41:30,109 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
296
+ 2024-04-01 09:41:30,403 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
297
+ 2024-04-01 09:41:30,403 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
298
+ 2024-04-01 09:41:30,403 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
299
+ 2024-04-01 09:41:30,403 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
300
+ 2024-04-01 09:41:30,404 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
301
+ 2024-04-01 09:41:30,405 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
302
+ 2024-04-01 09:41:30,405 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
303
+ 2024-04-01 09:41:30,406 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
304
+ 2024-04-01 09:41:30,406 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
305
+ 2024-04-01 09:41:30,410 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
306
+ 2024-04-01 09:41:30,410 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
307
+ 2024-04-01 09:41:30,410 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
308
+ 2024-04-01 09:41:30,411 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
309
+ 2024-04-01 09:41:30,411 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
310
+ 2024-04-01 09:41:30,411 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
311
+ 2024-04-01 09:41:30,411 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
312
+ 2024-04-01 09:41:30,412 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
313
+ 2024-04-01 09:41:30,428 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
314
+ 2024-04-01 09:41:30,429 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
315
+ 2024-04-01 09:41:30,429 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
316
+ 2024-04-01 09:41:30,429 DEBUG SenderThread:119 [sender.py:send():378] send: stats
317
+ 2024-04-01 09:41:30,807 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
318
+ 2024-04-01 09:41:30,807 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
319
+ 2024-04-01 09:41:30,807 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
320
+ 2024-04-01 09:41:30,807 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
321
+ 2024-04-01 09:41:30,808 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
322
+ 2024-04-01 09:41:30,809 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
323
+ 2024-04-01 09:41:30,809 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
324
+ 2024-04-01 09:41:30,810 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
325
+ 2024-04-01 09:41:30,812 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
326
+ 2024-04-01 09:41:31,309 DEBUG SenderThread:119 [sender.py:send():378] send: config
327
+ 2024-04-01 09:41:31,310 DEBUG SenderThread:119 [sender.py:send():378] send: metric
328
+ 2024-04-01 09:41:31,310 DEBUG SenderThread:119 [sender.py:send():378] send: metric
329
+ 2024-04-01 09:41:31,310 WARNING SenderThread:119 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
330
+ 2024-04-01 09:41:33,311 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
331
+ 2024-04-01 09:41:35,240 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
332
+ 2024-04-01 09:41:35,576 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
333
+ 2024-04-01 09:41:35,577 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
334
+ 2024-04-01 09:41:35,577 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
335
+ 2024-04-01 09:41:35,577 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
336
+ 2024-04-01 09:41:35,577 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
337
+ 2024-04-01 09:41:35,578 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
338
+ 2024-04-01 09:41:35,578 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
339
+ 2024-04-01 09:41:35,578 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
340
+ 2024-04-01 09:41:35,590 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
341
+ 2024-04-01 09:41:35,590 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
342
+ 2024-04-01 09:41:35,590 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
343
+ 2024-04-01 09:41:35,591 DEBUG SenderThread:119 [sender.py:send():378] send: stats
344
+ 2024-04-01 09:41:38,591 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
345
+ 2024-04-01 09:41:40,242 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
346
+ 2024-04-01 09:41:43,592 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
347
+ 2024-04-01 09:41:45,080 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
348
+ 2024-04-01 09:41:45,081 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
349
+ 2024-04-01 09:41:45,081 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
350
+ 2024-04-01 09:41:45,081 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
351
+ 2024-04-01 09:41:45,081 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
352
+ 2024-04-01 09:41:45,082 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
353
+ 2024-04-01 09:41:45,084 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
354
+ 2024-04-01 09:41:45,084 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
355
+ 2024-04-01 09:41:45,085 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
356
+ 2024-04-01 09:41:45,085 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
357
+ 2024-04-01 09:41:45,085 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
358
+ 2024-04-01 09:41:45,085 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
359
+ 2024-04-01 09:41:45,086 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
360
+ 2024-04-01 09:41:45,087 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
361
+ 2024-04-01 09:41:45,087 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
362
+ 2024-04-01 09:41:45,087 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
363
+ 2024-04-01 09:41:45,090 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
364
+ 2024-04-01 09:41:45,106 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
365
+ 2024-04-01 09:41:45,107 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
366
+ 2024-04-01 09:41:45,107 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
367
+ 2024-04-01 09:41:45,107 DEBUG SenderThread:119 [sender.py:send():378] send: stats
368
+ 2024-04-01 09:41:45,243 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
369
+ 2024-04-01 09:41:48,752 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
370
+ 2024-04-01 09:41:48,753 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
371
+ 2024-04-01 09:41:48,753 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
372
+ 2024-04-01 09:41:48,753 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
373
+ 2024-04-01 09:41:48,754 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
374
+ 2024-04-01 09:41:48,754 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
375
+ 2024-04-01 09:41:48,755 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
376
+ 2024-04-01 09:41:48,756 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
377
+ 2024-04-01 09:41:48,756 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
378
+ 2024-04-01 09:41:49,113 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
379
+ 2024-04-01 09:41:49,281 DEBUG SenderThread:119 [sender.py:send():378] send: config
380
+ 2024-04-01 09:41:49,283 DEBUG SenderThread:119 [sender.py:send():378] send: metric
381
+ 2024-04-01 09:41:49,283 DEBUG SenderThread:119 [sender.py:send():378] send: metric
382
+ 2024-04-01 09:41:49,283 WARNING SenderThread:119 [sender.py:send_metric():1316] Seen metric with glob (shouldn't happen)
383
+ 2024-04-01 09:41:49,848 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/config.yaml
384
+ 2024-04-01 09:41:50,247 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
385
+ 2024-04-01 09:41:54,284 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
386
+ 2024-04-01 09:41:55,248 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
387
+ 2024-04-01 09:41:59,285 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
388
+ 2024-04-01 09:42:00,249 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
389
+ 2024-04-01 09:42:00,541 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
390
+ 2024-04-01 09:42:00,542 DEBUG SenderThread:119 [sender.py:send():378] send: history
391
+ 2024-04-01 09:42:00,542 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
392
+ 2024-04-01 09:42:00,544 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
393
+ 2024-04-01 09:42:00,852 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
394
+ 2024-04-01 09:42:02,324 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
395
+ 2024-04-01 09:42:02,325 DEBUG SenderThread:119 [sender.py:send():378] send: history
396
+ 2024-04-01 09:42:02,326 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
397
+ 2024-04-01 09:42:02,326 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
398
+ 2024-04-01 09:42:02,853 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
399
+ 2024-04-01 09:42:03,853 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
400
+ 2024-04-01 09:42:04,516 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
401
+ 2024-04-01 09:42:05,250 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
402
+ 2024-04-01 09:42:09,517 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
403
+ 2024-04-01 09:42:10,251 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
404
+ 2024-04-01 09:42:13,817 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
405
+ 2024-04-01 09:42:13,818 DEBUG SenderThread:119 [sender.py:send():378] send: history
406
+ 2024-04-01 09:42:13,818 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
407
+ 2024-04-01 09:42:13,819 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
408
+ 2024-04-01 09:42:13,857 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
409
+ 2024-04-01 09:42:14,819 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
410
+ 2024-04-01 09:42:15,253 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
411
+ 2024-04-01 09:42:15,633 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
412
+ 2024-04-01 09:42:15,634 DEBUG SenderThread:119 [sender.py:send():378] send: history
413
+ 2024-04-01 09:42:15,634 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
414
+ 2024-04-01 09:42:15,635 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
415
+ 2024-04-01 09:42:15,858 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
416
+ 2024-04-01 09:42:17,858 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
417
+ 2024-04-01 09:42:19,836 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
418
+ 2024-04-01 09:42:20,254 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
419
+ 2024-04-01 09:42:20,860 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/config.yaml
420
+ 2024-04-01 09:42:24,974 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
421
+ 2024-04-01 09:42:25,255 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
422
+ 2024-04-01 09:42:28,954 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
423
+ 2024-04-01 09:42:28,955 DEBUG SenderThread:119 [sender.py:send():378] send: history
424
+ 2024-04-01 09:42:28,955 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
425
+ 2024-04-01 09:42:28,956 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
426
+ 2024-04-01 09:42:29,863 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
427
+ 2024-04-01 09:42:30,256 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
428
+ 2024-04-01 09:42:30,834 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
429
+ 2024-04-01 09:42:30,836 DEBUG SenderThread:119 [sender.py:send():378] send: history
430
+ 2024-04-01 09:42:30,836 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
431
+ 2024-04-01 09:42:30,836 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
432
+ 2024-04-01 09:42:30,837 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
433
+ 2024-04-01 09:42:30,863 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
434
+ 2024-04-01 09:42:33,864 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
435
+ 2024-04-01 09:42:35,257 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
436
+ 2024-04-01 09:42:36,054 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
437
+ 2024-04-01 09:42:40,258 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
438
+ 2024-04-01 09:42:41,055 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
439
+ 2024-04-01 09:42:43,043 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
440
+ 2024-04-01 09:42:43,045 DEBUG SenderThread:119 [sender.py:send():378] send: history
441
+ 2024-04-01 09:42:43,045 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
442
+ 2024-04-01 09:42:43,047 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
443
+ 2024-04-01 09:42:43,869 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
444
+ 2024-04-01 09:42:44,902 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
445
+ 2024-04-01 09:42:44,903 DEBUG SenderThread:119 [sender.py:send():378] send: history
446
+ 2024-04-01 09:42:44,904 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
447
+ 2024-04-01 09:42:44,904 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
448
+ 2024-04-01 09:42:45,259 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
449
+ 2024-04-01 09:42:45,870 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
450
+ 2024-04-01 09:42:46,685 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
451
+ 2024-04-01 09:42:47,871 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
452
+ 2024-04-01 09:42:48,757 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
453
+ 2024-04-01 09:42:48,758 DEBUG SenderThread:119 [sender.py:send():378] send: stats
454
+ 2024-04-01 09:42:50,260 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
455
+ 2024-04-01 09:42:51,759 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
456
+ 2024-04-01 09:42:55,261 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
457
+ 2024-04-01 09:42:56,760 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
458
+ 2024-04-01 09:42:58,884 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
459
+ 2024-04-01 09:42:58,885 DEBUG SenderThread:119 [sender.py:send():378] send: history
460
+ 2024-04-01 09:42:58,885 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
461
+ 2024-04-01 09:42:58,886 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
462
+ 2024-04-01 09:42:59,875 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
463
+ 2024-04-01 09:43:00,262 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
464
+ 2024-04-01 09:43:00,736 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
465
+ 2024-04-01 09:43:00,737 DEBUG SenderThread:119 [sender.py:send():378] send: history
466
+ 2024-04-01 09:43:00,738 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
467
+ 2024-04-01 09:43:00,738 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
468
+ 2024-04-01 09:43:00,876 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
469
+ 2024-04-01 09:43:01,944 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
470
+ 2024-04-01 09:43:03,877 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
471
+ 2024-04-01 09:43:05,263 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
472
+ 2024-04-01 09:43:06,945 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
473
+ 2024-04-01 09:43:10,264 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
474
+ 2024-04-01 09:43:11,946 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
475
+ 2024-04-01 09:43:13,849 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
476
+ 2024-04-01 09:43:13,850 DEBUG SenderThread:119 [sender.py:send():378] send: history
477
+ 2024-04-01 09:43:13,850 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
478
+ 2024-04-01 09:43:13,852 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
479
+ 2024-04-01 09:43:13,881 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
480
+ 2024-04-01 09:43:15,265 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
481
+ 2024-04-01 09:43:15,688 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
482
+ 2024-04-01 09:43:15,689 DEBUG SenderThread:119 [sender.py:send():378] send: history
483
+ 2024-04-01 09:43:15,690 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
484
+ 2024-04-01 09:43:15,690 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
485
+ 2024-04-01 09:43:15,882 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
486
+ 2024-04-01 09:43:17,883 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
487
+ 2024-04-01 09:43:17,909 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
488
+ 2024-04-01 09:43:18,759 DEBUG SenderThread:119 [sender.py:send():378] send: stats
489
+ 2024-04-01 09:43:20,266 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
490
+ 2024-04-01 09:43:23,761 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
491
+ 2024-04-01 09:43:25,268 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
492
+ 2024-04-01 09:43:27,359 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
493
+ 2024-04-01 09:43:27,360 DEBUG SenderThread:119 [sender.py:send():378] send: history
494
+ 2024-04-01 09:43:27,361 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
495
+ 2024-04-01 09:43:27,361 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
496
+ 2024-04-01 09:43:27,886 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
497
+ 2024-04-01 09:43:29,176 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
498
+ 2024-04-01 09:43:29,177 DEBUG SenderThread:119 [sender.py:send():378] send: history
499
+ 2024-04-01 09:43:29,178 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
500
+ 2024-04-01 09:43:29,178 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
501
+ 2024-04-01 09:43:29,179 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
502
+ 2024-04-01 09:43:29,887 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
503
+ 2024-04-01 09:43:30,269 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
504
+ 2024-04-01 09:43:31,888 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
505
+ 2024-04-01 09:43:34,386 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
506
+ 2024-04-01 09:43:35,270 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
507
+ 2024-04-01 09:43:39,387 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
508
+ 2024-04-01 09:43:40,271 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
509
+ 2024-04-01 09:43:42,423 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
510
+ 2024-04-01 09:43:42,425 DEBUG SenderThread:119 [sender.py:send():378] send: history
511
+ 2024-04-01 09:43:42,425 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
512
+ 2024-04-01 09:43:42,426 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
513
+ 2024-04-01 09:43:42,892 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
514
+ 2024-04-01 09:43:44,292 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
515
+ 2024-04-01 09:43:44,294 DEBUG SenderThread:119 [sender.py:send():378] send: history
516
+ 2024-04-01 09:43:44,294 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
517
+ 2024-04-01 09:43:44,296 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
518
+ 2024-04-01 09:43:44,496 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
519
+ 2024-04-01 09:43:44,893 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
520
+ 2024-04-01 09:43:45,272 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
521
+ 2024-04-01 09:43:45,893 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
522
+ 2024-04-01 09:43:48,760 DEBUG SenderThread:119 [sender.py:send():378] send: stats
523
+ 2024-04-01 09:43:49,761 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
524
+ 2024-04-01 09:43:50,273 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
525
+ 2024-04-01 09:43:54,762 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
526
+ 2024-04-01 09:43:55,274 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
527
+ 2024-04-01 09:43:57,622 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
528
+ 2024-04-01 09:43:57,623 DEBUG SenderThread:119 [sender.py:send():378] send: history
529
+ 2024-04-01 09:43:57,624 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
530
+ 2024-04-01 09:43:57,624 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
531
+ 2024-04-01 09:43:57,898 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
532
+ 2024-04-01 09:43:59,489 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
533
+ 2024-04-01 09:43:59,490 DEBUG SenderThread:119 [sender.py:send():378] send: history
534
+ 2024-04-01 09:43:59,490 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
535
+ 2024-04-01 09:43:59,490 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
536
+ 2024-04-01 09:43:59,899 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
537
+ 2024-04-01 09:44:00,275 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
538
+ 2024-04-01 09:44:00,695 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
539
+ 2024-04-01 09:44:01,899 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
540
+ 2024-04-01 09:44:05,276 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
541
+ 2024-04-01 09:44:05,696 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
542
+ 2024-04-01 09:44:10,277 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
543
+ 2024-04-01 09:44:10,697 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
544
+ 2024-04-01 09:44:11,309 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
545
+ 2024-04-01 09:44:11,310 DEBUG SenderThread:119 [sender.py:send():378] send: history
546
+ 2024-04-01 09:44:11,311 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
547
+ 2024-04-01 09:44:11,311 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
548
+ 2024-04-01 09:44:11,903 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
549
+ 2024-04-01 09:44:13,131 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
550
+ 2024-04-01 09:44:13,132 DEBUG SenderThread:119 [sender.py:send():378] send: history
551
+ 2024-04-01 09:44:13,132 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
552
+ 2024-04-01 09:44:13,133 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
553
+ 2024-04-01 09:44:13,903 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
554
+ 2024-04-01 09:44:15,278 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
555
+ 2024-04-01 09:44:15,904 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
556
+ 2024-04-01 09:44:16,334 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
557
+ 2024-04-01 09:44:18,761 DEBUG SenderThread:119 [sender.py:send():378] send: stats
558
+ 2024-04-01 09:44:20,279 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
559
+ 2024-04-01 09:44:21,762 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
560
+ 2024-04-01 09:44:24,824 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
561
+ 2024-04-01 09:44:24,825 DEBUG SenderThread:119 [sender.py:send():378] send: history
562
+ 2024-04-01 09:44:24,826 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
563
+ 2024-04-01 09:44:24,826 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
564
+ 2024-04-01 09:44:24,907 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
565
+ 2024-04-01 09:44:25,280 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
566
+ 2024-04-01 09:44:26,644 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
567
+ 2024-04-01 09:44:26,645 DEBUG SenderThread:119 [sender.py:send():378] send: history
568
+ 2024-04-01 09:44:26,646 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
569
+ 2024-04-01 09:44:26,646 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
570
+ 2024-04-01 09:44:26,848 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
571
+ 2024-04-01 09:44:26,908 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
572
+ 2024-04-01 09:44:29,909 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
573
+ 2024-04-01 09:44:30,281 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
574
+ 2024-04-01 09:44:31,848 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
575
+ 2024-04-01 09:44:35,282 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
576
+ 2024-04-01 09:44:36,849 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
577
+ 2024-04-01 09:44:39,728 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
578
+ 2024-04-01 09:44:39,729 DEBUG SenderThread:119 [sender.py:send():378] send: history
579
+ 2024-04-01 09:44:39,730 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
580
+ 2024-04-01 09:44:39,730 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
581
+ 2024-04-01 09:44:39,912 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
582
+ 2024-04-01 09:44:40,283 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
583
+ 2024-04-01 09:44:41,561 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: partial_history
584
+ 2024-04-01 09:44:41,562 DEBUG SenderThread:119 [sender.py:send():378] send: history
585
+ 2024-04-01 09:44:41,562 DEBUG SenderThread:119 [sender.py:send_request():405] send_request: summary_record
586
+ 2024-04-01 09:44:41,562 INFO SenderThread:119 [sender.py:_save_file():1365] saving file wandb-summary.json with policy end
587
+ 2024-04-01 09:44:41,913 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/wandb-summary.json
588
+ 2024-04-01 09:44:42,764 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
589
+ 2024-04-01 09:44:43,914 INFO Thread-12 :119 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/files/output.log
590
+ 2024-04-01 09:44:45,284 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
591
+ 2024-04-01 09:44:47,765 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
592
+ 2024-04-01 09:44:48,762 DEBUG SenderThread:119 [sender.py:send():378] send: stats
593
+ 2024-04-01 09:44:50,285 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
594
+ 2024-04-01 09:44:51,614 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
595
+ 2024-04-01 09:44:51,614 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
596
+ 2024-04-01 09:44:51,614 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
597
+ 2024-04-01 09:44:51,615 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
598
+ 2024-04-01 09:44:51,615 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
599
+ 2024-04-01 09:44:51,615 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
600
+ 2024-04-01 09:44:51,616 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
601
+ 2024-04-01 09:44:51,627 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
602
+ 2024-04-01 09:44:51,627 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
603
+ 2024-04-01 09:44:51,627 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
604
+ 2024-04-01 09:44:51,628 DEBUG SenderThread:119 [sender.py:send():378] send: stats
605
+ 2024-04-01 09:44:53,628 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
606
+ 2024-04-01 09:44:55,287 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
607
+ 2024-04-01 09:44:58,630 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
608
+ 2024-04-01 09:45:00,288 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
609
+ 2024-04-01 09:45:03,631 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
610
+ 2024-04-01 09:45:05,289 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
611
+ 2024-04-01 09:45:08,631 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
612
+ 2024-04-01 09:45:10,291 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
613
+ 2024-04-01 09:45:13,632 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
614
+ 2024-04-01 09:45:15,292 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
615
+ 2024-04-01 09:45:18,633 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
616
+ 2024-04-01 09:45:20,293 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
617
+ 2024-04-01 09:45:23,634 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
618
+ 2024-04-01 09:45:25,294 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
619
+ 2024-04-01 09:45:28,635 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
620
+ 2024-04-01 09:45:30,295 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
621
+ 2024-04-01 09:45:32,015 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
622
+ 2024-04-01 09:45:32,016 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
623
+ 2024-04-01 09:45:32,016 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
624
+ 2024-04-01 09:45:32,016 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
625
+ 2024-04-01 09:45:32,016 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
626
+ 2024-04-01 09:45:32,019 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
627
+ 2024-04-01 09:45:32,019 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
628
+ 2024-04-01 09:45:32,020 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
629
+ 2024-04-01 09:45:32,023 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
630
+ 2024-04-01 09:45:32,039 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
631
+ 2024-04-01 09:45:32,040 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
632
+ 2024-04-01 09:45:32,040 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
633
+ 2024-04-01 09:45:32,040 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
634
+ 2024-04-01 09:45:32,040 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
635
+ 2024-04-01 09:45:32,040 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
636
+ 2024-04-01 09:45:32,040 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
637
+ 2024-04-01 09:45:32,041 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
638
+ 2024-04-01 09:45:32,055 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
639
+ 2024-04-01 09:45:32,055 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
640
+ 2024-04-01 09:45:32,055 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
641
+ 2024-04-01 09:45:32,055 DEBUG SenderThread:119 [sender.py:send():378] send: stats
642
+ 2024-04-01 09:45:34,056 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
643
+ 2024-04-01 09:45:35,296 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
644
+ 2024-04-01 09:45:39,057 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
645
+ 2024-04-01 09:45:40,298 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
646
+ 2024-04-01 09:45:44,058 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
647
+ 2024-04-01 09:45:45,299 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
648
+ 2024-04-01 09:45:49,059 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
649
+ 2024-04-01 09:45:50,299 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
650
+ 2024-04-01 09:45:54,060 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
651
+ 2024-04-01 09:45:54,718 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
652
+ 2024-04-01 09:45:54,718 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
653
+ 2024-04-01 09:45:54,718 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
654
+ 2024-04-01 09:45:54,718 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
655
+ 2024-04-01 09:45:54,719 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
656
+ 2024-04-01 09:45:54,719 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
657
+ 2024-04-01 09:45:54,720 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: pause
658
+ 2024-04-01 09:45:54,720 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
659
+ 2024-04-01 09:45:54,720 INFO HandlerThread:119 [handler.py:handle_request_pause():708] stopping system metrics thread
660
+ 2024-04-01 09:45:54,721 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
661
+ 2024-04-01 09:45:54,721 INFO HandlerThread:119 [system_monitor.py:finish():203] Stopping system monitor
662
+ 2024-04-01 09:45:54,722 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
663
+ 2024-04-01 09:45:54,722 DEBUG SystemMonitor:119 [system_monitor.py:_start():172] Starting system metrics aggregation loop
664
+ 2024-04-01 09:45:54,722 DEBUG SystemMonitor:119 [system_monitor.py:_start():179] Finished system metrics aggregation loop
665
+ 2024-04-01 09:45:54,722 DEBUG SystemMonitor:119 [system_monitor.py:_start():183] Publishing last batch of metrics
666
+ 2024-04-01 09:45:54,725 INFO HandlerThread:119 [interfaces.py:finish():202] Joined cpu monitor
667
+ 2024-04-01 09:45:54,727 INFO HandlerThread:119 [interfaces.py:finish():202] Joined disk monitor
668
+ 2024-04-01 09:45:54,747 INFO HandlerThread:119 [interfaces.py:finish():202] Joined gpu monitor
669
+ 2024-04-01 09:45:54,747 INFO HandlerThread:119 [interfaces.py:finish():202] Joined memory monitor
670
+ 2024-04-01 09:45:54,747 INFO HandlerThread:119 [interfaces.py:finish():202] Joined network monitor
671
+ 2024-04-01 09:45:54,747 DEBUG SenderThread:119 [sender.py:send():378] send: stats
672
+ 2024-04-01 09:45:55,301 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
673
+ 2024-04-01 09:45:57,946 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: resume
674
+ 2024-04-01 09:45:57,946 INFO HandlerThread:119 [handler.py:handle_request_resume():699] starting system metrics thread
675
+ 2024-04-01 09:45:57,946 INFO HandlerThread:119 [system_monitor.py:start():194] Starting system monitor
676
+ 2024-04-01 09:45:57,947 INFO SystemMonitor:119 [system_monitor.py:_start():158] Starting system asset monitoring threads
677
+ 2024-04-01 09:45:57,947 INFO SystemMonitor:119 [interfaces.py:start():190] Started cpu monitoring
678
+ 2024-04-01 09:45:57,948 INFO SystemMonitor:119 [interfaces.py:start():190] Started disk monitoring
679
+ 2024-04-01 09:45:57,950 INFO SystemMonitor:119 [interfaces.py:start():190] Started gpu monitoring
680
+ 2024-04-01 09:45:57,951 INFO SystemMonitor:119 [interfaces.py:start():190] Started memory monitoring
681
+ 2024-04-01 09:45:57,951 INFO SystemMonitor:119 [interfaces.py:start():190] Started network monitoring
682
+ 2024-04-01 09:45:59,749 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
683
+ 2024-04-01 09:46:00,302 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
684
+ 2024-04-01 09:46:04,750 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
685
+ 2024-04-01 09:46:05,303 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
686
+ 2024-04-01 09:46:09,751 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: status_report
687
+ 2024-04-01 09:46:10,304 DEBUG HandlerThread:119 [handler.py:handle_request():146] handle_request: keepalive
wandb/run-20240401_093743-ie4jv7u5/logs/debug.log ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug.log
10
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug-internal.log
11
+ 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e30ad2cb760>
12
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-01 09:37:43,081 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-01 09:37:43,085 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-01 09:37:43,097 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-01 09:37:43,596 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-01 09:37:43,599 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-01 09:37:43,765 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-01 09:38:14,888 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-01 09:38:14,891 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-01 09:38:14,896 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-37-28_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-01 09:38:55,054 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-01 09:38:55,054 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-01 09:39:21,055 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-01 09:39:21,061 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-01 09:39:21,061 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-01 09:39:21,457 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-01 09:39:22,007 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-39-21_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
39
+ 2024-04-01 09:39:59,405 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
40
+ 2024-04-01 09:39:59,405 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
41
+ 2024-04-01 09:41:00,163 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
42
+ 2024-04-01 09:41:00,169 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
43
+ 2024-04-01 09:41:00,169 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
44
+ 2024-04-01 09:41:07,654 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
45
+ 2024-04-01 09:41:08,198 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-00_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
46
+ 2024-04-01 09:41:12,441 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
+ 2024-04-01 09:41:12,441 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
48
+ 2024-04-01 09:41:30,402 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
49
+ 2024-04-01 09:41:30,410 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
+ 2024-04-01 09:41:30,410 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
51
+ 2024-04-01 09:41:30,806 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
52
+ 2024-04-01 09:41:31,300 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-30_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
53
+ 2024-04-01 09:41:35,576 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-01 09:41:35,576 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
55
+ 2024-04-01 09:41:45,080 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
56
+ 2024-04-01 09:41:45,084 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-01 09:41:45,085 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
58
+ 2024-04-01 09:41:48,751 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
59
+ 2024-04-01 09:41:49,276 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 5, 'per_device_eval_batch_size': 5, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-45_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
60
+ 2024-04-01 09:44:51,613 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
61
+ 2024-04-01 09:44:51,613 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
62
+ 2024-04-01 09:45:32,015 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
63
+ 2024-04-01 09:45:32,039 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
64
+ 2024-04-01 09:45:32,039 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
65
+ 2024-04-01 09:45:54,716 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
66
+ 2024-04-01 09:45:54,719 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
67
+ 2024-04-01 09:45:54,719 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
68
+ 2024-04-01 09:45:57,946 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
wandb/run-20240401_093743-ie4jv7u5/run-ie4jv7u5.wandb ADDED
Binary file (61.2 kB). View file