Narkantak commited on
Commit
d034c21
1 Parent(s): cef7e76

Narkantak/mistral-7b-Intent-Classifier-Ashu

Browse files
README.md CHANGED
@@ -16,12 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - eval_loss: 1.0970
20
- - eval_runtime: 1.8296
21
- - eval_samples_per_second: 1.093
22
- - eval_steps_per_second: 0.547
23
- - epoch: 12.0
24
- - step: 16
25
 
26
  ## Model description
27
 
@@ -41,17 +36,52 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 0.0002
44
- - train_batch_size: 5
45
- - eval_batch_size: 5
46
  - seed: 42
47
- - gradient_accumulation_steps: 3
48
- - total_train_batch_size: 15
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
- - lr_scheduler_warmup_steps: 1
52
- - num_epochs: 50
53
  - mixed_precision_training: Native AMP
54
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
  ### Framework versions
56
 
57
  - PEFT 0.10.0
 
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.5491
 
 
 
 
 
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
+ - train_batch_size: 6
40
+ - eval_batch_size: 6
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 24
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 2
47
+ - num_epochs: 30
48
  - mixed_precision_training: Native AMP
49
 
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:-----:|:----:|:---------------:|
54
+ | 3.8811 | 0.96 | 12 | 2.3732 |
55
+ | 1.3483 | 2.0 | 25 | 0.7815 |
56
+ | 0.7479 | 2.96 | 37 | 0.6088 |
57
+ | 0.5459 | 4.0 | 50 | 0.5306 |
58
+ | 0.5392 | 4.96 | 62 | 0.5074 |
59
+ | 0.4672 | 6.0 | 75 | 0.4910 |
60
+ | 0.4799 | 6.96 | 87 | 0.4779 |
61
+ | 0.4205 | 8.0 | 100 | 0.4742 |
62
+ | 0.4357 | 8.96 | 112 | 0.4721 |
63
+ | 0.3848 | 10.0 | 125 | 0.4694 |
64
+ | 0.4006 | 10.96 | 137 | 0.4703 |
65
+ | 0.3579 | 12.0 | 150 | 0.4736 |
66
+ | 0.3768 | 12.96 | 162 | 0.4850 |
67
+ | 0.3373 | 14.0 | 175 | 0.4939 |
68
+ | 0.3582 | 14.96 | 187 | 0.4883 |
69
+ | 0.3239 | 16.0 | 200 | 0.4974 |
70
+ | 0.3445 | 16.96 | 212 | 0.5045 |
71
+ | 0.3129 | 18.0 | 225 | 0.5131 |
72
+ | 0.3362 | 18.96 | 237 | 0.5189 |
73
+ | 0.3075 | 20.0 | 250 | 0.5216 |
74
+ | 0.3302 | 20.96 | 262 | 0.5277 |
75
+ | 0.3028 | 22.0 | 275 | 0.5392 |
76
+ | 0.3268 | 22.96 | 287 | 0.5363 |
77
+ | 0.3006 | 24.0 | 300 | 0.5420 |
78
+ | 0.3249 | 24.96 | 312 | 0.5422 |
79
+ | 0.2994 | 26.0 | 325 | 0.5437 |
80
+ | 0.3236 | 26.96 | 337 | 0.5464 |
81
+ | 0.2979 | 28.0 | 350 | 0.5489 |
82
+ | 0.3094 | 28.8 | 360 | 0.5491 |
83
+
84
+
85
  ### Framework versions
86
 
87
  - PEFT 0.10.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bb4ad88bb45adc123b696e91c70736275c8d379919cb30b6f99cb1320a0d768
3
  size 8397056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fd9f83307a578ba77bc5360a29cd2aa2446c34c2de0de3756cd9007e24089ec
3
  size 8397056
runs/Apr02_07-24-52_0ecfc053b35f/events.out.tfevents.1712042696.0ecfc053b35f.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8721f24834dcf69bc4975ef1c816ec43becbb8c7b69d27597361091b7ee237a1
3
+ size 19480
test.csv CHANGED
@@ -1,34 +1,76 @@
1
  example
2
- "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
3
- Orion is known for its best precision and world class accuracy with the financial knowledge, \
4
- Document Type: Document Type 7
5
-
6
- , context associated to the document type is
7
-
8
- No Information Available [/INST] "
9
- "[INST] Orion, Act as a financial analyst with a photographic memory and respond to the user query based on the type of document type. \
10
- Orion is known for its best precision and world class accuracy with the financial knowledge, \
11
- Document Type: Document Type 1
12
-
13
- , context associated to the document type is
14
-
15
- (DT - OMB control number: 1545-0047/Expiration Date: 12/31/2019)(DOL - OMB control number: 1210-0147/Expiration date: 5/31/2022) Page 1 of 5
16
- (HHS - OMB control number: 0938-1146/Expiration date: 10/31/2022)
17
- Summary of Benefits and Coverage: What this Plan Covers & What You Pay for Covered Services Coverage Period: Beginning on or after 01/01/2021
18
- Community Blue – Faculty/Acad Active: Michigan State University Coverage for: Individual/Family | PlanType:PPO
19
- The Summary of Benefits and Coverage (SBC) document will help you choose a health plan. The SBC shows you how you and the plan would share the cost for covered health care services. NOTE: Information about the cost of this plan (called the premium) will be provided separately. This is only a summary. For more information about your coverage, or to get a copy of the complete terms of coverage, visit www.bcbsm.com or call 888-288-1726. For general definitions of common terms, such as allowed amount, balance billing, coinsurance, copayment, deductible, provider, or other underlined terms, see the Glossary. You can view the Glossary at https://www.healthcare.gov/sbc-glossary or call 888-288-1726 to request a copy.
20
- Important Questions Answers Why This Matters: In-Network Out-of-Network
21
- What is the overall deductible?
22
- $100 Individual/ $200 Family
23
- $250 Individual/
24
- $500 Family
25
- Generally, you must pay all of the costs from provider's up to the deductible amount before this plan begins to pay. If you have other family members on the plan, each family member must meet their own individual deductible until the total amount of deductible expenses paid by all family members meets the overall family deductible. Are there services covered before you meet your deductible? Yes This plan covers some items and services even if you haven’t yet met the deductible amount. But a copayment or coinsurance may apply. For example, this plan covers certain preventive services without cost-sharing and before you meet your deductible. See a list of covered preventive services at https://www.healthcare.gov/coverage/preventive-care-benefits/
26
- Are there other deductibles for specific services?
27
- No
28
- You don’t have to meet deductibles for specific services. What is the out-of-pocket limit for this plan? $2,000 Individual/ $4,000 Family $2,000 Individual/ $4,000 Family Plus deductible The out-of-pocket limit is the most you could pay in a year for covered services. If you have other family members in this plan, they have to meet their own out-of-pocket limits until the overall family out-of-pocket limit has been met.
29
- What is not included in the out-of-pocket limit?
30
- Premiums, balance-billed charges, any pharmacy penalty and health care this plan doesn’t cover.
31
- Even though you pay these expenses, they don’t count toward the out-of-pocket limit. Will you pay less if you use a network provider? Yes. For a list of in-network providers, see www.bcbsm.com or call the number on the back of your BCBSM ID card. This plan uses a provider network. You will pay less if you use a provider in the plan’s network. You will pay the most if you use an out-of-network provider, and you might receive a bill from a provider for the difference between the provider’s charge and what your plan pays (balance billing).
32
- Do you need a referral to see a specialist?
33
- No
34
- You can see the specialist you choose without a referral. [/INST] "
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  example
2
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for home health care? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
3
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Upgrade 'HIJ's employment status to full-time within 'Nu Assembly.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
4
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What are the services the plan does not cover? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
5
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Add a dependent to the insurance policy for 'Beta Group.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
6
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Plant a new signee within the 'Whiskey Command' organization. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
7
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the role of the Glossary in understanding the health plan? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
8
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Conduct the removal of an employee from the 'Zeta Coalition.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
9
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Transition 'ABC' from contingent worker to full-time staff in 'Psi Division.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
10
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Add a new born baby to group 'ABC' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
11
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : How does my compensation compare to the industry standard for my role? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
12
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : tell me my commission for year 'yyyy'? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
13
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Process a job status upgrade for 'PQR' to a full-time position in 'Mu Squad.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
14
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for home health care? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
15
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for home health care? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
16
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the overall deductible for in-network and out-of-network services in this health plan? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
17
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Will you pay less if you use a network provider? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
18
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Enroll a new apprentice into the 'Kilo Platoon' corps. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
19
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the percentage of cost covered for generic drugs? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
20
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Insert a new child under 'STU's medical coverage in 'Nu Division.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
21
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost of emergency room care? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
22
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Slot in a new addition to the 'Mike Regiment' lineup. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
23
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost of emergency room care? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
24
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Entrench a new specialist in the 'Delta Formation' echelon. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
25
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What are the specifics of my health and wellness benefits for this year? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
26
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the percentage of cost covered for preferred brand drugs? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
27
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : tell me my commission for year 2023? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
28
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost of emergency medical attention? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
29
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Is adult dental care covered? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
30
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What are the services the plan does not cover? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
31
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Attach a domestic partner to 'CDE's account in 'Iota Brigade.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
32
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Adjoin a new partner to the 'Hotel Taskforce' cluster. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
33
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the percentage of cost covered for generic drugs? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
34
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Extend full-time employment benefits to 'LMN' in the 'Pi Network.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
35
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for childbirth/delivery facility services? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
36
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Is long-term care covered? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
37
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the total cost Peg would pay for having a baby? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
38
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What's the total payout I received for hitting my annual targets? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
39
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the coverage period? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
40
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Can you provide a summary of my earnings from all sources for the past year? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
41
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Coordinate the exit of a staff member from the 'Pi Specialists.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
42
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What proportion of the costs for emergency medical attention does the plan cover? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
43
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What amount did I earn in referral bonuses for the past six months? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
44
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the total value of my compensation including benefits for the current year? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
45
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Can you provide a history of my commission rates over the past five years? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
46
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What are the costs associated with a hospital stay? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
47
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for generic drugs? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
48
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Conduct the end of service for a staffer in the 'Pi Consortium.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
49
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What costs are not included in the out-of-pocket limit? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
50
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Are children's eye exams covered? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
51
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the plan type? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
52
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Carry out the dismissal of a member from the 'Iota Faction.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
53
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is not included in the out-of-pocket limit? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC/Plan Doc</s>"
54
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the OMB control number and expiration date for HHS? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
55
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : How much does outpatient surgery cost? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
56
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost of prescription drug coverage? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
57
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Add a dependant to the 'Lambda Platoon' insurance policy for 'MNO.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
58
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the purpose of the Summary of Benefits and Coverage (SBC) document? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
59
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Are children's glasses covered? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
60
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Implement a marital status update for 'PQR' in the 'Chi Team' records. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
61
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Arrange for the departure of a contributor from the 'Theta Battalion.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
62
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Can you check if my compensation for overtime hours has been calculated correctly? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
63
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Fasten a new recruit to the 'Papa Command' company. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
64
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Amend the job classification from part-time to full-time for 'QRS' in 'Phi Troop.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Change</s>"
65
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for diagnostic tests such as x-rays and blood work? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
66
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What are the details of my incentive plan for the new business targets? [/INST] Hi! Centaurus at your service, here is the intent of your statement: Compensation</s>"
67
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for childbirth/delivery professional services? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
68
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Debut a new staff member in the 'November Coalition' alliance. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
69
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Are there other deductibles for specific services? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
70
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Knit a new participant into the 'Sierra Brigade' fabric. [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
71
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Are there any services that the plan does not cover? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
72
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the cost for rehabilitation services? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
73
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Organize the dismissal of a staffer from the 'Chi Network.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
74
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Finalize the termination of a signee from the 'Psi Brigade.' [/INST] Hi! Centaurus at your service, here is the intent of your statement: Terminate</s>"
75
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : Add a new member [/INST] Hi! Centaurus at your service, here is the intent of your statement: Install</s>"
76
+ "<s>[INST] Your Name is Centaurus, Start the response as, 'Hi! Centaurus at you service, here is ...' , Centaurus you are a intent classification wizard which can classify the intent of a statement from one among these: Compensation, SBC, SBC/Plan Docs, Install, Terminate, Change), Respond with intent of the statement : What is the overall deductible for this plan? [/INST] Hi! Centaurus at your service, here is the intent of your statement: SBC</s>"
test.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b71e3328ddf3c264c43471b314f5f9dd852688fedfae9484b9ee9f836b062f28
3
- size 16156
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10c069c6b3be0d3fd94e8c4783ff1e7003714427bc4774ec4065577066633014
3
+ size 9179
train.csv CHANGED
The diff for this file is too large to render. See raw diff
 
train.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3787510ef2fd47560b05aa86e99d6eb5c174a1b8d7ac31670e8fc45bf8fb5c39
3
- size 7445
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7daa5400b6b559de7964271f2c4744aa2bae329317741c2a9e4be45462433c2
3
+ size 22135
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10ec4641923a16c59aa087019cc91de846e9c70e43dec12655934e34195ad3fe
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:635892b8c2d90d49a146c0560d6d090ae50b7a606786fbe44ed5cad3448ca17e
3
  size 4920
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,68 +1,43 @@
1
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug.log
10
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240401_093743-ie4jv7u5/logs/debug-internal.log
11
- 2024-04-01 09:37:43,078 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e30ad2cb760>
12
- 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
- 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
- 2024-04-01 09:37:43,079 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
- 2024-04-01 09:37:43,081 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-01 09:37:43,085 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
- 2024-04-01 09:37:43,097 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
- 2024-04-01 09:37:43,596 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
- 2024-04-01 09:37:43,599 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
- 2024-04-01 09:37:43,765 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
- 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-01 09:37:43,855 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
- 2024-04-01 09:38:14,888 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
- 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
- 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
- 2024-04-01 09:38:14,889 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
- 2024-04-01 09:38:14,891 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
- 2024-04-01 09:38:14,896 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-37-28_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
- 2024-04-01 09:38:55,054 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
- 2024-04-01 09:38:55,054 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
- 2024-04-01 09:39:21,055 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
- 2024-04-01 09:39:21,061 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
- 2024-04-01 09:39:21,061 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
- 2024-04-01 09:39:21,457 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
- 2024-04-01 09:39:22,007 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-39-21_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
39
- 2024-04-01 09:39:59,405 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
40
- 2024-04-01 09:39:59,405 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
41
- 2024-04-01 09:41:00,163 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
42
- 2024-04-01 09:41:00,169 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
43
- 2024-04-01 09:41:00,169 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
44
- 2024-04-01 09:41:07,654 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
45
- 2024-04-01 09:41:08,198 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-00_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
46
- 2024-04-01 09:41:12,441 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
47
- 2024-04-01 09:41:12,441 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
48
- 2024-04-01 09:41:30,402 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
49
- 2024-04-01 09:41:30,410 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
50
- 2024-04-01 09:41:30,410 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
51
- 2024-04-01 09:41:30,806 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
52
- 2024-04-01 09:41:31,300 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-30_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
53
- 2024-04-01 09:41:35,576 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
- 2024-04-01 09:41:35,576 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
55
- 2024-04-01 09:41:45,080 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
56
- 2024-04-01 09:41:45,084 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
- 2024-04-01 09:41:45,085 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
58
- 2024-04-01 09:41:48,751 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
59
- 2024-04-01 09:41:49,276 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 5, 'per_device_eval_batch_size': 5, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 3, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 50, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr01_09-41-45_148e249fd8ed', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
60
- 2024-04-01 09:44:51,613 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
61
- 2024-04-01 09:44:51,613 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
62
- 2024-04-01 09:45:32,015 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
63
- 2024-04-01 09:45:32,039 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
64
- 2024-04-01 09:45:32,039 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
65
- 2024-04-01 09:45:54,716 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
66
- 2024-04-01 09:45:54,719 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
67
- 2024-04-01 09:45:54,719 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
68
- 2024-04-01 09:45:57,946 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
 
1
+ 2024-04-02 07:25:03,309 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-02 07:25:03,309 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240402_072503-b95w935d/logs/debug.log
10
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240402_072503-b95w935d/logs/debug-internal.log
11
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e03f1c07cd0>
12
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-02 07:25:03,311 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-02 07:25:03,311 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-02 07:25:03,312 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-02 07:25:03,315 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-02 07:25:03,328 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-02 07:25:03,908 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-02 07:25:03,911 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-02 07:25:04,116 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-02 07:25:04,182 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-02 07:25:04,183 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-02 07:25:35,178 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-02 07:25:35,178 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-02 07:25:35,179 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-02 07:25:35,179 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-02 07:25:35,180 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-02 07:25:35,186 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr02_07-24-52_0ecfc053b35f', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-02 09:10:25,449 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-02 09:10:25,449 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-02 09:10:45,681 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-02 09:10:45,704 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-02 09:10:45,704 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-02 09:11:24,497 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-02 09:11:24,498 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-04-02 09:11:24,498 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
+ 2024-04-02 09:11:34,304 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
41
+ 2024-04-02 09:11:34,306 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-04-02 09:11:34,306 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
43
+ 2024-04-02 09:11:35,541 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20240402_072503-b95w935d/files/conda-environment.yaml ADDED
@@ -0,0 +1,1072 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: base
2
+ channels:
3
+ - pytorch
4
+ - file:///tmp/conda
5
+ - rapidsai
6
+ - nvidia
7
+ - conda-forge
8
+ - defaults
9
+ dependencies:
10
+ - _libgcc_mutex=0.1=conda_forge
11
+ - _openmp_mutex=4.5=2_gnu
12
+ - aiohttp=3.9.1=py310h2372a71_0
13
+ - aiosignal=1.3.1=pyhd8ed1ab_0
14
+ - annotated-types=0.6.0=pyhd8ed1ab_0
15
+ - anyio=4.2.0=pyhd8ed1ab_0
16
+ - archspec=0.2.2=pyhd8ed1ab_0
17
+ - argon2-cffi=23.1.0=pyhd8ed1ab_0
18
+ - argon2-cffi-bindings=21.2.0=py310h2372a71_4
19
+ - arrow=1.3.0=pyhd8ed1ab_0
20
+ - arrow-cpp=11.0.0=ha770c72_9_cpu
21
+ - asttokens=2.4.1=pyhd8ed1ab_0
22
+ - async-timeout=4.0.3=pyhd8ed1ab_0
23
+ - atk-1.0=2.38.0=hd4edc92_1
24
+ - attrs=23.2.0=pyh71513ae_0
25
+ - aws-c-auth=0.6.26=h987a71b_2
26
+ - aws-c-cal=0.5.21=h48707d8_2
27
+ - aws-c-common=0.8.14=h0b41bf4_0
28
+ - aws-c-compression=0.2.16=h03acc5a_5
29
+ - aws-c-event-stream=0.2.20=h00877a2_4
30
+ - aws-c-http=0.7.6=hf342b9f_0
31
+ - aws-c-io=0.13.19=h5b20300_3
32
+ - aws-c-mqtt=0.8.6=hc4349f7_12
33
+ - aws-c-s3=0.2.7=h909e904_1
34
+ - aws-c-sdkutils=0.1.9=h03acc5a_0
35
+ - aws-checksums=0.1.14=h03acc5a_5
36
+ - aws-crt-cpp=0.19.8=hf7fbfca_12
37
+ - aws-sdk-cpp=1.10.57=h17c43bd_8
38
+ - beautifulsoup4=4.12.2=pyha770c72_0
39
+ - bleach=6.1.0=pyhd8ed1ab_0
40
+ - boltons=23.1.1=pyhd8ed1ab_0
41
+ - brotli=1.0.9=h166bdaf_9
42
+ - brotli-bin=1.0.9=h166bdaf_9
43
+ - brotli-python=1.0.9=py310hd8f1fbe_9
44
+ - brotlipy=0.7.0=py310h7f8727e_1002
45
+ - bzip2=1.0.8=h7b6447c_0
46
+ - c-ares=1.25.0=hd590300_0
47
+ - ca-certificates=2024.2.2=hbcca054_0
48
+ - cached-property=1.5.2=hd8ed1ab_1
49
+ - cached_property=1.5.2=pyha770c72_1
50
+ - cairo=1.18.0=h3faef2a_0
51
+ - cartopy=0.22.0=py310hcc13569_1
52
+ - catalogue=2.0.10=py310hff52083_0
53
+ - certifi=2024.2.2=pyhd8ed1ab_0
54
+ - cffi=1.16.0=py310h2fee648_0
55
+ - charset-normalizer=3.3.2=pyhd8ed1ab_0
56
+ - click=8.1.7=unix_pyh707e725_0
57
+ - cloudpathlib=0.16.0=pyhd8ed1ab_0
58
+ - colorama=0.4.6=pyhd8ed1ab_0
59
+ - comm=0.2.1=pyhd8ed1ab_0
60
+ - conda=23.7.4=py310hff52083_0
61
+ - conda-libmamba-solver=23.7.0=pyhd8ed1ab_0
62
+ - conda-package-handling=2.2.0=pyh38be061_0
63
+ - conda-package-streaming=0.9.0=pyhd8ed1ab_0
64
+ - confection=0.1.4=py310h17c5347_0
65
+ - contourpy=1.2.0=py310hd41b1e2_0
66
+ - cryptography=41.0.7=py310hb8475ec_1
67
+ - cuda-cccl=12.4.99=0
68
+ - cuda-cudart=12.4.99=0
69
+ - cuda-cudart-dev=12.4.99=0
70
+ - cuda-nvcc-dev_linux-64=12.1.105=ha770c72_0
71
+ - cuda-nvcc-impl=12.1.105=hd3aeb46_0
72
+ - cuda-nvcc-tools=12.1.105=hd3aeb46_0
73
+ - cuda-nvrtc=12.1.105=hd3aeb46_0
74
+ - cuda-profiler-api=12.4.99=0
75
+ - cuda-python=12.4.0=py310h52dc4f0_0
76
+ - cuda-version=12.1=h1d6eff3_3
77
+ - cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
78
+ - cuml=23.08.00=cuda12_py310_230809_gd7162cdea_0
79
+ - cupy=13.0.0=py310h7aad9d2_3
80
+ - cupy-core=13.0.0=py310had4011e_3
81
+ - curl=8.6.0=hca28451_0
82
+ - cycler=0.12.1=pyhd8ed1ab_0
83
+ - cymem=2.0.8=py310hc6cd4ac_1
84
+ - cython-blis=0.7.10=py310h1f7b6fc_2
85
+ - cytoolz=0.12.3=py310h2372a71_0
86
+ - dask-cuda=23.08.00=py310_230809_gefbd6ca_0
87
+ - dask-cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
88
+ - debugpy=1.8.0=py310hc6cd4ac_1
89
+ - decorator=5.1.1=pyhd8ed1ab_0
90
+ - defusedxml=0.7.1=pyhd8ed1ab_0
91
+ - distributed=2023.7.1=pyhd8ed1ab_0
92
+ - distro=1.9.0=pyhd8ed1ab_0
93
+ - dlenv-tf-2-15-gpu=1.0.20240111=py310ha20f8e0_0
94
+ - dlpack=0.5=h9c3ff4c_0
95
+ - entrypoints=0.4=pyhd8ed1ab_0
96
+ - exceptiongroup=1.2.0=pyhd8ed1ab_2
97
+ - executing=2.0.1=pyhd8ed1ab_0
98
+ - expat=2.6.2=h59595ed_0
99
+ - fastrlock=0.8.2=py310hc6cd4ac_2
100
+ - fftw=3.3.10=nompi_hc118613_108
101
+ - fmt=9.1.0=h924138e_0
102
+ - font-ttf-dejavu-sans-mono=2.37=hab24e00_0
103
+ - font-ttf-inconsolata=3.000=h77eed37_0
104
+ - font-ttf-source-code-pro=2.038=h77eed37_0
105
+ - font-ttf-ubuntu=0.83=h77eed37_1
106
+ - fontconfig=2.14.2=h14ed4e7_0
107
+ - fonts-conda-ecosystem=1=0
108
+ - fonts-conda-forge=1=0
109
+ - fqdn=1.5.1=pyhd8ed1ab_0
110
+ - freetype=2.12.1=h267a509_2
111
+ - fribidi=1.0.10=h36c2ea0_0
112
+ - frozenlist=1.4.1=py310h2372a71_0
113
+ - fsspec=2024.3.0=pyhca7485f_0
114
+ - gdk-pixbuf=2.42.10=h829c605_5
115
+ - geos=3.11.1=h27087fc_0
116
+ - gettext=0.21.1=h27087fc_0
117
+ - gflags=2.2.2=he1b5a44_1004
118
+ - ghostscript=10.03.0=h59595ed_0
119
+ - giflib=5.2.1=h0b41bf4_3
120
+ - glog=0.6.0=h6f12383_0
121
+ - gmock=1.14.0=ha770c72_1
122
+ - gmp=6.3.0=h59595ed_0
123
+ - google-api-core-grpc=2.11.1=hd8ed1ab_0
124
+ - google-auth=2.26.1=pyhca7485f_0
125
+ - google-cloud-core=2.4.1=pyhd8ed1ab_0
126
+ - google-cloud-datastore=2.19.0=pyhd8ed1ab_0
127
+ - googleapis-common-protos=1.62.0=pyhd8ed1ab_0
128
+ - graphite2=1.3.13=h58526e2_1001
129
+ - graphviz=9.0.0=h78e8752_1
130
+ - grpc-cpp=1.51.1=h27aab58_3
131
+ - gtest=1.14.0=h00ab1b0_1
132
+ - gtk2=2.24.33=h280cfa0_4
133
+ - gts=0.7.6=h977cf35_4
134
+ - harfbuzz=8.3.0=h3d44ed6_0
135
+ - icu=73.2=h59595ed_0
136
+ - idna=3.6=pyhd8ed1ab_0
137
+ - imagemagick=7.1.1_29=pl5321hb90aeea_0
138
+ - importlib_metadata=7.0.1=hd8ed1ab_0
139
+ - importlib_resources=6.1.1=pyhd8ed1ab_0
140
+ - intel-openmp=2023.1.0=hdb19cb5_46306
141
+ - ipykernel=6.28.0=pyhd33586a_0
142
+ - ipython=8.20.0=pyh707e725_0
143
+ - ipython_genutils=0.2.0=py_1
144
+ - isoduration=20.11.0=pyhd8ed1ab_0
145
+ - jbig=2.1=h7f98852_2003
146
+ - jedi=0.19.1=pyhd8ed1ab_0
147
+ - jinja2=3.1.2=pyhd8ed1ab_1
148
+ - joblib=1.3.2=pyhd8ed1ab_0
149
+ - jsonpatch=1.33=pyhd8ed1ab_0
150
+ - jsonpointer=2.4=py310hff52083_3
151
+ - jsonschema=4.20.0=pyhd8ed1ab_0
152
+ - jsonschema-specifications=2023.12.1=pyhd8ed1ab_0
153
+ - jsonschema-with-format-nongpl=4.20.0=pyhd8ed1ab_0
154
+ - jupyter_client=8.6.0=pyhd8ed1ab_0
155
+ - jupyter_core=5.7.1=py310hff52083_0
156
+ - jupyter_events=0.9.0=pyhd8ed1ab_0
157
+ - jupyter_server_terminals=0.5.1=pyhd8ed1ab_0
158
+ - jupyterlab_pygments=0.3.0=pyhd8ed1ab_0
159
+ - keyutils=1.6.1=h166bdaf_0
160
+ - kiwisolver=1.4.5=py310hd41b1e2_1
161
+ - krb5=1.21.2=h659d440_0
162
+ - langcodes=3.3.0=pyhd8ed1ab_0
163
+ - lcms2=2.16=hb7c19ff_0
164
+ - ld_impl_linux-64=2.40=h41732ed_0
165
+ - lerc=4.0.0=h27087fc_0
166
+ - libabseil=20230125.0=cxx17_hcb278e6_1
167
+ - libarchive=3.6.2=h039dbb9_1
168
+ - libarrow=11.0.0=h33598ff_9_cpu
169
+ - libblas=3.9.0=21_linux64_openblas
170
+ - libbrotlicommon=1.0.9=h166bdaf_9
171
+ - libbrotlidec=1.0.9=h166bdaf_9
172
+ - libbrotlienc=1.0.9=h166bdaf_9
173
+ - libcblas=3.9.0=21_linux64_openblas
174
+ - libcrc32c=1.1.2=h9c3ff4c_0
175
+ - libcublas=12.1.3.1=hd3aeb46_0
176
+ - libcublas-dev=12.1.3.1=0
177
+ - libcudf=23.08.00=cuda12_230809_g8150d38e08_0
178
+ - libcufft=11.0.2.54=hd3aeb46_0
179
+ - libcufile=1.9.0.20=0
180
+ - libcufile-dev=1.9.0.20=0
181
+ - libcuml=23.08.00=cuda12_230809_gd7162cdea_0
182
+ - libcumlprims=23.08.00=cuda12_230809_g71c0a86_0
183
+ - libcurand=10.3.2.106=hd3aeb46_0
184
+ - libcurand-dev=10.3.2.106=0
185
+ - libcurl=8.6.0=hca28451_0
186
+ - libcusolver=11.4.5.107=hd3aeb46_0
187
+ - libcusolver-dev=11.4.5.107=0
188
+ - libcusparse=12.1.0.106=hd3aeb46_0
189
+ - libcusparse-dev=12.1.0.106=0
190
+ - libdeflate=1.19=hd590300_0
191
+ - libedit=3.1.20191231=he28a2e2_2
192
+ - libev=4.33=hd590300_2
193
+ - libevent=2.1.10=h28343ad_4
194
+ - libexpat=2.6.2=h59595ed_0
195
+ - libffi=3.4.2=h7f98852_5
196
+ - libgcc-ng=13.2.0=h807b86a_3
197
+ - libgd=2.3.3=h119a65a_9
198
+ - libgfortran-ng=13.2.0=h69a702a_5
199
+ - libgfortran5=13.2.0=ha4646dd_5
200
+ - libglib=2.80.0=hf2295e7_0
201
+ - libgomp=13.2.0=h807b86a_3
202
+ - libgoogle-cloud=2.8.0=h3c06191_0
203
+ - libgrpc=1.51.1=hcf146ea_3
204
+ - libhwloc=2.9.3=default_h554bfaf_1009
205
+ - libiconv=1.17=hd590300_2
206
+ - libjpeg-turbo=3.0.0=hd590300_1
207
+ - libkvikio=23.08.00=cuda12_230809_g51a9036_0
208
+ - liblapack=3.9.0=21_linux64_openblas
209
+ - libllvm14=14.0.6=hcd5def8_4
210
+ - libmamba=1.5.0=h658169a_0
211
+ - libmambapy=1.5.0=py310h8aae740_0
212
+ - libnghttp2=1.58.0=h47da74e_1
213
+ - libnsl=2.0.1=hd590300_0
214
+ - libnuma=2.0.18=hd590300_0
215
+ - libnvjitlink=12.1.105=hd3aeb46_0
216
+ - libopenblas=0.3.26=pthreads_h413a1c8_0
217
+ - libpng=1.6.43=h2797004_0
218
+ - libprotobuf=3.21.12=hfc55251_2
219
+ - libraft=23.08.00=cuda12_230809_ge588d7b5_0
220
+ - libraft-headers=23.08.00=cuda12_230809_ge588d7b5_0
221
+ - libraft-headers-only=23.08.00=cuda12_230809_ge588d7b5_0
222
+ - librmm=23.08.00=cuda12_230809_gf3af0e8d_0
223
+ - librsvg=2.56.3=he3f83f7_1
224
+ - libsodium=1.0.18=h36c2ea0_1
225
+ - libsolv=0.7.27=hfc55251_0
226
+ - libsqlite=3.44.2=h2797004_0
227
+ - libssh2=1.11.0=h0841786_0
228
+ - libstdcxx-ng=13.2.0=h7e041cc_3
229
+ - libthrift=0.18.0=h5e4af38_0
230
+ - libtiff=4.6.0=ha9c0a0a_2
231
+ - libutf8proc=2.8.0=h166bdaf_0
232
+ - libuuid=2.38.1=h0b41bf4_0
233
+ - libuv=1.46.0=hd590300_0
234
+ - libwebp=1.3.2=h658648e_1
235
+ - libwebp-base=1.3.2=hd590300_0
236
+ - libxcb=1.15=h0b41bf4_0
237
+ - libxcrypt=4.4.36=hd590300_1
238
+ - libxml2=2.12.6=h232c23b_0
239
+ - libzlib=1.2.13=hd590300_5
240
+ - llvm-openmp=8.0.1=hc9558a2_0
241
+ - locket=1.0.0=pyhd8ed1ab_0
242
+ - lz4=4.3.3=py310h350c4a5_0
243
+ - lz4-c=1.9.4=hcb278e6_0
244
+ - lzo=2.10=h516909a_1000
245
+ - magma-cuda121=2.6.1=1
246
+ - mamba=1.5.0=py310h51d5547_0
247
+ - markdown-it-py=3.0.0=pyhd8ed1ab_0
248
+ - matplotlib-base=3.8.3=py310h62c0568_0
249
+ - matplotlib-inline=0.1.6=pyhd8ed1ab_0
250
+ - mdurl=0.1.2=pyhd8ed1ab_0
251
+ - menuinst=2.0.1=py310hff52083_0
252
+ - mkl=2023.1.0=h213fc3f_46344
253
+ - msgpack-python=1.0.7=py310hd41b1e2_0
254
+ - multidict=6.0.4=py310h2372a71_1
255
+ - munkres=1.1.4=pyh9f0ad1d_0
256
+ - murmurhash=1.0.10=py310hc6cd4ac_1
257
+ - nb_conda=2.2.1=unix_7
258
+ - nb_conda_kernels=2.3.1=pyhd8ed1ab_3
259
+ - nbclassic=1.0.0=pyhb4ecaf3_1
260
+ - nbconvert-pandoc=7.14.0=pyhd8ed1ab_0
261
+ - nbformat=5.9.2=pyhd8ed1ab_0
262
+ - nccl=2.20.5.1=h3a97aeb_0
263
+ - ncurses=6.4=h59595ed_2
264
+ - nest-asyncio=1.5.8=pyhd8ed1ab_0
265
+ - nodejs=20.9.0=hb753e55_0
266
+ - notebook-shim=0.2.3=pyhd8ed1ab_0
267
+ - numpy=1.26.4=py310hb13e2d6_0
268
+ - nvcomp=2.6.1=h10b603f_3
269
+ - nvtx=0.2.10=py310h2372a71_0
270
+ - openjpeg=2.5.2=h488ebb8_0
271
+ - openmp=8.0.1=0
272
+ - openssl=3.2.1=hd590300_0
273
+ - orc=1.8.2=hfdbbad2_2
274
+ - overrides=7.4.0=pyhd8ed1ab_0
275
+ - pandoc=3.1.3=h32600fe_0
276
+ - pandocfilters=1.5.0=pyhd8ed1ab_0
277
+ - pango=1.52.1=ha41ecd1_0
278
+ - parquet-cpp=1.5.1=2
279
+ - parso=0.8.3=pyhd8ed1ab_0
280
+ - partd=1.4.1=pyhd8ed1ab_0
281
+ - pathy=0.10.3=py310h06a4308_0
282
+ - pcre2=10.43=hcad00b1_0
283
+ - perl=5.32.1=7_hd590300_perl5
284
+ - pickleshare=0.7.5=py_1003
285
+ - pip=23.3.2=pyhd8ed1ab_0
286
+ - pixman=0.43.2=h59595ed_0
287
+ - pkg-config=0.29.2=h36c2ea0_1008
288
+ - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_1
289
+ - preshed=3.0.9=py310hc6cd4ac_1
290
+ - proj=9.3.1=h1d62c97_0
291
+ - prometheus_client=0.19.0=pyhd8ed1ab_0
292
+ - proto-plus=1.23.0=pyhd8ed1ab_0
293
+ - pthread-stubs=0.4=h36c2ea0_1001
294
+ - ptyprocess=0.7.0=pyhd3deb0d_0
295
+ - pure_eval=0.2.2=pyhd8ed1ab_0
296
+ - pyarrow=11.0.0=py310h633f555_9_cpu
297
+ - pyasn1=0.5.1=pyhd8ed1ab_0
298
+ - pyasn1-modules=0.3.0=pyhd8ed1ab_0
299
+ - pybind11-abi=4=hd8ed1ab_3
300
+ - pycosat=0.6.6=py310h2372a71_0
301
+ - pygments=2.17.2=pyhd8ed1ab_0
302
+ - pylibraft=23.08.00=cuda12_py310_230809_ge588d7b5_0
303
+ - pynvml=11.4.1=pyhd8ed1ab_0
304
+ - pyopenssl=23.3.0=pyhd8ed1ab_0
305
+ - pyproj=3.6.1=py310hd5c30f3_5
306
+ - pyshp=2.3.1=pyhd8ed1ab_0
307
+ - pysocks=1.7.1=py310h06a4308_0
308
+ - python=3.10.13=hd12c33a_1_cpython
309
+ - python-fastjsonschema=2.19.1=pyhd8ed1ab_0
310
+ - python-json-logger=2.0.7=pyhd8ed1ab_0
311
+ - python_abi=3.10=4_cp310
312
+ - pyu2f=0.1.5=pyhd8ed1ab_0
313
+ - pyyaml=6.0.1=py310h2372a71_1
314
+ - raft-dask=23.08.00=cuda12_py310_230809_ge588d7b5_0
315
+ - rdma-core=28.9=h59595ed_1
316
+ - re2=2023.02.02=hcb278e6_0
317
+ - readline=8.2=h8228510_1
318
+ - referencing=0.32.1=pyhd8ed1ab_0
319
+ - reproc=14.2.4.post0=hd590300_1
320
+ - reproc-cpp=14.2.4.post0=h59595ed_1
321
+ - requests=2.31.0=pyhd8ed1ab_0
322
+ - rfc3339-validator=0.1.4=pyhd8ed1ab_0
323
+ - rfc3986-validator=0.1.1=pyh9f0ad1d_0
324
+ - rmm=23.08.00=cuda12_py310_230809_gf3af0e8d_0
325
+ - rpds-py=0.16.2=py310hcb5633a_0
326
+ - rsa=4.9=pyhd8ed1ab_0
327
+ - ruamel.yaml=0.17.40=py310h2372a71_0
328
+ - ruamel.yaml.clib=0.2.7=py310h2372a71_2
329
+ - ruamel_yaml=0.15.100=py310h7f8727e_0
330
+ - s2n=1.3.41=h3358134_0
331
+ - send2trash=1.8.2=pyh41d4057_0
332
+ - setuptools=69.0.3=pyhd8ed1ab_0
333
+ - shellingham=1.5.4=pyhd8ed1ab_0
334
+ - smart_open=6.4.0=pyhd8ed1ab_0
335
+ - snappy=1.1.10=h9fff704_0
336
+ - sniffio=1.3.0=pyhd8ed1ab_0
337
+ - sortedcontainers=2.4.0=pyhd8ed1ab_0
338
+ - soupsieve=2.5=pyhd8ed1ab_1
339
+ - spacy=3.7.2=py310hcb52e73_0
340
+ - spacy-legacy=3.0.12=pyhd8ed1ab_0
341
+ - spacy-loggers=1.0.5=pyhd8ed1ab_0
342
+ - spdlog=1.11.0=h9b3ece8_1
343
+ - sqlite=3.38.2=hc218d9a_0
344
+ - srsly=2.4.8=py310hc6cd4ac_1
345
+ - stack_data=0.6.2=pyhd8ed1ab_0
346
+ - tblib=3.0.0=pyhd8ed1ab_0
347
+ - terminado=0.18.0=pyh0d859eb_0
348
+ - thinc=8.2.2=py310hcb52e73_0
349
+ - tinycss2=1.2.1=pyhd8ed1ab_0
350
+ - tk=8.6.13=noxft_h4845f30_101
351
+ - toolz=0.12.1=pyhd8ed1ab_0
352
+ - tornado=6.3.3=py310h2372a71_1
353
+ - tqdm=4.66.1=pyhd8ed1ab_0
354
+ - traitlets=5.9.0=pyhd8ed1ab_0
355
+ - treelite=3.2.0=py310h1be96d9_0
356
+ - truststore=0.8.0=pyhd8ed1ab_0
357
+ - typer=0.9.0=pyhd8ed1ab_0
358
+ - types-python-dateutil=2.8.19.20240106=pyhd8ed1ab_0
359
+ - typing-extensions=4.9.0=hd8ed1ab_0
360
+ - typing_extensions=4.9.0=pyha770c72_0
361
+ - typing_utils=0.1.0=pyhd8ed1ab_0
362
+ - ucx=1.14.1=h195a15c_5
363
+ - ucx-proc=1.0.0=gpu
364
+ - ucx-py=0.33.00=py310_230809_gea1eb8f_0
365
+ - unicodedata2=15.1.0=py310h2372a71_0
366
+ - uri-template=1.3.0=pyhd8ed1ab_0
367
+ - wasabi=1.1.2=py310hff52083_0
368
+ - wcwidth=0.2.13=pyhd8ed1ab_0
369
+ - weasel=0.3.4=pyhd8ed1ab_0
370
+ - webcolors=1.13=pyhd8ed1ab_0
371
+ - webencodings=0.5.1=pyhd8ed1ab_2
372
+ - websocket-client=1.7.0=pyhd8ed1ab_0
373
+ - wheel=0.42.0=pyhd8ed1ab_0
374
+ - xorg-kbproto=1.0.7=h7f98852_1002
375
+ - xorg-libice=1.1.1=hd590300_0
376
+ - xorg-libsm=1.2.4=h7391055_0
377
+ - xorg-libx11=1.8.7=h8ee46fc_0
378
+ - xorg-libxau=1.0.11=hd590300_0
379
+ - xorg-libxdmcp=1.1.3=h7f98852_0
380
+ - xorg-libxext=1.3.4=h0b41bf4_2
381
+ - xorg-libxrender=0.9.11=hd590300_0
382
+ - xorg-libxt=1.3.0=hd590300_1
383
+ - xorg-renderproto=0.11.1=h7f98852_1002
384
+ - xorg-xextproto=7.3.0=h0b41bf4_1003
385
+ - xorg-xproto=7.0.31=h7f98852_1007
386
+ - xyzservices=2023.10.1=pyhd8ed1ab_0
387
+ - xz=5.2.6=h166bdaf_0
388
+ - yaml=0.2.5=h7b6447c_0
389
+ - yaml-cpp=0.7.0=h59595ed_3
390
+ - zeromq=4.3.5=h59595ed_0
391
+ - zict=3.0.0=pyhd8ed1ab_0
392
+ - zipp=3.17.0=pyhd8ed1ab_0
393
+ - zlib=1.2.13=hd590300_5
394
+ - zstandard=0.22.0=py310h1275a96_0
395
+ - zstd=1.5.5=hfc55251_0
396
+ - pip:
397
+ - absl-py==1.4.0
398
+ - accelerate==0.28.0
399
+ - access==1.1.9
400
+ - affine==2.4.0
401
+ - aiobotocore==2.12.1
402
+ - aiofiles==22.1.0
403
+ - aiohttp-cors==0.7.0
404
+ - aioitertools==0.11.0
405
+ - aiorwlock==1.3.0
406
+ - aiosqlite==0.19.0
407
+ - albumentations==1.4.0
408
+ - alembic==1.13.1
409
+ - altair==5.2.0
410
+ - annoy==1.17.3
411
+ - apache-beam==2.46.0
412
+ - aplus==0.11.0
413
+ - appdirs==1.4.4
414
+ - array-record==0.5.0
415
+ - arviz==0.17.1
416
+ - astroid==3.0.3
417
+ - astropy==6.0.0
418
+ - astropy-iers-data==0.2024.3.18.0.29.47
419
+ - astunparse==1.6.3
420
+ - async-lru==2.0.4
421
+ - audioread==3.0.1
422
+ - auto-gptq==0.7.1
423
+ - autopep8==2.0.4
424
+ - babel==2.14.0
425
+ - backoff==2.2.1
426
+ - bayesian-optimization==1.4.3
427
+ - beatrix-jupyterlab==2023.128.151533
428
+ - bidict==0.23.1
429
+ - bitsandbytes==0.43.0
430
+ - blake3==0.2.1
431
+ - blessed==1.20.0
432
+ - blinker==1.7.0
433
+ - blosc2==2.5.1
434
+ - bokeh==3.3.4
435
+ - boruta==0.3
436
+ - boto3==1.26.100
437
+ - botocore==1.34.51
438
+ - bqplot==0.12.43
439
+ - branca==0.7.1
440
+ - brewer2mpl==1.4.1
441
+ - cachetools==4.2.4
442
+ - catalyst==22.4
443
+ - catboost==1.2.3
444
+ - category-encoders==2.6.3
445
+ - cesium==0.12.1
446
+ - chex==0.1.85
447
+ - cleverhans==4.0.0
448
+ - click-plugins==1.1.1
449
+ - cligj==0.7.2
450
+ - cloud-tpu-client==0.10
451
+ - cloud-tpu-profiler==2.4.0
452
+ - cloudpickle==2.2.1
453
+ - cmdstanpy==1.2.1
454
+ - cmudict==1.0.21
455
+ - colorcet==3.1.0
456
+ - coloredlogs==15.0.1
457
+ - colorful==0.5.6
458
+ - colorlog==6.8.2
459
+ - colorlover==0.3.0
460
+ - contextily==1.5.2
461
+ - convertdate==2.4.0
462
+ - cramjam==2.8.3
463
+ - crcmod==1.7
464
+ - cufflinks==0.17.3
465
+ - cvxcanon==0.1.2
466
+ - cython==3.0.8
467
+ - daal==2024.1.0
468
+ - daal4py==2024.1.0
469
+ - dacite==1.8.1
470
+ - dask==2024.3.1
471
+ - dask-expr==1.0.4
472
+ - dataclasses-json==0.6.4
473
+ - dataproc-jupyter-plugin==0.1.66
474
+ - datasets==2.1.0
475
+ - datashader==0.16.0
476
+ - datatile==1.0.3
477
+ - db-dtypes==1.2.0
478
+ - deap==1.4.1
479
+ - deepdiff==6.7.1
480
+ - deprecated==1.2.14
481
+ - deprecation==2.1.0
482
+ - descartes==1.1.0
483
+ - dill==0.3.8
484
+ - dipy==1.9.0
485
+ - distlib==0.3.8
486
+ - dm-tree==0.1.8
487
+ - docker==7.0.0
488
+ - docker-pycreds==0.4.0
489
+ - docopt==0.6.2
490
+ - docstring-parser==0.15
491
+ - docstring-to-markdown==0.15
492
+ - docutils==0.20.1
493
+ - earthengine-api==0.1.394
494
+ - easydict==1.13
495
+ - easyocr==1.7.1
496
+ - ecos==2.0.13
497
+ - eli5==0.13.0
498
+ - emoji==2.10.1
499
+ - en-core-web-lg==3.7.1
500
+ - en-core-web-sm==3.7.1
501
+ - ephem==4.1.5
502
+ - esda==2.5.1
503
+ - essentia==2.1b6.dev1110
504
+ - et-xmlfile==1.1.0
505
+ - etils==1.6.0
506
+ - explainable-ai-sdk==1.3.3
507
+ - farama-notifications==0.0.4
508
+ - fastai==2.7.14
509
+ - fastapi==0.108.0
510
+ - fastavro==1.9.3
511
+ - fastcore==1.5.29
512
+ - fastdownload==0.0.7
513
+ - fasteners==0.19
514
+ - fastparquet==2024.2.0
515
+ - fastprogress==1.0.3
516
+ - fasttext==0.9.2
517
+ - feather-format==0.4.1
518
+ - featuretools==1.30.0
519
+ - filelock==3.13.1
520
+ - fiona==1.9.6
521
+ - fitter==1.7.0
522
+ - flake8==7.0.0
523
+ - flashtext==2.7
524
+ - flask==3.0.2
525
+ - flatbuffers==23.5.26
526
+ - flax==0.8.2
527
+ - folium==0.16.0
528
+ - fonttools==4.47.0
529
+ - frozendict==2.4.0
530
+ - funcy==2.0
531
+ - fury==0.10.0
532
+ - future==1.0.0
533
+ - fuzzywuzzy==0.18.0
534
+ - gast==0.5.4
535
+ - gatspy==0.3
536
+ - gcsfs==2023.12.2.post1
537
+ - gekko==1.1.0
538
+ - gensim==4.3.2
539
+ - geographiclib==2.0
540
+ - geohash==1.0
541
+ - geojson==3.1.0
542
+ - geopandas==0.14.3
543
+ - geoplot==0.5.1
544
+ - geopy==2.4.1
545
+ - geoviews==1.11.1
546
+ - ggplot==0.11.5
547
+ - giddy==2.3.5
548
+ - gitdb==4.0.11
549
+ - gitpython==3.1.41
550
+ - google-ai-generativelanguage==0.4.0
551
+ - google-api-core==2.17.1
552
+ - google-api-python-client==2.122.0
553
+ - google-apitools==0.5.31
554
+ - google-auth-httplib2==0.1.1
555
+ - google-auth-oauthlib==1.2.0
556
+ - google-cloud-aiplatform==0.6.0a1
557
+ - google-cloud-artifact-registry==1.10.0
558
+ - google-cloud-automl==1.0.1
559
+ - google-cloud-bigquery==2.34.4
560
+ - google-cloud-bigtable==1.7.3
561
+ - google-cloud-dlp==3.14.0
562
+ - google-cloud-jupyter-config==0.0.5
563
+ - google-cloud-language==2.13.3
564
+ - google-cloud-monitoring==2.18.0
565
+ - google-cloud-pubsub==2.19.0
566
+ - google-cloud-pubsublite==1.9.0
567
+ - google-cloud-recommendations-ai==0.7.1
568
+ - google-cloud-resource-manager==1.11.0
569
+ - google-cloud-spanner==3.40.1
570
+ - google-cloud-storage==1.44.0
571
+ - google-cloud-translate==3.12.1
572
+ - google-cloud-videointelligence==2.13.3
573
+ - google-cloud-vision==2.8.0
574
+ - google-crc32c==1.5.0
575
+ - google-generativeai==0.4.1
576
+ - google-pasta==0.2.0
577
+ - google-resumable-media==2.7.0
578
+ - gplearn==0.4.2
579
+ - gpustat==1.0.0
580
+ - gpxpy==1.6.2
581
+ - greenlet==3.0.3
582
+ - grpc-google-iam-v1==0.12.7
583
+ - grpcio==1.60.0
584
+ - grpcio-status==1.48.2
585
+ - gviz-api==1.10.0
586
+ - gym==0.26.2
587
+ - gym-notices==0.0.8
588
+ - gymnasium==0.29.0
589
+ - h11==0.14.0
590
+ - h2o==3.46.0.1
591
+ - h5netcdf==1.3.0
592
+ - h5py==3.10.0
593
+ - haversine==2.8.1
594
+ - hdfs==2.7.3
595
+ - hep-ml==0.7.2
596
+ - hijri-converter==2.3.1
597
+ - hmmlearn==0.3.2
598
+ - holidays==0.24
599
+ - holoviews==1.18.3
600
+ - hpsklearn==0.1.0
601
+ - html5lib==1.1
602
+ - htmlmin==0.1.12
603
+ - httpcore==1.0.4
604
+ - httplib2==0.21.0
605
+ - httptools==0.6.1
606
+ - httpx==0.27.0
607
+ - huggingface-hub==0.21.4
608
+ - humanfriendly==10.0
609
+ - hunspell==0.5.5
610
+ - husl==4.0.3
611
+ - hydra-slayer==0.5.0
612
+ - hyperopt==0.2.7
613
+ - hypertools==0.8.0
614
+ - igraph==0.11.4
615
+ - imagecodecs==2024.1.1
616
+ - imagehash==4.3.1
617
+ - imageio==2.33.1
618
+ - imbalanced-learn==0.12.0
619
+ - imgaug==0.4.0
620
+ - importlib-metadata==6.11.0
621
+ - inequality==1.0.1
622
+ - iniconfig==2.0.0
623
+ - ipydatawidgets==4.3.5
624
+ - ipyleaflet==0.18.2
625
+ - ipympl==0.7.0
626
+ - ipython-genutils==0.2.0
627
+ - ipython-sql==0.5.0
628
+ - ipyvolume==0.6.3
629
+ - ipyvue==1.10.2
630
+ - ipyvuetify==1.9.2
631
+ - ipywebrtc==0.6.0
632
+ - ipywidgets==7.7.1
633
+ - isort==5.13.2
634
+ - isoweek==1.3.3
635
+ - itsdangerous==2.1.2
636
+ - janome==0.5.0
637
+ - jaraco-classes==3.3.0
638
+ - jax==0.4.23
639
+ - jax-jumpy==1.0.0
640
+ - jaxlib==0.4.23.dev20240116
641
+ - jeepney==0.8.0
642
+ - jieba==0.42.1
643
+ - jmespath==1.0.1
644
+ - json5==0.9.14
645
+ - jupyter-client==7.4.9
646
+ - jupyter-console==6.6.3
647
+ - jupyter-http-over-ws==0.0.8
648
+ - jupyter-lsp==1.5.1
649
+ - jupyter-server==2.13.0
650
+ - jupyter-server-fileid==0.9.1
651
+ - jupyter-server-mathjax==0.2.6
652
+ - jupyter-server-proxy==4.1.0
653
+ - jupyter-server-ydoc==0.8.0
654
+ - jupyter-ydoc==0.2.5
655
+ - jupyterlab==4.1.5
656
+ - jupyterlab-git==0.44.0
657
+ - jupyterlab-lsp==5.1.0
658
+ - jupyterlab-server==2.25.2
659
+ - jupyterlab-widgets==3.0.9
660
+ - jupytext==1.16.0
661
+ - kaggle==1.6.6
662
+ - kaggle-environments==1.14.3
663
+ - kagglehub==0.2.0
664
+ - keras==3.0.5
665
+ - keras-cv==0.8.2
666
+ - keras-nlp==0.8.2
667
+ - keras-tuner==1.4.6
668
+ - kernels-mixer==0.0.7
669
+ - keyring==24.3.0
670
+ - keyrings-google-artifactregistry-auth==1.1.2
671
+ - kfp==2.5.0
672
+ - kfp-pipeline-spec==0.2.2
673
+ - kfp-server-api==2.0.5
674
+ - kmapper==2.0.1
675
+ - kmodes==0.12.2
676
+ - korean-lunar-calendar==0.3.1
677
+ - kornia==0.7.2
678
+ - kornia-rs==0.1.2
679
+ - kt-legacy==1.0.5
680
+ - kubernetes==26.1.0
681
+ - langid==1.1.6
682
+ - lazy-loader==0.3
683
+ - learntools==0.3.4
684
+ - leven==1.0.4
685
+ - levenshtein==0.25.0
686
+ - libclang==16.0.6
687
+ - libpysal==4.9.2
688
+ - librosa==0.10.1
689
+ - lightgbm==4.2.0
690
+ - lightning-utilities==0.10.1
691
+ - lime==0.2.0.1
692
+ - line-profiler==4.1.2
693
+ - linkify-it-py==2.0.3
694
+ - llvmlite==0.41.1
695
+ - lml==0.1.0
696
+ - loguru==0.7.2
697
+ - lunarcalendar==0.0.9
698
+ - lxml==5.1.0
699
+ - mako==1.3.2
700
+ - mapclassify==2.6.1
701
+ - markdown==3.5.2
702
+ - markovify==0.9.4
703
+ - markupsafe==2.1.5
704
+ - marshmallow==3.21.1
705
+ - matplotlib==3.7.5
706
+ - matplotlib-venn==0.11.10
707
+ - mccabe==0.7.0
708
+ - mdit-py-plugins==0.4.0
709
+ - memory-profiler==0.61.0
710
+ - mercantile==1.2.1
711
+ - mgwr==2.2.1
712
+ - missingno==0.5.2
713
+ - mistune==0.8.4
714
+ - mizani==0.11.0
715
+ - ml-dtypes==0.2.0
716
+ - mlcrate==0.2.0
717
+ - mlens==0.2.3
718
+ - mlxtend==0.23.1
719
+ - mmh3==4.1.0
720
+ - mne==1.6.1
721
+ - mnist==0.2.2
722
+ - mock==5.1.0
723
+ - momepy==0.7.0
724
+ - more-itertools==10.2.0
725
+ - mpld3==0.5.10
726
+ - mpmath==1.3.0
727
+ - msgpack-numpy==0.4.8
728
+ - multimethod==1.10
729
+ - multipledispatch==1.0.0
730
+ - multiprocess==0.70.16
731
+ - mypy-extensions==1.0.0
732
+ - namex==0.0.7
733
+ - nbclient==0.5.13
734
+ - nbconvert==6.4.5
735
+ - nbdime==3.2.0
736
+ - ndindex==1.8
737
+ - networkx==3.2.1
738
+ - nibabel==5.2.1
739
+ - nilearn==0.10.3
740
+ - ninja==1.11.1.1
741
+ - nltk==3.2.4
742
+ - nose==1.3.7
743
+ - notebook==6.5.6
744
+ - notebook-executor==0.2
745
+ - numba==0.58.1
746
+ - numexpr==2.9.0
747
+ - nvidia-ml-py==11.495.46
748
+ - oauth2client==4.1.3
749
+ - oauthlib==3.2.2
750
+ - objsize==0.6.1
751
+ - odfpy==1.4.1
752
+ - olefile==0.47
753
+ - onnx==1.15.0
754
+ - opencensus==0.11.4
755
+ - opencensus-context==0.1.3
756
+ - opencv-contrib-python==4.9.0.80
757
+ - opencv-python==4.9.0.80
758
+ - opencv-python-headless==4.9.0.80
759
+ - openpyxl==3.1.2
760
+ - openslide-python==1.3.1
761
+ - opentelemetry-api==1.22.0
762
+ - opentelemetry-exporter-otlp==1.22.0
763
+ - opentelemetry-exporter-otlp-proto-common==1.22.0
764
+ - opentelemetry-exporter-otlp-proto-grpc==1.22.0
765
+ - opentelemetry-exporter-otlp-proto-http==1.22.0
766
+ - opentelemetry-proto==1.22.0
767
+ - opentelemetry-sdk==1.22.0
768
+ - opentelemetry-semantic-conventions==0.43b0
769
+ - opt-einsum==3.3.0
770
+ - optax==0.2.1
771
+ - optimum==1.18.0
772
+ - optuna==3.6.0
773
+ - orbax-checkpoint==0.5.6
774
+ - ordered-set==4.1.0
775
+ - orjson==3.9.10
776
+ - ortools==9.4.1874
777
+ - osmnx==1.9.1
778
+ - packaging==21.3
779
+ - pandas==2.2.1
780
+ - pandas-datareader==0.10.0
781
+ - pandas-profiling==3.6.6
782
+ - pandas-summary==0.2.0
783
+ - pandasql==0.7.3
784
+ - panel==1.3.8
785
+ - papermill==2.5.0
786
+ - param==2.0.2
787
+ - path==16.10.0
788
+ - path-py==12.5.0
789
+ - pathos==0.3.2
790
+ - patsy==0.5.6
791
+ - pdf2image==1.17.0
792
+ - peft==0.10.0
793
+ - pettingzoo==1.24.0
794
+ - pexpect==4.9.0
795
+ - phik==0.12.4
796
+ - pillow==9.5.0
797
+ - platformdirs==4.2.0
798
+ - plotly==5.18.0
799
+ - plotly-express==0.4.1
800
+ - plotnine==0.13.2
801
+ - pluggy==1.4.0
802
+ - pointpats==2.4.0
803
+ - polars==0.20.15
804
+ - polyglot==16.7.4
805
+ - pooch==1.8.1
806
+ - pox==0.3.4
807
+ - ppca==0.0.4
808
+ - ppft==1.7.6.8
809
+ - preprocessing==0.1.13
810
+ - prettytable==3.9.0
811
+ - progressbar2==4.4.2
812
+ - promise==2.3
813
+ - prompt-toolkit==3.0.43
814
+ - pronouncing==0.2.0
815
+ - prophet==1.1.1
816
+ - protobuf==3.20.3
817
+ - psutil==5.9.3
818
+ - pudb==2024.1
819
+ - pulp==2.8.0
820
+ - py-cpuinfo==9.0.0
821
+ - py-spy==0.3.14
822
+ - py4j==0.10.9.7
823
+ - pyaml==23.12.0
824
+ - pyarabic==0.6.15
825
+ - pyastronomy==0.21.0
826
+ - pybind11==2.11.1
827
+ - pyclipper==1.3.0.post5
828
+ - pycodestyle==2.11.1
829
+ - pycparser==2.21
830
+ - pycryptodome==3.20.0
831
+ - pyct==0.5.0
832
+ - pycuda==2024.1
833
+ - pydantic==2.5.3
834
+ - pydantic-core==2.14.6
835
+ - pydegensac==0.1.2
836
+ - pydicom==2.4.4
837
+ - pydocstyle==6.3.0
838
+ - pydot==1.4.2
839
+ - pydub==0.25.1
840
+ - pyemd==1.0.0
841
+ - pyerfa==2.0.1.1
842
+ - pyexcel-io==0.6.6
843
+ - pyexcel-ods==0.6.0
844
+ - pyflakes==3.2.0
845
+ - pygltflib==1.16.2
846
+ - pyjwt==2.8.0
847
+ - pykalman==0.9.5
848
+ - pyldavis==3.4.1
849
+ - pylint==3.0.4
850
+ - pymc3==3.11.4
851
+ - pymeeus==0.5.12
852
+ - pymongo==3.13.0
853
+ - pympler==1.0.1
854
+ - pynndescent==0.5.11
855
+ - pynvrtc==9.2
856
+ - pyocr==0.8.5
857
+ - pyparsing==3.1.1
858
+ - pypdf==4.1.0
859
+ - pysal==24.1
860
+ - pytesseract==0.3.10
861
+ - pytest==8.1.1
862
+ - python-bidi==0.4.2
863
+ - python-dateutil==2.9.0.post0
864
+ - python-dotenv==1.0.0
865
+ - python-graphviz==0.20.2
866
+ - python-levenshtein==0.25.0
867
+ - python-louvain==0.16
868
+ - python-lsp-jsonrpc==1.1.2
869
+ - python-lsp-server==1.10.1
870
+ - python-slugify==8.0.4
871
+ - python-utils==3.8.2
872
+ - pythreejs==2.4.2
873
+ - pytoolconfig==1.3.1
874
+ - pytools==2023.1.1
875
+ - pytorch-ignite==0.4.13
876
+ - pytorch-lightning==2.2.1
877
+ - pytz==2023.3.post1
878
+ - pyupset==0.1.1.post7
879
+ - pyviz-comms==3.0.1
880
+ - pywavelets==1.5.0
881
+ - pyzmq==24.0.1
882
+ - qgrid==1.3.1
883
+ - qtconsole==5.5.1
884
+ - qtpy==2.4.1
885
+ - quantecon==0.7.2
886
+ - quantities==0.15.0
887
+ - qudida==0.0.4
888
+ - rapidfuzz==3.6.2
889
+ - rasterio==1.3.9
890
+ - rasterstats==0.19.0
891
+ - ray==2.9.0
892
+ - ray-cpp==2.9.0
893
+ - regex==2023.12.25
894
+ - requests-oauthlib==1.3.1
895
+ - requests-toolbelt==0.10.1
896
+ - responses==0.18.0
897
+ - retrying==1.3.4
898
+ - rgf-python==3.12.0
899
+ - rich==13.7.0
900
+ - rich-click==1.7.4
901
+ - rope==1.12.0
902
+ - rouge==1.0.1
903
+ - rtree==1.2.0
904
+ - s2sphere==0.2.5
905
+ - s3fs==2024.3.0
906
+ - s3transfer==0.6.2
907
+ - safetensors==0.4.2
908
+ - scattertext==0.1.19
909
+ - scikit-image==0.22.0
910
+ - scikit-learn==1.2.2
911
+ - scikit-learn-intelex==2024.1.0
912
+ - scikit-multilearn==0.2.0
913
+ - scikit-optimize==0.10.1
914
+ - scikit-plot==0.3.7
915
+ - scikit-surprise==1.1.3
916
+ - scipy==1.11.4
917
+ - seaborn==0.12.2
918
+ - secretstorage==3.3.3
919
+ - segment-anything==1.0
920
+ - segregation==2.5
921
+ - semver==3.0.2
922
+ - sentencepiece==0.2.0
923
+ - sentry-sdk==1.42.0
924
+ - setproctitle==1.3.3
925
+ - setuptools-git==1.2
926
+ - setuptools-scm==8.0.4
927
+ - shap==0.44.1
928
+ - shapely==2.0.3
929
+ - shimmy==1.3.0
930
+ - simpervisor==1.0.0
931
+ - simpleitk==2.3.1
932
+ - simplejson==3.19.2
933
+ - six==1.16.0
934
+ - sklearn-pandas==2.2.0
935
+ - slicer==0.0.7
936
+ - smmap==5.0.1
937
+ - snowballstemmer==2.2.0
938
+ - snuggs==1.4.7
939
+ - soundfile==0.12.1
940
+ - soxr==0.3.7
941
+ - spaghetti==1.7.5.post1
942
+ - spectral==0.23.1
943
+ - spglm==1.1.0
944
+ - sphinx-rtd-theme==0.2.4
945
+ - spint==1.0.7
946
+ - splot==1.1.5.post1
947
+ - spopt==0.6.0
948
+ - spreg==1.4.2
949
+ - spvcm==0.3.0
950
+ - sqlalchemy==2.0.25
951
+ - sqlparse==0.4.4
952
+ - squarify==0.4.3
953
+ - stable-baselines3==2.1.0
954
+ - stack-data==0.6.3
955
+ - stanio==0.3.0
956
+ - starlette==0.32.0.post1
957
+ - statsmodels==0.14.1
958
+ - stemming==1.0.1
959
+ - stop-words==2018.7.23
960
+ - stopit==1.1.2
961
+ - stumpy==1.12.0
962
+ - sympy==1.12
963
+ - tables==3.9.2
964
+ - tabulate==0.9.0
965
+ - tangled-up-in-unicode==0.2.0
966
+ - tbb==2021.11.0
967
+ - tenacity==8.2.3
968
+ - tensorboard==2.15.1
969
+ - tensorboard-data-server==0.7.2
970
+ - tensorboard-plugin-profile==2.15.0
971
+ - tensorboardx==2.6.2.2
972
+ - tensorflow==2.15.0
973
+ - tensorflow-cloud==0.1.16
974
+ - tensorflow-datasets==4.9.4
975
+ - tensorflow-decision-forests==1.8.1
976
+ - tensorflow-estimator==2.15.0
977
+ - tensorflow-hub==0.16.1
978
+ - tensorflow-io==0.35.0
979
+ - tensorflow-io-gcs-filesystem==0.35.0
980
+ - tensorflow-metadata==0.14.0
981
+ - tensorflow-probability==0.23.0
982
+ - tensorflow-serving-api==2.14.1
983
+ - tensorflow-text==2.15.0
984
+ - tensorflow-transform==0.14.0
985
+ - tensorpack==0.11
986
+ - tensorstore==0.1.56
987
+ - termcolor==2.4.0
988
+ - testpath==0.6.0
989
+ - text-unidecode==1.3
990
+ - textblob==0.18.0.post0
991
+ - texttable==1.7.0
992
+ - tf-keras==2.15.1
993
+ - tfp-nightly==0.24.0.dev0
994
+ - theano==1.0.5
995
+ - theano-pymc==1.1.2
996
+ - threadpoolctl==3.2.0
997
+ - tifffile==2023.12.9
998
+ - timm==0.9.16
999
+ - tobler==0.11.2
1000
+ - tokenizers==0.15.2
1001
+ - toml==0.10.2
1002
+ - tomli==2.0.1
1003
+ - tomlkit==0.12.4
1004
+ - torch==2.1.2
1005
+ - torchaudio==2.1.2
1006
+ - torchdata==0.7.1
1007
+ - torchinfo==1.8.0
1008
+ - torchmetrics==1.3.2
1009
+ - torchtext==0.16.2
1010
+ - torchvision==0.16.2
1011
+ - tpot==0.12.1
1012
+ - traceml==1.0.8
1013
+ - traittypes==0.2.1
1014
+ - transformers==4.38.2
1015
+ - treelite-runtime==3.2.0
1016
+ - trueskill==0.4.5
1017
+ - trx-python==0.2.9
1018
+ - tsfresh==0.20.2
1019
+ - typeguard==4.1.5
1020
+ - typing-inspect==0.9.0
1021
+ - tzdata==2023.4
1022
+ - uc-micro-py==1.0.3
1023
+ - ujson==5.9.0
1024
+ - umap-learn==0.5.5
1025
+ - unidecode==1.3.8
1026
+ - update-checker==0.18.0
1027
+ - uritemplate==3.0.1
1028
+ - urllib3==1.26.18
1029
+ - urwid==2.6.9
1030
+ - urwid-readline==0.14
1031
+ - uvicorn==0.25.0
1032
+ - uvloop==0.19.0
1033
+ - vaex==4.17.0
1034
+ - vaex-astro==0.9.3
1035
+ - vaex-core==4.17.1
1036
+ - vaex-hdf5==0.14.1
1037
+ - vaex-jupyter==0.8.2
1038
+ - vaex-ml==0.18.3
1039
+ - vaex-server==0.9.0
1040
+ - vaex-viz==0.5.4
1041
+ - vec-noise==1.1.4
1042
+ - vecstack==0.4.0
1043
+ - virtualenv==20.21.0
1044
+ - visions==0.7.5
1045
+ - vowpalwabbit==9.9.0
1046
+ - vtk==9.3.0
1047
+ - wand==0.6.13
1048
+ - wandb==0.16.4
1049
+ - watchfiles==0.21.0
1050
+ - wavio==0.0.8
1051
+ - websockets==12.0
1052
+ - werkzeug==3.0.1
1053
+ - wfdb==4.1.2
1054
+ - whatthepatch==1.0.5
1055
+ - widgetsnbextension==3.6.6
1056
+ - witwidget==1.8.1
1057
+ - woodwork==0.29.0
1058
+ - wordcloud==1.9.3
1059
+ - wordsegment==1.3.1
1060
+ - wrapt==1.14.1
1061
+ - xarray==2024.2.0
1062
+ - xarray-einstats==0.7.0
1063
+ - xgboost==2.0.3
1064
+ - xvfbwrapper==0.2.9
1065
+ - xxhash==3.4.1
1066
+ - y-py==0.6.2
1067
+ - yapf==0.40.2
1068
+ - yarl==1.9.4
1069
+ - ydata-profiling==4.6.4
1070
+ - yellowbrick==1.5
1071
+ - ypy-websocket==0.8.4
1072
+ prefix: /opt/conda
wandb/run-20240402_072503-b95w935d/files/config.yaml ADDED
@@ -0,0 +1,724 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.4
8
+ framework: huggingface
9
+ huggingface_version: 4.38.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712042703.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ - 99
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 98
42
+ - 99
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.4
49
+ 6: 4.38.2
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ - 1: eval/loss
78
+ 5: 1
79
+ 6:
80
+ - 1
81
+ - 1: eval/runtime
82
+ 5: 1
83
+ 6:
84
+ - 1
85
+ - 1: eval/samples_per_second
86
+ 5: 1
87
+ 6:
88
+ - 1
89
+ - 1: eval/steps_per_second
90
+ 5: 1
91
+ 6:
92
+ - 1
93
+ - 1: train/train_runtime
94
+ 5: 1
95
+ 6:
96
+ - 1
97
+ - 1: train/train_samples_per_second
98
+ 5: 1
99
+ 6:
100
+ - 1
101
+ - 1: train/train_steps_per_second
102
+ 5: 1
103
+ 6:
104
+ - 1
105
+ - 1: train/total_flos
106
+ 5: 1
107
+ 6:
108
+ - 1
109
+ - 1: train/train_loss
110
+ 5: 1
111
+ 6:
112
+ - 1
113
+ vocab_size:
114
+ desc: null
115
+ value: 32000
116
+ max_position_embeddings:
117
+ desc: null
118
+ value: 32768
119
+ hidden_size:
120
+ desc: null
121
+ value: 4096
122
+ intermediate_size:
123
+ desc: null
124
+ value: 14336
125
+ num_hidden_layers:
126
+ desc: null
127
+ value: 32
128
+ num_attention_heads:
129
+ desc: null
130
+ value: 32
131
+ sliding_window:
132
+ desc: null
133
+ value: null
134
+ num_key_value_heads:
135
+ desc: null
136
+ value: 8
137
+ hidden_act:
138
+ desc: null
139
+ value: silu
140
+ initializer_range:
141
+ desc: null
142
+ value: 0.02
143
+ rms_norm_eps:
144
+ desc: null
145
+ value: 1.0e-05
146
+ use_cache:
147
+ desc: null
148
+ value: false
149
+ rope_theta:
150
+ desc: null
151
+ value: 1000000.0
152
+ attention_dropout:
153
+ desc: null
154
+ value: 0.0
155
+ return_dict:
156
+ desc: null
157
+ value: true
158
+ output_hidden_states:
159
+ desc: null
160
+ value: false
161
+ output_attentions:
162
+ desc: null
163
+ value: false
164
+ torchscript:
165
+ desc: null
166
+ value: false
167
+ torch_dtype:
168
+ desc: null
169
+ value: bfloat16
170
+ use_bfloat16:
171
+ desc: null
172
+ value: false
173
+ tf_legacy_loss:
174
+ desc: null
175
+ value: false
176
+ pruned_heads:
177
+ desc: null
178
+ value: {}
179
+ tie_word_embeddings:
180
+ desc: null
181
+ value: false
182
+ chunk_size_feed_forward:
183
+ desc: null
184
+ value: 0
185
+ is_encoder_decoder:
186
+ desc: null
187
+ value: false
188
+ is_decoder:
189
+ desc: null
190
+ value: false
191
+ cross_attention_hidden_size:
192
+ desc: null
193
+ value: null
194
+ add_cross_attention:
195
+ desc: null
196
+ value: false
197
+ tie_encoder_decoder:
198
+ desc: null
199
+ value: false
200
+ max_length:
201
+ desc: null
202
+ value: 20
203
+ min_length:
204
+ desc: null
205
+ value: 0
206
+ do_sample:
207
+ desc: null
208
+ value: false
209
+ early_stopping:
210
+ desc: null
211
+ value: false
212
+ num_beams:
213
+ desc: null
214
+ value: 1
215
+ num_beam_groups:
216
+ desc: null
217
+ value: 1
218
+ diversity_penalty:
219
+ desc: null
220
+ value: 0.0
221
+ temperature:
222
+ desc: null
223
+ value: 1.0
224
+ top_k:
225
+ desc: null
226
+ value: 50
227
+ top_p:
228
+ desc: null
229
+ value: 1.0
230
+ typical_p:
231
+ desc: null
232
+ value: 1.0
233
+ repetition_penalty:
234
+ desc: null
235
+ value: 1.0
236
+ length_penalty:
237
+ desc: null
238
+ value: 1.0
239
+ no_repeat_ngram_size:
240
+ desc: null
241
+ value: 0
242
+ encoder_no_repeat_ngram_size:
243
+ desc: null
244
+ value: 0
245
+ bad_words_ids:
246
+ desc: null
247
+ value: null
248
+ num_return_sequences:
249
+ desc: null
250
+ value: 1
251
+ output_scores:
252
+ desc: null
253
+ value: false
254
+ return_dict_in_generate:
255
+ desc: null
256
+ value: false
257
+ forced_bos_token_id:
258
+ desc: null
259
+ value: null
260
+ forced_eos_token_id:
261
+ desc: null
262
+ value: null
263
+ remove_invalid_values:
264
+ desc: null
265
+ value: false
266
+ exponential_decay_length_penalty:
267
+ desc: null
268
+ value: null
269
+ suppress_tokens:
270
+ desc: null
271
+ value: null
272
+ begin_suppress_tokens:
273
+ desc: null
274
+ value: null
275
+ architectures:
276
+ desc: null
277
+ value:
278
+ - MistralForCausalLM
279
+ finetuning_task:
280
+ desc: null
281
+ value: null
282
+ id2label:
283
+ desc: null
284
+ value:
285
+ '0': LABEL_0
286
+ '1': LABEL_1
287
+ label2id:
288
+ desc: null
289
+ value:
290
+ LABEL_0: 0
291
+ LABEL_1: 1
292
+ tokenizer_class:
293
+ desc: null
294
+ value: null
295
+ prefix:
296
+ desc: null
297
+ value: null
298
+ bos_token_id:
299
+ desc: null
300
+ value: 1
301
+ pad_token_id:
302
+ desc: null
303
+ value: 0
304
+ eos_token_id:
305
+ desc: null
306
+ value: 2
307
+ sep_token_id:
308
+ desc: null
309
+ value: null
310
+ decoder_start_token_id:
311
+ desc: null
312
+ value: null
313
+ task_specific_params:
314
+ desc: null
315
+ value: null
316
+ problem_type:
317
+ desc: null
318
+ value: null
319
+ _name_or_path:
320
+ desc: null
321
+ value: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
322
+ transformers_version:
323
+ desc: null
324
+ value: 4.38.2
325
+ model_type:
326
+ desc: null
327
+ value: mistral
328
+ pretraining_tp:
329
+ desc: null
330
+ value: 1
331
+ quantization_config:
332
+ desc: null
333
+ value:
334
+ quant_method: QuantizationMethod.GPTQ
335
+ bits: 4
336
+ tokenizer: null
337
+ dataset: null
338
+ group_size: 128
339
+ damp_percent: 0.1
340
+ desc_act: true
341
+ sym: true
342
+ true_sequential: true
343
+ use_cuda_fp16: false
344
+ model_seqlen: null
345
+ block_name_to_quantize: null
346
+ module_name_preceding_first_block: null
347
+ batch_size: 1
348
+ pad_token_id: null
349
+ use_exllama: true
350
+ max_input_length: null
351
+ exllama_config:
352
+ version: ExllamaVersion.ONE
353
+ cache_block_outputs: true
354
+ modules_in_block_to_quantize: null
355
+ output_dir:
356
+ desc: null
357
+ value: /kaggle/working/
358
+ overwrite_output_dir:
359
+ desc: null
360
+ value: false
361
+ do_train:
362
+ desc: null
363
+ value: false
364
+ do_eval:
365
+ desc: null
366
+ value: true
367
+ do_predict:
368
+ desc: null
369
+ value: false
370
+ evaluation_strategy:
371
+ desc: null
372
+ value: epoch
373
+ prediction_loss_only:
374
+ desc: null
375
+ value: false
376
+ per_device_train_batch_size:
377
+ desc: null
378
+ value: 6
379
+ per_device_eval_batch_size:
380
+ desc: null
381
+ value: 6
382
+ per_gpu_train_batch_size:
383
+ desc: null
384
+ value: null
385
+ per_gpu_eval_batch_size:
386
+ desc: null
387
+ value: null
388
+ gradient_accumulation_steps:
389
+ desc: null
390
+ value: 4
391
+ eval_accumulation_steps:
392
+ desc: null
393
+ value: null
394
+ eval_delay:
395
+ desc: null
396
+ value: 0
397
+ learning_rate:
398
+ desc: null
399
+ value: 0.0002
400
+ weight_decay:
401
+ desc: null
402
+ value: 0.01
403
+ adam_beta1:
404
+ desc: null
405
+ value: 0.9
406
+ adam_beta2:
407
+ desc: null
408
+ value: 0.999
409
+ adam_epsilon:
410
+ desc: null
411
+ value: 1.0e-08
412
+ max_grad_norm:
413
+ desc: null
414
+ value: 1.0
415
+ num_train_epochs:
416
+ desc: null
417
+ value: 30
418
+ max_steps:
419
+ desc: null
420
+ value: -1
421
+ lr_scheduler_type:
422
+ desc: null
423
+ value: linear
424
+ lr_scheduler_kwargs:
425
+ desc: null
426
+ value: {}
427
+ warmup_ratio:
428
+ desc: null
429
+ value: 0.0
430
+ warmup_steps:
431
+ desc: null
432
+ value: 2
433
+ log_level:
434
+ desc: null
435
+ value: passive
436
+ log_level_replica:
437
+ desc: null
438
+ value: warning
439
+ log_on_each_node:
440
+ desc: null
441
+ value: true
442
+ logging_dir:
443
+ desc: null
444
+ value: /kaggle/working/runs/Apr02_07-24-52_0ecfc053b35f
445
+ logging_strategy:
446
+ desc: null
447
+ value: epoch
448
+ logging_first_step:
449
+ desc: null
450
+ value: false
451
+ logging_steps:
452
+ desc: null
453
+ value: 500
454
+ logging_nan_inf_filter:
455
+ desc: null
456
+ value: true
457
+ save_strategy:
458
+ desc: null
459
+ value: epoch
460
+ save_steps:
461
+ desc: null
462
+ value: 500
463
+ save_total_limit:
464
+ desc: null
465
+ value: null
466
+ save_safetensors:
467
+ desc: null
468
+ value: true
469
+ save_on_each_node:
470
+ desc: null
471
+ value: false
472
+ save_only_model:
473
+ desc: null
474
+ value: false
475
+ no_cuda:
476
+ desc: null
477
+ value: false
478
+ use_cpu:
479
+ desc: null
480
+ value: false
481
+ use_mps_device:
482
+ desc: null
483
+ value: false
484
+ seed:
485
+ desc: null
486
+ value: 42
487
+ data_seed:
488
+ desc: null
489
+ value: null
490
+ jit_mode_eval:
491
+ desc: null
492
+ value: false
493
+ use_ipex:
494
+ desc: null
495
+ value: false
496
+ bf16:
497
+ desc: null
498
+ value: false
499
+ fp16:
500
+ desc: null
501
+ value: true
502
+ fp16_opt_level:
503
+ desc: null
504
+ value: O1
505
+ half_precision_backend:
506
+ desc: null
507
+ value: auto
508
+ bf16_full_eval:
509
+ desc: null
510
+ value: false
511
+ fp16_full_eval:
512
+ desc: null
513
+ value: false
514
+ tf32:
515
+ desc: null
516
+ value: null
517
+ local_rank:
518
+ desc: null
519
+ value: 0
520
+ ddp_backend:
521
+ desc: null
522
+ value: null
523
+ tpu_num_cores:
524
+ desc: null
525
+ value: null
526
+ tpu_metrics_debug:
527
+ desc: null
528
+ value: false
529
+ debug:
530
+ desc: null
531
+ value: []
532
+ dataloader_drop_last:
533
+ desc: null
534
+ value: false
535
+ eval_steps:
536
+ desc: null
537
+ value: null
538
+ dataloader_num_workers:
539
+ desc: null
540
+ value: 0
541
+ dataloader_prefetch_factor:
542
+ desc: null
543
+ value: null
544
+ past_index:
545
+ desc: null
546
+ value: -1
547
+ run_name:
548
+ desc: null
549
+ value: /kaggle/working/
550
+ disable_tqdm:
551
+ desc: null
552
+ value: false
553
+ remove_unused_columns:
554
+ desc: null
555
+ value: true
556
+ label_names:
557
+ desc: null
558
+ value: null
559
+ load_best_model_at_end:
560
+ desc: null
561
+ value: true
562
+ metric_for_best_model:
563
+ desc: null
564
+ value: loss
565
+ greater_is_better:
566
+ desc: null
567
+ value: false
568
+ ignore_data_skip:
569
+ desc: null
570
+ value: false
571
+ fsdp:
572
+ desc: null
573
+ value: []
574
+ fsdp_min_num_params:
575
+ desc: null
576
+ value: 0
577
+ fsdp_config:
578
+ desc: null
579
+ value:
580
+ min_num_params: 0
581
+ xla: false
582
+ xla_fsdp_v2: false
583
+ xla_fsdp_grad_ckpt: false
584
+ fsdp_transformer_layer_cls_to_wrap:
585
+ desc: null
586
+ value: null
587
+ accelerator_config:
588
+ desc: null
589
+ value:
590
+ split_batches: false
591
+ dispatch_batches: null
592
+ even_batches: true
593
+ use_seedable_sampler: true
594
+ deepspeed:
595
+ desc: null
596
+ value: null
597
+ label_smoothing_factor:
598
+ desc: null
599
+ value: 0.0
600
+ optim:
601
+ desc: null
602
+ value: paged_adamw_8bit
603
+ optim_args:
604
+ desc: null
605
+ value: null
606
+ adafactor:
607
+ desc: null
608
+ value: false
609
+ group_by_length:
610
+ desc: null
611
+ value: false
612
+ length_column_name:
613
+ desc: null
614
+ value: length
615
+ report_to:
616
+ desc: null
617
+ value:
618
+ - tensorboard
619
+ - wandb
620
+ ddp_find_unused_parameters:
621
+ desc: null
622
+ value: null
623
+ ddp_bucket_cap_mb:
624
+ desc: null
625
+ value: null
626
+ ddp_broadcast_buffers:
627
+ desc: null
628
+ value: null
629
+ dataloader_pin_memory:
630
+ desc: null
631
+ value: true
632
+ dataloader_persistent_workers:
633
+ desc: null
634
+ value: false
635
+ skip_memory_metrics:
636
+ desc: null
637
+ value: true
638
+ use_legacy_prediction_loop:
639
+ desc: null
640
+ value: false
641
+ push_to_hub:
642
+ desc: null
643
+ value: false
644
+ resume_from_checkpoint:
645
+ desc: null
646
+ value: null
647
+ hub_model_id:
648
+ desc: null
649
+ value: null
650
+ hub_strategy:
651
+ desc: null
652
+ value: every_save
653
+ hub_token:
654
+ desc: null
655
+ value: <HUB_TOKEN>
656
+ hub_private_repo:
657
+ desc: null
658
+ value: false
659
+ hub_always_push:
660
+ desc: null
661
+ value: false
662
+ gradient_checkpointing:
663
+ desc: null
664
+ value: false
665
+ gradient_checkpointing_kwargs:
666
+ desc: null
667
+ value: null
668
+ include_inputs_for_metrics:
669
+ desc: null
670
+ value: false
671
+ fp16_backend:
672
+ desc: null
673
+ value: auto
674
+ push_to_hub_model_id:
675
+ desc: null
676
+ value: null
677
+ push_to_hub_organization:
678
+ desc: null
679
+ value: null
680
+ push_to_hub_token:
681
+ desc: null
682
+ value: <PUSH_TO_HUB_TOKEN>
683
+ mp_parameters:
684
+ desc: null
685
+ value: ''
686
+ auto_find_batch_size:
687
+ desc: null
688
+ value: false
689
+ full_determinism:
690
+ desc: null
691
+ value: false
692
+ torchdynamo:
693
+ desc: null
694
+ value: null
695
+ ray_scope:
696
+ desc: null
697
+ value: last
698
+ ddp_timeout:
699
+ desc: null
700
+ value: 1800
701
+ torch_compile:
702
+ desc: null
703
+ value: false
704
+ torch_compile_backend:
705
+ desc: null
706
+ value: null
707
+ torch_compile_mode:
708
+ desc: null
709
+ value: null
710
+ dispatch_batches:
711
+ desc: null
712
+ value: null
713
+ split_batches:
714
+ desc: null
715
+ value: null
716
+ include_tokens_per_second:
717
+ desc: null
718
+ value: false
719
+ include_num_input_tokens_seen:
720
+ desc: null
721
+ value: false
722
+ neftune_noise_alpha:
723
+ desc: null
724
+ value: null
wandb/run-20240402_072503-b95w935d/files/output.log ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
2
+ warnings.warn(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
6
+ warnings.warn(
7
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
8
+ warnings.warn(
9
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
10
+ warnings.warn(
11
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
12
+ warnings.warn(
13
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
14
+ warnings.warn(
15
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
16
+ warnings.warn(
17
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
18
+ warnings.warn(
19
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
20
+ warnings.warn(
21
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
22
+ warnings.warn(
23
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
24
+ warnings.warn(
25
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
26
+ warnings.warn(
27
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
28
+ warnings.warn(
29
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
30
+ warnings.warn(
31
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
32
+ warnings.warn(
33
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
34
+ warnings.warn(
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
38
+ warnings.warn(
39
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
40
+ warnings.warn(
41
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
42
+ warnings.warn(
43
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
44
+ warnings.warn(
45
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
46
+ warnings.warn(
47
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
48
+ warnings.warn(
49
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
50
+ warnings.warn(
51
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
52
+ warnings.warn(
53
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
54
+ warnings.warn(
55
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
56
+ warnings.warn(
57
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
wandb/run-20240402_072503-b95w935d/files/requirements.txt ADDED
@@ -0,0 +1,882 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ Levenshtein==0.25.0
16
+ LunarCalendar==0.0.9
17
+ Mako==1.3.2
18
+ Markdown==3.5.2
19
+ MarkupSafe==2.1.3
20
+ MarkupSafe==2.1.5
21
+ Pillow==9.5.0
22
+ PuLP==2.8.0
23
+ PyArabic==0.6.15
24
+ PyAstronomy==0.21.0
25
+ PyJWT==2.8.0
26
+ PyMeeus==0.5.12
27
+ PySocks==1.7.1
28
+ PyUpSet==0.1.1.post7
29
+ PyWavelets==1.5.0
30
+ PyYAML==6.0.1
31
+ Pygments==2.17.2
32
+ Pympler==1.0.1
33
+ QtPy==2.4.1
34
+ Rtree==1.2.0
35
+ SQLAlchemy==2.0.25
36
+ SecretStorage==3.3.3
37
+ Send2Trash==1.8.2
38
+ Shapely==1.8.5.post1
39
+ Shimmy==1.3.0
40
+ SimpleITK==2.3.1
41
+ TPOT==0.12.1
42
+ Theano-PyMC==1.1.2
43
+ Theano==1.0.5
44
+ Unidecode==1.3.8
45
+ Wand==0.6.13
46
+ Werkzeug==3.0.1
47
+ absl-py==1.4.0
48
+ accelerate==0.28.0
49
+ access==1.1.9
50
+ affine==2.4.0
51
+ aiobotocore==2.12.1
52
+ aiofiles==22.1.0
53
+ aiohttp-cors==0.7.0
54
+ aiohttp==3.9.1
55
+ aioitertools==0.11.0
56
+ aiorwlock==1.3.0
57
+ aiosignal==1.3.1
58
+ aiosqlite==0.19.0
59
+ albumentations==1.4.0
60
+ alembic==1.13.1
61
+ altair==5.2.0
62
+ annotated-types==0.6.0
63
+ annoy==1.17.3
64
+ anyio==4.2.0
65
+ apache-beam==2.46.0
66
+ aplus==0.11.0
67
+ appdirs==1.4.4
68
+ archspec==0.2.2
69
+ argon2-cffi-bindings==21.2.0
70
+ argon2-cffi==23.1.0
71
+ array-record==0.5.0
72
+ arrow==1.3.0
73
+ arviz==0.17.1
74
+ astroid==3.0.3
75
+ astropy-iers-data==0.2024.3.18.0.29.47
76
+ astropy==6.0.0
77
+ asttokens==2.4.1
78
+ astunparse==1.6.3
79
+ async-lru==2.0.4
80
+ async-timeout==4.0.3
81
+ attrs==23.2.0
82
+ audioread==3.0.1
83
+ auto_gptq==0.7.1
84
+ autopep8==2.0.4
85
+ backoff==2.2.1
86
+ bayesian-optimization==1.4.3
87
+ beatrix_jupyterlab==2023.128.151533
88
+ beautifulsoup4==4.12.2
89
+ bidict==0.23.1
90
+ bitsandbytes==0.43.0
91
+ blake3==0.2.1
92
+ bleach==6.1.0
93
+ blessed==1.20.0
94
+ blinker==1.7.0
95
+ blis==0.7.10
96
+ blosc2==2.5.1
97
+ bokeh==3.3.4
98
+ boltons==23.1.1
99
+ boto3==1.26.100
100
+ botocore==1.34.51
101
+ bq_helper==0.4.1
102
+ bqplot==0.12.43
103
+ branca==0.7.1
104
+ brewer2mpl==1.4.1
105
+ brotlipy==0.7.0
106
+ cached-property==1.5.2
107
+ cachetools==4.2.4
108
+ cachetools==5.3.2
109
+ catalogue==2.0.10
110
+ catalyst==22.4
111
+ catboost==1.2.3
112
+ category-encoders==2.6.3
113
+ certifi==2024.2.2
114
+ cesium==0.12.1
115
+ cffi==1.16.0
116
+ charset-normalizer==3.3.2
117
+ chex==0.1.85
118
+ cleverhans==4.0.0
119
+ click-plugins==1.1.1
120
+ click==8.1.7
121
+ cligj==0.7.2
122
+ cloud-tpu-client==0.10
123
+ cloud-tpu-profiler==2.4.0
124
+ cloudpathlib==0.16.0
125
+ cloudpickle==2.2.1
126
+ cloudpickle==3.0.0
127
+ cmdstanpy==1.2.1
128
+ cmudict==1.0.21
129
+ colorama==0.4.6
130
+ colorcet==3.1.0
131
+ coloredlogs==15.0.1
132
+ colorful==0.5.6
133
+ colorlog==6.8.2
134
+ colorlover==0.3.0
135
+ comm==0.2.1
136
+ conda-libmamba-solver==23.7.0
137
+ conda-package-handling==2.2.0
138
+ conda==23.7.4
139
+ conda_package_streaming==0.9.0
140
+ confection==0.1.4
141
+ contextily==1.5.2
142
+ contourpy==1.2.0
143
+ convertdate==2.4.0
144
+ cramjam==2.8.3
145
+ crcmod==1.7
146
+ cryptography==41.0.7
147
+ cuda-python==12.4.0
148
+ cudf==23.8.0
149
+ cufflinks==0.17.3
150
+ cuml==23.8.0
151
+ cupy==13.0.0
152
+ cycler==0.12.1
153
+ cymem==2.0.8
154
+ cytoolz==0.12.3
155
+ daal4py==2024.1.0
156
+ daal==2024.1.0
157
+ dacite==1.8.1
158
+ dask-cuda==23.8.0
159
+ dask-cudf==23.8.0
160
+ dask-expr==1.0.4
161
+ dask==2024.3.1
162
+ dataclasses-json==0.6.4
163
+ dataproc_jupyter_plugin==0.1.66
164
+ datasets==2.1.0
165
+ datashader==0.16.0
166
+ datatile==1.0.3
167
+ db-dtypes==1.2.0
168
+ deap==1.4.1
169
+ debugpy==1.8.0
170
+ decorator==5.1.1
171
+ deepdiff==6.7.1
172
+ defusedxml==0.7.1
173
+ deprecation==2.1.0
174
+ descartes==1.1.0
175
+ dill==0.3.8
176
+ dipy==1.9.0
177
+ distlib==0.3.8
178
+ distributed==2023.7.1
179
+ distro==1.9.0
180
+ dm-tree==0.1.8
181
+ docker-pycreds==0.4.0
182
+ docker==7.0.0
183
+ docopt==0.6.2
184
+ docstring-parser==0.15
185
+ docstring-to-markdown==0.15
186
+ docutils==0.20.1
187
+ earthengine-api==0.1.394
188
+ easydict==1.13
189
+ easyocr==1.7.1
190
+ ecos==2.0.13
191
+ eli5==0.13.0
192
+ emoji==2.10.1
193
+ en-core-web-lg==3.7.1
194
+ en-core-web-sm==3.7.1
195
+ entrypoints==0.4
196
+ ephem==4.1.5
197
+ esda==2.5.1
198
+ essentia==2.1b6.dev1110
199
+ et-xmlfile==1.1.0
200
+ etils==1.6.0
201
+ exceptiongroup==1.2.0
202
+ executing==2.0.1
203
+ explainable-ai-sdk==1.3.3
204
+ fastai==2.7.14
205
+ fastapi==0.108.0
206
+ fastavro==1.9.3
207
+ fastcore==1.5.29
208
+ fastdownload==0.0.7
209
+ fasteners==0.19
210
+ fastjsonschema==2.19.1
211
+ fastparquet==2024.2.0
212
+ fastprogress==1.0.3
213
+ fastrlock==0.8.2
214
+ fasttext==0.9.2
215
+ feather-format==0.4.1
216
+ featuretools==1.30.0
217
+ filelock==3.13.1
218
+ fiona==1.9.6
219
+ fitter==1.7.0
220
+ flake8==7.0.0
221
+ flashtext==2.7
222
+ flatbuffers==23.5.26
223
+ flax==0.8.2
224
+ folium==0.16.0
225
+ fonttools==4.47.0
226
+ fonttools==4.49.0
227
+ fqdn==1.5.1
228
+ frozendict==2.4.0
229
+ frozenlist==1.4.1
230
+ fsspec==2024.3.0
231
+ funcy==2.0
232
+ fury==0.10.0
233
+ future==1.0.0
234
+ fuzzywuzzy==0.18.0
235
+ gast==0.5.4
236
+ gatspy==0.3
237
+ gcsfs==2023.12.2.post1
238
+ gekko==1.1.0
239
+ gensim==4.3.2
240
+ geographiclib==2.0
241
+ geojson==3.1.0
242
+ geopandas==0.14.3
243
+ geoplot==0.5.1
244
+ geopy==2.4.1
245
+ geoviews==1.11.1
246
+ ggplot==0.11.5
247
+ giddy==2.3.5
248
+ gitdb==4.0.11
249
+ google-ai-generativelanguage==0.4.0
250
+ google-api-core==2.11.1
251
+ google-api-core==2.17.1
252
+ google-api-python-client==2.122.0
253
+ google-apitools==0.5.31
254
+ google-auth-httplib2==0.1.1
255
+ google-auth-oauthlib==1.2.0
256
+ google-auth==2.26.1
257
+ google-cloud-aiplatform==0.6.0a1
258
+ google-cloud-artifact-registry==1.10.0
259
+ google-cloud-automl==1.0.1
260
+ google-cloud-bigquery==2.34.4
261
+ google-cloud-bigtable==1.7.3
262
+ google-cloud-core==2.4.1
263
+ google-cloud-datastore==2.19.0
264
+ google-cloud-dlp==3.14.0
265
+ google-cloud-jupyter-config==0.0.5
266
+ google-cloud-language==2.13.3
267
+ google-cloud-monitoring==2.18.0
268
+ google-cloud-pubsub==2.19.0
269
+ google-cloud-pubsublite==1.9.0
270
+ google-cloud-recommendations-ai==0.7.1
271
+ google-cloud-resource-manager==1.11.0
272
+ google-cloud-spanner==3.40.1
273
+ google-cloud-storage==1.44.0
274
+ google-cloud-translate==3.12.1
275
+ google-cloud-videointelligence==2.13.3
276
+ google-cloud-vision==2.8.0
277
+ google-crc32c==1.5.0
278
+ google-generativeai==0.4.1
279
+ google-pasta==0.2.0
280
+ google-resumable-media==2.7.0
281
+ googleapis-common-protos==1.62.0
282
+ gplearn==0.4.2
283
+ gpustat==1.0.0
284
+ gpxpy==1.6.2
285
+ graphviz==0.20.2
286
+ greenlet==3.0.3
287
+ grpc-google-iam-v1==0.12.7
288
+ grpcio-status==1.48.1
289
+ grpcio-status==1.48.2
290
+ grpcio==1.51.1
291
+ grpcio==1.60.0
292
+ gviz-api==1.10.0
293
+ gym-notices==0.0.8
294
+ gym==0.26.2
295
+ gymnasium==0.29.0
296
+ h11==0.14.0
297
+ h2o==3.46.0.1
298
+ h5netcdf==1.3.0
299
+ h5py==3.10.0
300
+ haversine==2.8.1
301
+ hdfs==2.7.3
302
+ hep-ml==0.7.2
303
+ hijri-converter==2.3.1
304
+ hmmlearn==0.3.2
305
+ holidays==0.24
306
+ holoviews==1.18.3
307
+ hpsklearn==0.1.0
308
+ html5lib==1.1
309
+ htmlmin==0.1.12
310
+ httpcore==1.0.4
311
+ httplib2==0.21.0
312
+ httptools==0.6.1
313
+ httpx==0.27.0
314
+ huggingface-hub==0.21.4
315
+ humanfriendly==10.0
316
+ hunspell==0.5.5
317
+ husl==4.0.3
318
+ hydra-slayer==0.5.0
319
+ hyperopt==0.2.7
320
+ hypertools==0.8.0
321
+ idna==3.6
322
+ igraph==0.11.4
323
+ imagecodecs==2024.1.1
324
+ imageio==2.33.1
325
+ imbalanced-learn==0.12.0
326
+ imgaug==0.4.0
327
+ importlib-metadata==6.11.0
328
+ importlib-metadata==7.0.1
329
+ importlib-resources==6.1.1
330
+ inequality==1.0.1
331
+ iniconfig==2.0.0
332
+ ipydatawidgets==4.3.5
333
+ ipykernel==6.28.0
334
+ ipyleaflet==0.18.2
335
+ ipympl==0.7.0
336
+ ipython-genutils==0.2.0
337
+ ipython-genutils==0.2.0
338
+ ipython-sql==0.5.0
339
+ ipython==8.20.0
340
+ ipyvolume==0.6.3
341
+ ipyvue==1.10.2
342
+ ipyvuetify==1.9.2
343
+ ipywebrtc==0.6.0
344
+ ipywidgets==7.7.1
345
+ isoduration==20.11.0
346
+ isort==5.13.2
347
+ isoweek==1.3.3
348
+ itsdangerous==2.1.2
349
+ jaraco.classes==3.3.0
350
+ jax-jumpy==1.0.0
351
+ jax==0.4.23
352
+ jaxlib==0.4.23.dev20240116
353
+ jedi==0.19.1
354
+ jeepney==0.8.0
355
+ jieba==0.42.1
356
+ jmespath==1.0.1
357
+ joblib==1.3.2
358
+ json5==0.9.14
359
+ jsonpatch==1.33
360
+ jsonpointer==2.4
361
+ jsonschema-specifications==2023.12.1
362
+ jsonschema==4.20.0
363
+ jupyter-console==6.6.3
364
+ jupyter-events==0.9.0
365
+ jupyter-http-over-ws==0.0.8
366
+ jupyter-lsp==1.5.1
367
+ jupyter-server-mathjax==0.2.6
368
+ jupyter-ydoc==0.2.5
369
+ jupyter_client==7.4.9
370
+ jupyter_client==8.6.0
371
+ jupyter_core==5.7.1
372
+ jupyter_server==2.13.0
373
+ jupyter_server_fileid==0.9.1
374
+ jupyter_server_proxy==4.1.0
375
+ jupyter_server_terminals==0.5.1
376
+ jupyter_server_ydoc==0.8.0
377
+ jupyterlab-lsp==5.1.0
378
+ jupyterlab-widgets==3.0.9
379
+ jupyterlab==4.1.5
380
+ jupyterlab_git==0.44.0
381
+ jupyterlab_pygments==0.3.0
382
+ jupyterlab_server==2.25.2
383
+ jupytext==1.16.0
384
+ kaggle-environments==1.14.3
385
+ kaggle==1.6.6
386
+ kagglehub==0.2.0
387
+ keras-cv==0.8.2
388
+ keras-nlp==0.8.2
389
+ keras-tuner==1.4.6
390
+ keras==3.0.5
391
+ kernels-mixer==0.0.7
392
+ keyring==24.3.0
393
+ keyrings.google-artifactregistry-auth==1.1.2
394
+ kfp-pipeline-spec==0.2.2
395
+ kfp-server-api==2.0.5
396
+ kfp==2.5.0
397
+ kiwisolver==1.4.5
398
+ kmapper==2.0.1
399
+ kmodes==0.12.2
400
+ korean-lunar-calendar==0.3.1
401
+ kornia==0.7.2
402
+ kornia_rs==0.1.2
403
+ kt-legacy==1.0.5
404
+ kubernetes==26.1.0
405
+ langcodes==3.3.0
406
+ langid==1.1.6
407
+ lazy_loader==0.3
408
+ learntools==0.3.4
409
+ leven==1.0.4
410
+ libclang==16.0.6
411
+ libmambapy==1.5.0
412
+ libpysal==4.9.2
413
+ librosa==0.10.1
414
+ lightgbm==4.2.0
415
+ lightning-utilities==0.10.1
416
+ lime==0.2.0.1
417
+ line-profiler==4.1.2
418
+ linkify-it-py==2.0.3
419
+ llvmlite==0.41.1
420
+ llvmlite==0.42.0
421
+ lml==0.1.0
422
+ locket==1.0.0
423
+ loguru==0.7.2
424
+ lxml==5.1.0
425
+ lz4==4.3.3
426
+ mamba==1.5.0
427
+ mapclassify==2.6.1
428
+ markdown-it-py==3.0.0
429
+ markovify==0.9.4
430
+ marshmallow==3.21.1
431
+ matplotlib-inline==0.1.6
432
+ matplotlib-venn==0.11.10
433
+ matplotlib==3.7.5
434
+ matplotlib==3.8.3
435
+ mccabe==0.7.0
436
+ mdit-py-plugins==0.4.0
437
+ mdurl==0.1.2
438
+ memory-profiler==0.61.0
439
+ menuinst==2.0.1
440
+ mercantile==1.2.1
441
+ mgwr==2.2.1
442
+ missingno==0.5.2
443
+ mistune==0.8.4
444
+ mizani==0.11.0
445
+ ml-dtypes==0.2.0
446
+ mlcrate==0.2.0
447
+ mlens==0.2.3
448
+ mlxtend==0.23.1
449
+ mmh3==4.1.0
450
+ mne==1.6.1
451
+ mnist==0.2.2
452
+ mock==5.1.0
453
+ momepy==0.7.0
454
+ more-itertools==10.2.0
455
+ mpld3==0.5.10
456
+ mpmath==1.3.0
457
+ msgpack-numpy==0.4.8
458
+ msgpack==1.0.7
459
+ multidict==6.0.4
460
+ multimethod==1.10
461
+ multipledispatch==1.0.0
462
+ multiprocess==0.70.16
463
+ munkres==1.1.4
464
+ murmurhash==1.0.10
465
+ mypy-extensions==1.0.0
466
+ namex==0.0.7
467
+ nb-conda-kernels==2.3.1
468
+ nb_conda==2.2.1
469
+ nbclassic==1.0.0
470
+ nbclient==0.5.13
471
+ nbconvert==6.4.5
472
+ nbdime==3.2.0
473
+ nbformat==5.9.2
474
+ ndindex==1.8
475
+ nest-asyncio==1.5.8
476
+ networkx==3.2.1
477
+ nibabel==5.2.1
478
+ nilearn==0.10.3
479
+ ninja==1.11.1.1
480
+ nltk==3.2.4
481
+ nose==1.3.7
482
+ notebook==6.5.4
483
+ notebook==6.5.6
484
+ notebook_executor==0.2
485
+ notebook_shim==0.2.3
486
+ numba==0.58.1
487
+ numba==0.59.0
488
+ numexpr==2.9.0
489
+ numpy==1.26.4
490
+ nvidia-ml-py==11.495.46
491
+ nvtx==0.2.10
492
+ oauth2client==4.1.3
493
+ oauthlib==3.2.2
494
+ objsize==0.6.1
495
+ odfpy==1.4.1
496
+ olefile==0.47
497
+ onnx==1.15.0
498
+ opencensus-context==0.1.3
499
+ opencensus==0.11.4
500
+ opencv-contrib-python==4.9.0.80
501
+ opencv-python-headless==4.9.0.80
502
+ opencv-python==4.9.0.80
503
+ openpyxl==3.1.2
504
+ openslide-python==1.3.1
505
+ opentelemetry-api==1.22.0
506
+ opentelemetry-exporter-otlp-proto-common==1.22.0
507
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
508
+ opentelemetry-exporter-otlp-proto-http==1.22.0
509
+ opentelemetry-exporter-otlp==1.22.0
510
+ opentelemetry-proto==1.22.0
511
+ opentelemetry-sdk==1.22.0
512
+ opentelemetry-semantic-conventions==0.43b0
513
+ opt-einsum==3.3.0
514
+ optax==0.2.1
515
+ optimum==1.18.0
516
+ optuna==3.6.0
517
+ orbax-checkpoint==0.5.6
518
+ ordered-set==4.1.0
519
+ orjson==3.9.10
520
+ ortools==9.4.1874
521
+ osmnx==1.9.1
522
+ overrides==7.4.0
523
+ packaging==21.3
524
+ pandas-datareader==0.10.0
525
+ pandas-profiling==3.6.6
526
+ pandas-summary==0.2.0
527
+ pandas==2.1.4
528
+ pandas==2.2.1
529
+ pandasql==0.7.3
530
+ pandocfilters==1.5.0
531
+ panel==1.3.8
532
+ papermill==2.5.0
533
+ param==2.0.2
534
+ parso==0.8.3
535
+ partd==1.4.1
536
+ path.py==12.5.0
537
+ path==16.10.0
538
+ pathos==0.3.2
539
+ pathy==0.10.3
540
+ patsy==0.5.6
541
+ pdf2image==1.17.0
542
+ peft==0.10.0
543
+ pettingzoo==1.24.0
544
+ pexpect==4.8.0
545
+ pexpect==4.9.0
546
+ phik==0.12.4
547
+ pickleshare==0.7.5
548
+ pip==23.3.2
549
+ pkgutil_resolve_name==1.3.10
550
+ platformdirs==4.2.0
551
+ plotly-express==0.4.1
552
+ plotly==5.18.0
553
+ plotnine==0.13.2
554
+ pluggy==1.4.0
555
+ pointpats==2.4.0
556
+ polars==0.20.15
557
+ polyglot==16.7.4
558
+ pooch==1.8.1
559
+ pox==0.3.4
560
+ ppca==0.0.4
561
+ ppft==1.7.6.8
562
+ preprocessing==0.1.13
563
+ preshed==3.0.9
564
+ prettytable==3.9.0
565
+ progressbar2==4.4.2
566
+ prometheus-client==0.19.0
567
+ promise==2.3
568
+ prompt-toolkit==3.0.42
569
+ prompt-toolkit==3.0.43
570
+ pronouncing==0.2.0
571
+ prophet==1.1.1
572
+ proto-plus==1.23.0
573
+ protobuf==3.20.3
574
+ protobuf==4.21.12
575
+ psutil==5.9.3
576
+ psutil==5.9.7
577
+ ptyprocess==0.7.0
578
+ pudb==2024.1
579
+ pure-eval==0.2.2
580
+ py-cpuinfo==9.0.0
581
+ py-spy==0.3.14
582
+ py4j==0.10.9.7
583
+ pyLDAvis==3.4.1
584
+ pyOpenSSL==23.3.0
585
+ pyaml==23.12.0
586
+ pyarrow==11.0.0
587
+ pyasn1-modules==0.3.0
588
+ pyasn1==0.5.1
589
+ pybind11==2.11.1
590
+ pyclipper==1.3.0.post5
591
+ pycodestyle==2.11.1
592
+ pycosat==0.6.6
593
+ pycparser==2.21
594
+ pycryptodome==3.20.0
595
+ pyct==0.5.0
596
+ pycuda==2024.1
597
+ pydantic==2.5.3
598
+ pydantic==2.6.4
599
+ pydantic_core==2.14.6
600
+ pydantic_core==2.16.3
601
+ pydegensac==0.1.2
602
+ pydicom==2.4.4
603
+ pydocstyle==6.3.0
604
+ pydot==1.4.2
605
+ pydub==0.25.1
606
+ pyemd==1.0.0
607
+ pyerfa==2.0.1.1
608
+ pyexcel-io==0.6.6
609
+ pyexcel-ods==0.6.0
610
+ pyflakes==3.2.0
611
+ pygltflib==1.16.2
612
+ pykalman==0.9.5
613
+ pylibraft==23.8.0
614
+ pylint==3.0.4
615
+ pymc3==3.11.4
616
+ pymongo==3.13.0
617
+ pynndescent==0.5.11
618
+ pynvml==11.4.1
619
+ pynvrtc==9.2
620
+ pyocr==0.8.5
621
+ pyparsing==3.1.1
622
+ pyparsing==3.1.2
623
+ pypdf==4.1.0
624
+ pyproj==3.6.1
625
+ pysal==24.1
626
+ pyshp==2.3.1
627
+ pytesseract==0.3.10
628
+ pytest==8.1.1
629
+ python-Levenshtein==0.25.0
630
+ python-bidi==0.4.2
631
+ python-dateutil==2.9.0.post0
632
+ python-dotenv==1.0.0
633
+ python-json-logger==2.0.7
634
+ python-louvain==0.16
635
+ python-lsp-jsonrpc==1.1.2
636
+ python-lsp-server==1.10.1
637
+ python-slugify==8.0.4
638
+ python-utils==3.8.2
639
+ pythreejs==2.4.2
640
+ pytoolconfig==1.3.1
641
+ pytools==2023.1.1
642
+ pytorch-ignite==0.4.13
643
+ pytorch-lightning==2.2.1
644
+ pytz==2023.3.post1
645
+ pytz==2024.1
646
+ pyu2f==0.1.5
647
+ pyviz_comms==3.0.1
648
+ pyzmq==24.0.1
649
+ pyzmq==25.1.2
650
+ qgrid==1.3.1
651
+ qtconsole==5.5.1
652
+ quantecon==0.7.2
653
+ quantities==0.15.0
654
+ qudida==0.0.4
655
+ raft-dask==23.8.0
656
+ rapidfuzz==3.6.2
657
+ rasterio==1.3.9
658
+ rasterstats==0.19.0
659
+ ray-cpp==2.9.0
660
+ ray==2.9.0
661
+ referencing==0.32.1
662
+ regex==2023.12.25
663
+ requests-oauthlib==1.3.1
664
+ requests-toolbelt==0.10.1
665
+ requests==2.31.0
666
+ responses==0.18.0
667
+ retrying==1.3.3
668
+ retrying==1.3.4
669
+ rfc3339-validator==0.1.4
670
+ rfc3986-validator==0.1.1
671
+ rgf-python==3.12.0
672
+ rich-click==1.7.4
673
+ rich==13.7.0
674
+ rich==13.7.1
675
+ rmm==23.8.0
676
+ rope==1.12.0
677
+ rouge==1.0.1
678
+ rpds-py==0.16.2
679
+ rsa==4.9
680
+ ruamel-yaml-conda==0.15.100
681
+ ruamel.yaml.clib==0.2.7
682
+ ruamel.yaml==0.17.40
683
+ s2sphere==0.2.5
684
+ s3fs==2024.3.0
685
+ s3transfer==0.6.2
686
+ safetensors==0.4.2
687
+ scattertext==0.1.19
688
+ scikit-image==0.22.0
689
+ scikit-learn-intelex==2024.1.0
690
+ scikit-learn==1.2.2
691
+ scikit-multilearn==0.2.0
692
+ scikit-optimize==0.10.1
693
+ scikit-plot==0.3.7
694
+ scikit-surprise==1.1.3
695
+ scipy==1.11.4
696
+ scipy==1.12.0
697
+ seaborn==0.12.2
698
+ segment_anything==1.0
699
+ segregation==2.5
700
+ semver==3.0.2
701
+ sentencepiece==0.2.0
702
+ sentry-sdk==1.42.0
703
+ setproctitle==1.3.3
704
+ setuptools-git==1.2
705
+ setuptools-scm==8.0.4
706
+ setuptools==69.0.3
707
+ shap==0.44.1
708
+ shapely==2.0.3
709
+ shellingham==1.5.4
710
+ simpervisor==1.0.0
711
+ simplejson==3.19.2
712
+ six==1.16.0
713
+ sklearn-pandas==2.2.0
714
+ slicer==0.0.7
715
+ smart-open==6.4.0
716
+ smmap==5.0.1
717
+ sniffio==1.3.0
718
+ snowballstemmer==2.2.0
719
+ snuggs==1.4.7
720
+ sortedcontainers==2.4.0
721
+ soundfile==0.12.1
722
+ soupsieve==2.5
723
+ soxr==0.3.7
724
+ spacy-legacy==3.0.12
725
+ spacy-loggers==1.0.5
726
+ spacy==3.7.2
727
+ spaghetti==1.7.5.post1
728
+ spectral==0.23.1
729
+ spglm==1.1.0
730
+ sphinx-rtd-theme==0.2.4
731
+ spint==1.0.7
732
+ splot==1.1.5.post1
733
+ spopt==0.6.0
734
+ spreg==1.4.2
735
+ spvcm==0.3.0
736
+ sqlparse==0.4.4
737
+ squarify==0.4.3
738
+ srsly==2.4.8
739
+ stable-baselines3==2.1.0
740
+ stack-data==0.6.2
741
+ stack-data==0.6.3
742
+ stanio==0.3.0
743
+ starlette==0.32.0.post1
744
+ statsmodels==0.14.1
745
+ stemming==1.0.1
746
+ stop-words==2018.7.23
747
+ stopit==1.1.2
748
+ stumpy==1.12.0
749
+ sympy==1.12
750
+ tables==3.9.2
751
+ tabulate==0.9.0
752
+ tangled-up-in-unicode==0.2.0
753
+ tbb==2021.11.0
754
+ tblib==3.0.0
755
+ tenacity==8.2.3
756
+ tensorboard-data-server==0.7.2
757
+ tensorboard-plugin-profile==2.15.0
758
+ tensorboard==2.15.1
759
+ tensorboardX==2.6.2.2
760
+ tensorflow-cloud==0.1.16
761
+ tensorflow-datasets==4.9.4
762
+ tensorflow-decision-forests==1.8.1
763
+ tensorflow-estimator==2.15.0
764
+ tensorflow-hub==0.16.1
765
+ tensorflow-io-gcs-filesystem==0.35.0
766
+ tensorflow-io==0.35.0
767
+ tensorflow-metadata==0.14.0
768
+ tensorflow-probability==0.23.0
769
+ tensorflow-serving-api==2.14.1
770
+ tensorflow-text==2.15.0
771
+ tensorflow-transform==0.14.0
772
+ tensorflow==2.15.0
773
+ tensorpack==0.11
774
+ tensorstore==0.1.56
775
+ termcolor==2.4.0
776
+ terminado==0.18.0
777
+ testpath==0.6.0
778
+ text-unidecode==1.3
779
+ textblob==0.18.0.post0
780
+ texttable==1.7.0
781
+ tf_keras==2.15.1
782
+ tfp-nightly==0.24.0.dev0
783
+ thinc==8.2.2
784
+ threadpoolctl==3.2.0
785
+ tifffile==2023.12.9
786
+ timm==0.9.16
787
+ tinycss2==1.2.1
788
+ tobler==0.11.2
789
+ tokenizers==0.15.2
790
+ toml==0.10.2
791
+ tomli==2.0.1
792
+ tomlkit==0.12.4
793
+ toolz==0.12.1
794
+ torch==2.1.2
795
+ torchaudio==2.1.2
796
+ torchdata==0.7.1
797
+ torchinfo==1.8.0
798
+ torchmetrics==1.3.2
799
+ torchtext==0.16.2
800
+ torchvision==0.16.2
801
+ tornado==6.3.3
802
+ tqdm==4.66.1
803
+ traceml==1.0.8
804
+ traitlets==5.9.0
805
+ traittypes==0.2.1
806
+ transformers==4.38.2
807
+ treelite-runtime==3.2.0
808
+ treelite==3.2.0
809
+ trueskill==0.4.5
810
+ truststore==0.8.0
811
+ trx-python==0.2.9
812
+ tsfresh==0.20.2
813
+ typeguard==4.1.5
814
+ typer==0.9.0
815
+ types-python-dateutil==2.8.19.20240106
816
+ typing-inspect==0.9.0
817
+ typing-utils==0.1.0
818
+ typing_extensions==4.9.0
819
+ tzdata==2023.4
820
+ uc-micro-py==1.0.3
821
+ ucx-py==0.33.0
822
+ ujson==5.9.0
823
+ umap-learn==0.5.5
824
+ unicodedata2==15.1.0
825
+ update-checker==0.18.0
826
+ uri-template==1.3.0
827
+ uritemplate==3.0.1
828
+ urllib3==1.26.18
829
+ urllib3==2.1.0
830
+ urwid==2.6.9
831
+ urwid_readline==0.14
832
+ uvicorn==0.25.0
833
+ uvloop==0.19.0
834
+ vaex-astro==0.9.3
835
+ vaex-core==4.17.1
836
+ vaex-hdf5==0.14.1
837
+ vaex-jupyter==0.8.2
838
+ vaex-ml==0.18.3
839
+ vaex-server==0.9.0
840
+ vaex-viz==0.5.4
841
+ vaex==4.17.0
842
+ vec_noise==1.1.4
843
+ vecstack==0.4.0
844
+ virtualenv==20.21.0
845
+ visions==0.7.5
846
+ vowpalwabbit==9.9.0
847
+ vtk==9.3.0
848
+ wandb==0.16.4
849
+ wasabi==1.1.2
850
+ watchfiles==0.21.0
851
+ wavio==0.0.8
852
+ wcwidth==0.2.13
853
+ weasel==0.3.4
854
+ webcolors==1.13
855
+ webencodings==0.5.1
856
+ websocket-client==1.7.0
857
+ websockets==12.0
858
+ wfdb==4.1.2
859
+ whatthepatch==1.0.5
860
+ wheel==0.42.0
861
+ widgetsnbextension==3.6.6
862
+ witwidget==1.8.1
863
+ woodwork==0.29.0
864
+ wordcloud==1.9.3
865
+ wordsegment==1.3.1
866
+ wrapt==1.14.1
867
+ xarray-einstats==0.7.0
868
+ xarray==2024.2.0
869
+ xgboost==2.0.3
870
+ xvfbwrapper==0.2.9
871
+ xxhash==3.4.1
872
+ xyzservices==2023.10.1
873
+ y-py==0.6.2
874
+ yapf==0.40.2
875
+ yarl==1.9.3
876
+ yarl==1.9.4
877
+ ydata-profiling==4.6.4
878
+ yellowbrick==1.5
879
+ ypy-websocket==0.8.4
880
+ zict==3.0.0
881
+ zipp==3.17.0
882
+ zstandard==0.22.0
wandb/run-20240402_072503-b95w935d/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-02T07:25:04.219238",
5
+ "startedAt": "2024-04-02T07:25:03.308167",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "0ecfc053b35f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5524.099773406982
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240402_072503-b95w935d/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.3094, "train/grad_norm": 0.5635281801223755, "train/learning_rate": 0.0, "train/epoch": 28.8, "train/global_step": 360, "_timestamp": 1712049025.4445002, "_runtime": 6322.128933191299, "_step": 58, "eval/loss": 0.549111545085907, "eval/runtime": 19.991, "eval/samples_per_second": 3.752, "eval/steps_per_second": 0.65, "train/train_runtime": 6328.6782, "train/train_samples_per_second": 1.422, "train/train_steps_per_second": 0.057, "train/total_flos": 861990497353728.0, "train/train_loss": 0.5323640836609734}
wandb/run-20240402_072503-b95w935d/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240402_072503-b95w935d/logs/debug.log ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-02 07:25:03,309 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.4
2
+ 2024-04-02 07:25:03,309 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240402_072503-b95w935d/logs/debug.log
10
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240402_072503-b95w935d/logs/debug-internal.log
11
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7e03f1c07cd0>
12
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-04-02 07:25:03,310 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-02 07:25:03,311 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-04-02 07:25:03,311 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-04-02 07:25:03,312 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-02 07:25:03,315 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-04-02 07:25:03,328 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1295] probe notebook
20
+ 2024-04-02 07:25:03,908 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-04-02 07:25:03,911 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-04-02 07:25:04,116 INFO MainThread:34 [wandb_run.py:_on_init():2262] communicating current version
23
+ 2024-04-02 07:25:04,182 INFO MainThread:34 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.5 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-02 07:25:04,183 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-04-02 07:25:35,178 INFO MainThread:34 [wandb_run.py:_console_start():2241] atexit reg
27
+ 2024-04-02 07:25:35,178 INFO MainThread:34 [wandb_run.py:_redirect():2096] redirect: wrap_raw
28
+ 2024-04-02 07:25:35,179 INFO MainThread:34 [wandb_run.py:_redirect():2161] Wrapping output streams.
29
+ 2024-04-02 07:25:35,179 INFO MainThread:34 [wandb_run.py:_redirect():2186] Redirects installed.
30
+ 2024-04-02 07:25:35,180 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-04-02 07:25:35,186 INFO MainThread:34 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 32768, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'sliding_window': None, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'use_cache': False, 'rope_theta': 1000000.0, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['MistralForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TheBloke/Mistral-7B-Instruct-v0.2-GPTQ', 'transformers_version': '4.38.2', 'model_type': 'mistral', 'pretraining_tp': 1, 'quantization_config': {'quant_method': 'QuantizationMethod.GPTQ', 'bits': 4, 'tokenizer': None, 'dataset': None, 'group_size': 128, 'damp_percent': 0.1, 'desc_act': True, 'sym': True, 'true_sequential': True, 'use_cuda_fp16': False, 'model_seqlen': None, 'block_name_to_quantize': None, 'module_name_preceding_first_block': None, 'batch_size': 1, 'pad_token_id': None, 'use_exllama': True, 'max_input_length': None, 'exllama_config': {'version': 'ExllamaVersion.ONE'}, 'cache_block_outputs': True, 'modules_in_block_to_quantize': None}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 6, 'per_device_eval_batch_size': 6, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0002, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 2, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr02_07-24-52_0ecfc053b35f', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
+ 2024-04-02 09:10:25,449 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-02 09:10:25,449 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-04-02 09:10:45,681 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-04-02 09:10:45,704 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
+ 2024-04-02 09:10:45,704 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
37
+ 2024-04-02 09:11:24,497 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
38
+ 2024-04-02 09:11:24,498 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
+ 2024-04-02 09:11:24,498 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
40
+ 2024-04-02 09:11:34,304 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
41
+ 2024-04-02 09:11:34,306 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
+ 2024-04-02 09:11:34,306 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
43
+ 2024-04-02 09:11:35,541 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
wandb/run-20240402_072503-b95w935d/run-b95w935d.wandb ADDED
Binary file (218 kB). View file