Femboyuwu2000 commited on
Commit
9fbd070
1 Parent(s): 5415c3f
Files changed (50) hide show
  1. README.md +1 -1
  2. adapter_config.json +3 -2
  3. adapter_model.safetensors +2 -2
  4. runs/Apr11_11-38-55_55cd788bb41a/events.out.tfevents.1712836020.55cd788bb41a.94.0 +3 -0
  5. runs/Apr11_11-52-16_55cd788bb41a/events.out.tfevents.1712836797.55cd788bb41a.427.0 +3 -0
  6. runs/Apr11_12-04-45_55cd788bb41a/events.out.tfevents.1712837287.55cd788bb41a.427.1 +3 -0
  7. runs/Apr11_12-10-45_55cd788bb41a/events.out.tfevents.1712837449.55cd788bb41a.999.0 +3 -0
  8. runs/Apr11_12-17-04_55cd788bb41a/events.out.tfevents.1712838278.55cd788bb41a.1425.0 +3 -0
  9. special_tokens_map.json +7 -0
  10. tokenizer.json +2 -2
  11. tokenizer_config.json +16 -8
  12. training_args.bin +1 -1
  13. wandb/debug-internal.log +0 -0
  14. wandb/debug.log +29 -29
  15. wandb/run-20240411_114743-9c27btud/files/conda-environment.yaml +0 -0
  16. wandb/run-20240411_114743-9c27btud/files/config.yaml +685 -0
  17. wandb/run-20240411_114743-9c27btud/files/output.log +134 -0
  18. wandb/run-20240411_114743-9c27btud/files/requirements.txt +864 -0
  19. wandb/run-20240411_114743-9c27btud/files/wandb-metadata.json +66 -0
  20. wandb/run-20240411_114743-9c27btud/files/wandb-summary.json +1 -0
  21. wandb/run-20240411_114743-9c27btud/logs/debug-internal.log +271 -0
  22. wandb/run-20240411_114743-9c27btud/logs/debug.log +34 -0
  23. wandb/run-20240411_114743-9c27btud/run-9c27btud.wandb +0 -0
  24. wandb/run-20240411_115958-sm65qpc7/files/conda-environment.yaml +0 -0
  25. wandb/run-20240411_115958-sm65qpc7/files/config.yaml +701 -0
  26. wandb/run-20240411_115958-sm65qpc7/files/output.log +275 -0
  27. wandb/run-20240411_115958-sm65qpc7/files/requirements.txt +864 -0
  28. wandb/run-20240411_115958-sm65qpc7/files/wandb-metadata.json +66 -0
  29. wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json +1 -0
  30. wandb/run-20240411_115958-sm65qpc7/logs/debug-internal.log +483 -0
  31. wandb/run-20240411_115958-sm65qpc7/logs/debug.log +37 -0
  32. wandb/run-20240411_115958-sm65qpc7/run-sm65qpc7.wandb +0 -0
  33. wandb/run-20240411_121051-1qmakf52/files/conda-environment.yaml +0 -0
  34. wandb/run-20240411_121051-1qmakf52/files/config.yaml +701 -0
  35. wandb/run-20240411_121051-1qmakf52/files/output.log +14 -0
  36. wandb/run-20240411_121051-1qmakf52/files/requirements.txt +864 -0
  37. wandb/run-20240411_121051-1qmakf52/files/wandb-metadata.json +66 -0
  38. wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json +1 -0
  39. wandb/run-20240411_121051-1qmakf52/logs/debug-internal.log +269 -0
  40. wandb/run-20240411_121051-1qmakf52/logs/debug.log +33 -0
  41. wandb/run-20240411_121051-1qmakf52/run-1qmakf52.wandb +0 -0
  42. wandb/run-20240411_122441-3wiow4fo/files/conda-environment.yaml +0 -0
  43. wandb/run-20240411_122441-3wiow4fo/files/config.yaml +690 -0
  44. wandb/run-20240411_122441-3wiow4fo/files/output.log +103 -0
  45. wandb/run-20240411_122441-3wiow4fo/files/requirements.txt +864 -0
  46. wandb/run-20240411_122441-3wiow4fo/files/wandb-metadata.json +66 -0
  47. wandb/run-20240411_122441-3wiow4fo/files/wandb-summary.json +1 -0
  48. wandb/run-20240411_122441-3wiow4fo/logs/debug-internal.log +127 -0
  49. wandb/run-20240411_122441-3wiow4fo/logs/debug.log +31 -0
  50. wandb/run-20240411_122441-3wiow4fo/run-3wiow4fo.wandb +0 -0
README.md CHANGED
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
- - training_steps: 20000
46
 
47
  ### Training results
48
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
+ - training_steps: 10
46
 
47
  ### Training results
48
 
adapter_config.json CHANGED
@@ -20,9 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "query_key_value",
24
  "lm_head",
25
- "dense_h_to_4h"
 
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "lm_head",
24
+ "dense_h_to_4h",
25
+ "query_key_value",
26
+ "word_embeddings"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ac730ac4465d0ec9fbe49ed9c5ad2a69ec11ddf9dc1d9cb0f5595f60817af39
3
- size 1042757704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b6634512ad03e3f4ebc6d111357c12706dde03297921ad8a59889665659e4b1
3
+ size 2078423528
runs/Apr11_11-38-55_55cd788bb41a/events.out.tfevents.1712836020.55cd788bb41a.94.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:720200355a94b4a797a1a91707db83746af1b302f10d5dfa32f5b5d25957baf1
3
+ size 20632
runs/Apr11_11-52-16_55cd788bb41a/events.out.tfevents.1712836797.55cd788bb41a.427.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:480709c9a9c2751c4226237550a295b35f078662084627b69529c800fd614d0b
3
+ size 20632
runs/Apr11_12-04-45_55cd788bb41a/events.out.tfevents.1712837287.55cd788bb41a.427.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:329802ecf9648166758fc23aa17501d530a5d5ff3d438bdc2dbc3463f43f2439
3
+ size 19799
runs/Apr11_12-10-45_55cd788bb41a/events.out.tfevents.1712837449.55cd788bb41a.999.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97f862a1c6d6aaba7e809468d2e122e134bc9966b2a772a68f2b23bc6e2e9b33
3
+ size 20839
runs/Apr11_12-17-04_55cd788bb41a/events.out.tfevents.1712838278.55cd788bb41a.1425.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c93e0f11e3c0cff9a65eb8c5395851334df8fee2abc3f52c243bce81b00fbb1
3
+ size 15841
special_tokens_map.json CHANGED
@@ -1,4 +1,11 @@
1
  {
 
 
 
 
 
 
 
2
  "eos_token": {
3
  "content": "</s>",
4
  "lstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
  "eos_token": {
10
  "content": "</s>",
11
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54e5c72a5ea09da48b2f316760b8bc5a445683ab9a5bc6b68db5d8db624ecceb
3
- size 16315213
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:490b3432398e9c45dda0581bf34f6daa9411bda59cd0cd831670d7c86f90d2dd
3
+ size 14500570
tokenizer_config.json CHANGED
@@ -1,7 +1,8 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "<pad>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
@@ -9,7 +10,7 @@
9
  "special": true
10
  },
11
  "1": {
12
- "content": "</s>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
@@ -17,7 +18,15 @@
17
  "special": true
18
  },
19
  "2": {
20
- "content": "<unk>",
 
 
 
 
 
 
 
 
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
@@ -25,15 +34,14 @@
25
  "special": true
26
  }
27
  },
28
- "additional_special_tokens": [],
29
- "clean_up_tokenization_spaces": true,
30
  "eos_token": "</s>",
31
- "extra_ids": 0,
32
  "max_length": 1024,
33
  "model_max_length": 1000000000000000019884624838656,
34
  "pad_token": "</s>",
35
- "sp_model_kwargs": {},
36
- "tokenizer_class": "T5Tokenizer",
37
  "truncation": true,
38
  "unk_token": "<unk>"
39
  }
 
1
  {
2
+ "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
+ "content": "<unk>",
6
  "lstrip": false,
7
  "normalized": false,
8
  "rstrip": false,
 
10
  "special": true
11
  },
12
  "1": {
13
+ "content": "<s>",
14
  "lstrip": false,
15
  "normalized": false,
16
  "rstrip": false,
 
18
  "special": true
19
  },
20
  "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<pad>",
30
  "lstrip": false,
31
  "normalized": false,
32
  "rstrip": false,
 
34
  "special": true
35
  }
36
  },
37
+ "bos_token": "<s>",
38
+ "clean_up_tokenization_spaces": false,
39
  "eos_token": "</s>",
 
40
  "max_length": 1024,
41
  "model_max_length": 1000000000000000019884624838656,
42
  "pad_token": "</s>",
43
+ "padding_side": "left",
44
+ "tokenizer_class": "BloomTokenizer",
45
  "truncation": true,
46
  "unk_token": "<unk>"
47
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ba44607bb84c5fd76d74e8d257eb1a18a8dceedf164c95e0e7b4eb4771e56e1
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7936773bf7232091b96c6d72977f6ab9f7e29932f703ce05f5fef485d916ec5d
3
  size 4920
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,31 +1,31 @@
1
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Configure stats pid to 1050
3
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug.log
10
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_040722-rimyh9a1/logs/debug-internal.log
11
- 2024-04-11 04:07:22,843 INFO MainThread:1050 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7d563f5f5120>
12
- 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():567] calling init triggers
13
- 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():617] starting backend
16
- 2024-04-11 04:07:22,844 INFO MainThread:1050 [wandb_init.py:init():621] setting up manager
17
- 2024-04-11 04:07:22,847 INFO MainThread:1050 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-11 04:07:22,848 INFO MainThread:1050 [wandb_init.py:init():629] backend started and connected
19
- 2024-04-11 04:07:22,865 INFO MainThread:1050 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
- 2024-04-11 04:07:23,106 INFO MainThread:1050 [wandb_init.py:init():721] updated telemetry
21
- 2024-04-11 04:07:23,108 INFO MainThread:1050 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
- 2024-04-11 04:07:23,235 INFO MainThread:1050 [wandb_run.py:_on_init():2344] communicating current version
23
- 2024-04-11 04:07:23,323 INFO MainThread:1050 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-11 04:07:23,324 INFO MainThread:1050 [wandb_init.py:init():805] starting run threads in backend
26
- 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_console_start():2323] atexit reg
27
- 2024-04-11 04:07:39,317 INFO MainThread:1050 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
- 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
- 2024-04-11 04:07:39,318 INFO MainThread:1050 [wandb_run.py:_redirect():2268] Redirects installed.
30
- 2024-04-11 04:07:39,319 INFO MainThread:1050 [wandb_init.py:init():848] run started, returning control to user process
31
- 2024-04-11 04:07:39,324 INFO MainThread:1050 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_04-07-17_321ddc84c049', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
1
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Configure stats pid to 1425
3
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_122441-3wiow4fo/logs/debug.log
10
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_122441-3wiow4fo/logs/debug-internal.log
11
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7bd820cc4d00>
12
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 12:24:41,036 INFO MainThread:1425 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 12:24:41,038 INFO MainThread:1425 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 12:24:41,050 INFO MainThread:1425 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 12:24:41,423 INFO MainThread:1425 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 12:24:41,432 INFO MainThread:1425 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 12:24:41,583 INFO MainThread:1425 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 12:24:41,671 INFO MainThread:1425 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-11 12:24:41,672 INFO MainThread:1425 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 12:24:57,689 INFO MainThread:1425 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 12:24:57,689 INFO MainThread:1425 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 12:24:57,690 INFO MainThread:1425 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 12:24:57,690 INFO MainThread:1425 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 12:24:57,691 INFO MainThread:1425 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 12:24:57,696 INFO MainThread:1425 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_12-17-04_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240411_114743-9c27btud/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_114743-9c27btud/files/config.yaml ADDED
@@ -0,0 +1,685 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712836063.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ vocab_size:
62
+ desc: null
63
+ value: 250880
64
+ hidden_size:
65
+ desc: null
66
+ value: 1536
67
+ n_layer:
68
+ desc: null
69
+ value: 24
70
+ n_head:
71
+ desc: null
72
+ value: 16
73
+ layer_norm_epsilon:
74
+ desc: null
75
+ value: 1.0e-05
76
+ initializer_range:
77
+ desc: null
78
+ value: 0.02
79
+ use_cache:
80
+ desc: null
81
+ value: false
82
+ pretraining_tp:
83
+ desc: null
84
+ value: 1
85
+ apply_residual_connection_post_layernorm:
86
+ desc: null
87
+ value: false
88
+ hidden_dropout:
89
+ desc: null
90
+ value: 0.0
91
+ attention_dropout:
92
+ desc: null
93
+ value: 0.0
94
+ bos_token_id:
95
+ desc: null
96
+ value: 1
97
+ eos_token_id:
98
+ desc: null
99
+ value: 2
100
+ slow_but_exact:
101
+ desc: null
102
+ value: false
103
+ return_dict:
104
+ desc: null
105
+ value: true
106
+ output_hidden_states:
107
+ desc: null
108
+ value: false
109
+ output_attentions:
110
+ desc: null
111
+ value: false
112
+ torchscript:
113
+ desc: null
114
+ value: false
115
+ torch_dtype:
116
+ desc: null
117
+ value: null
118
+ use_bfloat16:
119
+ desc: null
120
+ value: false
121
+ tf_legacy_loss:
122
+ desc: null
123
+ value: false
124
+ pruned_heads:
125
+ desc: null
126
+ value: {}
127
+ tie_word_embeddings:
128
+ desc: null
129
+ value: true
130
+ chunk_size_feed_forward:
131
+ desc: null
132
+ value: 0
133
+ is_encoder_decoder:
134
+ desc: null
135
+ value: false
136
+ is_decoder:
137
+ desc: null
138
+ value: false
139
+ cross_attention_hidden_size:
140
+ desc: null
141
+ value: null
142
+ add_cross_attention:
143
+ desc: null
144
+ value: false
145
+ tie_encoder_decoder:
146
+ desc: null
147
+ value: false
148
+ max_length:
149
+ desc: null
150
+ value: 20
151
+ min_length:
152
+ desc: null
153
+ value: 0
154
+ do_sample:
155
+ desc: null
156
+ value: false
157
+ early_stopping:
158
+ desc: null
159
+ value: false
160
+ num_beams:
161
+ desc: null
162
+ value: 1
163
+ num_beam_groups:
164
+ desc: null
165
+ value: 1
166
+ diversity_penalty:
167
+ desc: null
168
+ value: 0.0
169
+ temperature:
170
+ desc: null
171
+ value: 1.0
172
+ top_k:
173
+ desc: null
174
+ value: 50
175
+ top_p:
176
+ desc: null
177
+ value: 1.0
178
+ typical_p:
179
+ desc: null
180
+ value: 1.0
181
+ repetition_penalty:
182
+ desc: null
183
+ value: 1.0
184
+ length_penalty:
185
+ desc: null
186
+ value: 1.0
187
+ no_repeat_ngram_size:
188
+ desc: null
189
+ value: 0
190
+ encoder_no_repeat_ngram_size:
191
+ desc: null
192
+ value: 0
193
+ bad_words_ids:
194
+ desc: null
195
+ value: null
196
+ num_return_sequences:
197
+ desc: null
198
+ value: 1
199
+ output_scores:
200
+ desc: null
201
+ value: false
202
+ return_dict_in_generate:
203
+ desc: null
204
+ value: false
205
+ forced_bos_token_id:
206
+ desc: null
207
+ value: null
208
+ forced_eos_token_id:
209
+ desc: null
210
+ value: null
211
+ remove_invalid_values:
212
+ desc: null
213
+ value: false
214
+ exponential_decay_length_penalty:
215
+ desc: null
216
+ value: null
217
+ suppress_tokens:
218
+ desc: null
219
+ value: null
220
+ begin_suppress_tokens:
221
+ desc: null
222
+ value: null
223
+ architectures:
224
+ desc: null
225
+ value:
226
+ - BloomForCausalLM
227
+ finetuning_task:
228
+ desc: null
229
+ value: null
230
+ id2label:
231
+ desc: null
232
+ value:
233
+ '0': LABEL_0
234
+ '1': LABEL_1
235
+ label2id:
236
+ desc: null
237
+ value:
238
+ LABEL_0: 0
239
+ LABEL_1: 1
240
+ tokenizer_class:
241
+ desc: null
242
+ value: null
243
+ prefix:
244
+ desc: null
245
+ value: null
246
+ pad_token_id:
247
+ desc: null
248
+ value: 3
249
+ sep_token_id:
250
+ desc: null
251
+ value: null
252
+ decoder_start_token_id:
253
+ desc: null
254
+ value: null
255
+ task_specific_params:
256
+ desc: null
257
+ value: null
258
+ problem_type:
259
+ desc: null
260
+ value: null
261
+ _name_or_path:
262
+ desc: null
263
+ value: bigscience/bloomz-1b1
264
+ transformers_version:
265
+ desc: null
266
+ value: 4.39.3
267
+ attention_softmax_in_fp32:
268
+ desc: null
269
+ value: true
270
+ bias_dropout_fusion:
271
+ desc: null
272
+ value: true
273
+ unk_token_id:
274
+ desc: null
275
+ value: 0
276
+ masked_softmax_fusion:
277
+ desc: null
278
+ value: true
279
+ model_type:
280
+ desc: null
281
+ value: bloom
282
+ n_inner:
283
+ desc: null
284
+ value: null
285
+ offset_alibi:
286
+ desc: null
287
+ value: 100
288
+ seq_length:
289
+ desc: null
290
+ value: 2048
291
+ skip_bias_add:
292
+ desc: null
293
+ value: true
294
+ skip_bias_add_qkv:
295
+ desc: null
296
+ value: false
297
+ quantization_config:
298
+ desc: null
299
+ value:
300
+ quant_method: QuantizationMethod.BITS_AND_BYTES
301
+ _load_in_8bit: false
302
+ _load_in_4bit: true
303
+ llm_int8_threshold: 6.0
304
+ llm_int8_skip_modules: null
305
+ llm_int8_enable_fp32_cpu_offload: false
306
+ llm_int8_has_fp16_weight: false
307
+ bnb_4bit_quant_type: nf4
308
+ bnb_4bit_use_double_quant: false
309
+ bnb_4bit_compute_dtype: float16
310
+ bnb_4bit_quant_storage: uint8
311
+ load_in_4bit: true
312
+ load_in_8bit: false
313
+ output_dir:
314
+ desc: null
315
+ value: /kaggle/working/
316
+ overwrite_output_dir:
317
+ desc: null
318
+ value: false
319
+ do_train:
320
+ desc: null
321
+ value: false
322
+ do_eval:
323
+ desc: null
324
+ value: false
325
+ do_predict:
326
+ desc: null
327
+ value: false
328
+ evaluation_strategy:
329
+ desc: null
330
+ value: 'no'
331
+ prediction_loss_only:
332
+ desc: null
333
+ value: false
334
+ per_device_train_batch_size:
335
+ desc: null
336
+ value: 8
337
+ per_device_eval_batch_size:
338
+ desc: null
339
+ value: 8
340
+ per_gpu_train_batch_size:
341
+ desc: null
342
+ value: null
343
+ per_gpu_eval_batch_size:
344
+ desc: null
345
+ value: null
346
+ gradient_accumulation_steps:
347
+ desc: null
348
+ value: 1
349
+ eval_accumulation_steps:
350
+ desc: null
351
+ value: null
352
+ eval_delay:
353
+ desc: null
354
+ value: 0
355
+ learning_rate:
356
+ desc: null
357
+ value: 3.0e-05
358
+ weight_decay:
359
+ desc: null
360
+ value: 0.001
361
+ adam_beta1:
362
+ desc: null
363
+ value: 0.9
364
+ adam_beta2:
365
+ desc: null
366
+ value: 0.999
367
+ adam_epsilon:
368
+ desc: null
369
+ value: 1.0e-08
370
+ max_grad_norm:
371
+ desc: null
372
+ value: 0.3
373
+ num_train_epochs:
374
+ desc: null
375
+ value: 5
376
+ max_steps:
377
+ desc: null
378
+ value: 20000
379
+ lr_scheduler_type:
380
+ desc: null
381
+ value: cosine
382
+ lr_scheduler_kwargs:
383
+ desc: null
384
+ value: {}
385
+ warmup_ratio:
386
+ desc: null
387
+ value: 0.03
388
+ warmup_steps:
389
+ desc: null
390
+ value: 0
391
+ log_level:
392
+ desc: null
393
+ value: passive
394
+ log_level_replica:
395
+ desc: null
396
+ value: warning
397
+ log_on_each_node:
398
+ desc: null
399
+ value: true
400
+ logging_dir:
401
+ desc: null
402
+ value: /kaggle/working/runs/Apr11_11-38-55_55cd788bb41a
403
+ logging_strategy:
404
+ desc: null
405
+ value: steps
406
+ logging_first_step:
407
+ desc: null
408
+ value: false
409
+ logging_steps:
410
+ desc: null
411
+ value: 20
412
+ logging_nan_inf_filter:
413
+ desc: null
414
+ value: true
415
+ save_strategy:
416
+ desc: null
417
+ value: steps
418
+ save_steps:
419
+ desc: null
420
+ value: 20
421
+ save_total_limit:
422
+ desc: null
423
+ value: 1
424
+ save_safetensors:
425
+ desc: null
426
+ value: true
427
+ save_on_each_node:
428
+ desc: null
429
+ value: false
430
+ save_only_model:
431
+ desc: null
432
+ value: false
433
+ no_cuda:
434
+ desc: null
435
+ value: false
436
+ use_cpu:
437
+ desc: null
438
+ value: false
439
+ use_mps_device:
440
+ desc: null
441
+ value: false
442
+ seed:
443
+ desc: null
444
+ value: 42
445
+ data_seed:
446
+ desc: null
447
+ value: null
448
+ jit_mode_eval:
449
+ desc: null
450
+ value: false
451
+ use_ipex:
452
+ desc: null
453
+ value: false
454
+ bf16:
455
+ desc: null
456
+ value: false
457
+ fp16:
458
+ desc: null
459
+ value: false
460
+ fp16_opt_level:
461
+ desc: null
462
+ value: O1
463
+ half_precision_backend:
464
+ desc: null
465
+ value: auto
466
+ bf16_full_eval:
467
+ desc: null
468
+ value: false
469
+ fp16_full_eval:
470
+ desc: null
471
+ value: false
472
+ tf32:
473
+ desc: null
474
+ value: null
475
+ local_rank:
476
+ desc: null
477
+ value: 0
478
+ ddp_backend:
479
+ desc: null
480
+ value: null
481
+ tpu_num_cores:
482
+ desc: null
483
+ value: null
484
+ tpu_metrics_debug:
485
+ desc: null
486
+ value: false
487
+ debug:
488
+ desc: null
489
+ value: []
490
+ dataloader_drop_last:
491
+ desc: null
492
+ value: false
493
+ eval_steps:
494
+ desc: null
495
+ value: null
496
+ dataloader_num_workers:
497
+ desc: null
498
+ value: 8
499
+ dataloader_prefetch_factor:
500
+ desc: null
501
+ value: null
502
+ past_index:
503
+ desc: null
504
+ value: -1
505
+ run_name:
506
+ desc: null
507
+ value: /kaggle/working/
508
+ disable_tqdm:
509
+ desc: null
510
+ value: false
511
+ remove_unused_columns:
512
+ desc: null
513
+ value: true
514
+ label_names:
515
+ desc: null
516
+ value: null
517
+ load_best_model_at_end:
518
+ desc: null
519
+ value: false
520
+ metric_for_best_model:
521
+ desc: null
522
+ value: null
523
+ greater_is_better:
524
+ desc: null
525
+ value: null
526
+ ignore_data_skip:
527
+ desc: null
528
+ value: false
529
+ fsdp:
530
+ desc: null
531
+ value: []
532
+ fsdp_min_num_params:
533
+ desc: null
534
+ value: 0
535
+ fsdp_config:
536
+ desc: null
537
+ value:
538
+ min_num_params: 0
539
+ xla: false
540
+ xla_fsdp_v2: false
541
+ xla_fsdp_grad_ckpt: false
542
+ fsdp_transformer_layer_cls_to_wrap:
543
+ desc: null
544
+ value: null
545
+ accelerator_config:
546
+ desc: null
547
+ value:
548
+ split_batches: false
549
+ dispatch_batches: null
550
+ even_batches: true
551
+ use_seedable_sampler: true
552
+ deepspeed:
553
+ desc: null
554
+ value: null
555
+ label_smoothing_factor:
556
+ desc: null
557
+ value: 0.0
558
+ optim:
559
+ desc: null
560
+ value: paged_adamw_32bit
561
+ optim_args:
562
+ desc: null
563
+ value: null
564
+ adafactor:
565
+ desc: null
566
+ value: false
567
+ group_by_length:
568
+ desc: null
569
+ value: false
570
+ length_column_name:
571
+ desc: null
572
+ value: length
573
+ report_to:
574
+ desc: null
575
+ value:
576
+ - tensorboard
577
+ - wandb
578
+ ddp_find_unused_parameters:
579
+ desc: null
580
+ value: null
581
+ ddp_bucket_cap_mb:
582
+ desc: null
583
+ value: null
584
+ ddp_broadcast_buffers:
585
+ desc: null
586
+ value: null
587
+ dataloader_pin_memory:
588
+ desc: null
589
+ value: true
590
+ dataloader_persistent_workers:
591
+ desc: null
592
+ value: false
593
+ skip_memory_metrics:
594
+ desc: null
595
+ value: true
596
+ use_legacy_prediction_loop:
597
+ desc: null
598
+ value: false
599
+ push_to_hub:
600
+ desc: null
601
+ value: false
602
+ resume_from_checkpoint:
603
+ desc: null
604
+ value: null
605
+ hub_model_id:
606
+ desc: null
607
+ value: null
608
+ hub_strategy:
609
+ desc: null
610
+ value: every_save
611
+ hub_token:
612
+ desc: null
613
+ value: <HUB_TOKEN>
614
+ hub_private_repo:
615
+ desc: null
616
+ value: false
617
+ hub_always_push:
618
+ desc: null
619
+ value: false
620
+ gradient_checkpointing:
621
+ desc: null
622
+ value: true
623
+ gradient_checkpointing_kwargs:
624
+ desc: null
625
+ value: null
626
+ include_inputs_for_metrics:
627
+ desc: null
628
+ value: false
629
+ fp16_backend:
630
+ desc: null
631
+ value: auto
632
+ push_to_hub_model_id:
633
+ desc: null
634
+ value: null
635
+ push_to_hub_organization:
636
+ desc: null
637
+ value: null
638
+ push_to_hub_token:
639
+ desc: null
640
+ value: <PUSH_TO_HUB_TOKEN>
641
+ mp_parameters:
642
+ desc: null
643
+ value: ''
644
+ auto_find_batch_size:
645
+ desc: null
646
+ value: true
647
+ full_determinism:
648
+ desc: null
649
+ value: false
650
+ torchdynamo:
651
+ desc: null
652
+ value: null
653
+ ray_scope:
654
+ desc: null
655
+ value: last
656
+ ddp_timeout:
657
+ desc: null
658
+ value: 1800
659
+ torch_compile:
660
+ desc: null
661
+ value: false
662
+ torch_compile_backend:
663
+ desc: null
664
+ value: null
665
+ torch_compile_mode:
666
+ desc: null
667
+ value: null
668
+ dispatch_batches:
669
+ desc: null
670
+ value: null
671
+ split_batches:
672
+ desc: null
673
+ value: null
674
+ include_tokens_per_second:
675
+ desc: null
676
+ value: false
677
+ include_num_input_tokens_seen:
678
+ desc: null
679
+ value: false
680
+ neftune_noise_alpha:
681
+ desc: null
682
+ value: null
683
+ optim_target_modules:
684
+ desc: null
685
+ value: null
wandb/run-20240411_114743-9c27btud/files/output.log ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
38
+ warnings.warn('Was asked to gather along dimension 0, but all '
39
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
40
+ To disable this warning, you can either:
41
+ - Avoid using `tokenizers` before the fork if possible
42
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
44
+ To disable this warning, you can either:
45
+ - Avoid using `tokenizers` before the fork if possible
46
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
47
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
48
+ To disable this warning, you can either:
49
+ - Avoid using `tokenizers` before the fork if possible
50
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
51
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
52
+ To disable this warning, you can either:
53
+ - Avoid using `tokenizers` before the fork if possible
54
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
55
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
56
+ To disable this warning, you can either:
57
+ - Avoid using `tokenizers` before the fork if possible
58
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
59
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
60
+ To disable this warning, you can either:
61
+ - Avoid using `tokenizers` before the fork if possible
62
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
63
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
64
+ To disable this warning, you can either:
65
+ - Avoid using `tokenizers` before the fork if possible
66
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
67
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
68
+ To disable this warning, you can either:
69
+ - Avoid using `tokenizers` before the fork if possible
70
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
71
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
72
+ To disable this warning, you can either:
73
+ - Avoid using `tokenizers` before the fork if possible
74
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
75
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
76
+ To disable this warning, you can either:
77
+ - Avoid using `tokenizers` before the fork if possible
78
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
79
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
80
+ To disable this warning, you can either:
81
+ - Avoid using `tokenizers` before the fork if possible
82
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
83
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
84
+ To disable this warning, you can either:
85
+ - Avoid using `tokenizers` before the fork if possible
86
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
87
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
88
+ To disable this warning, you can either:
89
+ - Avoid using `tokenizers` before the fork if possible
90
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
91
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
92
+ To disable this warning, you can either:
93
+ - Avoid using `tokenizers` before the fork if possible
94
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
95
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
96
+ To disable this warning, you can either:
97
+ - Avoid using `tokenizers` before the fork if possible
98
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
99
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
100
+ To disable this warning, you can either:
101
+ - Avoid using `tokenizers` before the fork if possible
102
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
104
+ To disable this warning, you can either:
105
+ - Avoid using `tokenizers` before the fork if possible
106
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
107
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
108
+ To disable this warning, you can either:
109
+ - Avoid using `tokenizers` before the fork if possible
110
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
111
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
112
+ To disable this warning, you can either:
113
+ - Avoid using `tokenizers` before the fork if possible
114
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
115
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
116
+ To disable this warning, you can either:
117
+ - Avoid using `tokenizers` before the fork if possible
118
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
119
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
120
+ To disable this warning, you can either:
121
+ - Avoid using `tokenizers` before the fork if possible
122
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
123
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
124
+ To disable this warning, you can either:
125
+ - Avoid using `tokenizers` before the fork if possible
126
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
127
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
128
+ To disable this warning, you can either:
129
+ - Avoid using `tokenizers` before the fork if possible
130
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
131
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
132
+ To disable this warning, you can either:
133
+ - Avoid using `tokenizers` before the fork if possible
134
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
wandb/run-20240411_114743-9c27btud/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_114743-9c27btud/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T11:47:44.334234",
5
+ "startedAt": "2024-04-11T11:47:43.611001",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "55cd788bb41a",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.18,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.18,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.18,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.18,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.18,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5567.39563369751
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_114743-9c27btud/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 84}}
wandb/run-20240411_114743-9c27btud/logs/debug-internal.log ADDED
@@ -0,0 +1,271 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 11:47:43,617 INFO StreamThr :141 [internal.py:wandb_internal():86] W&B internal server running at pid: 141, started at: 2024-04-11 11:47:43.616921
2
+ 2024-04-11 11:47:43,618 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 11:47:44,043 INFO WriterThread:141 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_114743-9c27btud/run-9c27btud.wandb
4
+ 2024-04-11 11:47:44,044 DEBUG SenderThread:141 [sender.py:send():379] send: header
5
+ 2024-04-11 11:47:44,047 DEBUG SenderThread:141 [sender.py:send():379] send: run
6
+ 2024-04-11 11:47:44,202 INFO SenderThread:141 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_114743-9c27btud/files
7
+ 2024-04-11 11:47:44,202 INFO SenderThread:141 [sender.py:_start_run_threads():1124] run started: 9c27btud with start time 1712836063.619587
8
+ 2024-04-11 11:47:44,210 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 11:47:44,211 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 11:47:44,303 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 11:47:44,314 DEBUG HandlerThread:141 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 11:47:44,314 DEBUG HandlerThread:141 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 11:47:44,314 INFO HandlerThread:141 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 11:47:44,315 INFO SystemMonitor:141 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 11:47:44,315 INFO HandlerThread:141 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-11 11:47:44,315 INFO SystemMonitor:141 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-11 11:47:44,316 INFO SystemMonitor:141 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-11 11:47:44,317 INFO SystemMonitor:141 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 11:47:44,319 INFO SystemMonitor:141 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 11:47:44,319 INFO SystemMonitor:141 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 11:47:44,334 DEBUG HandlerThread:141 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 11:47:44,336 DEBUG HandlerThread:141 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 11:47:44,336 DEBUG HandlerThread:141 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 11:47:44,336 DEBUG HandlerThread:141 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T11:47:44.334234', 'startedAt': '2024-04-11T11:47:43.611001', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '55cd788bb41a', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.18, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5567.39563369751}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 11:47:44,336 INFO HandlerThread:141 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 11:47:44,336 INFO HandlerThread:141 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 11:47:44,336 DEBUG HandlerThread:141 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 11:47:45,205 INFO Thread-12 :141 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/conda-environment.yaml
29
+ 2024-04-11 11:47:59,351 ERROR HandlerThread:141 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 11:47:59,356 DEBUG HandlerThread:141 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 11:47:59,357 INFO HandlerThread:141 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 11:47:59,366 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 11:47:59,366 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 11:47:59,366 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 11:47:59,366 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 11:47:59,366 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 11:47:59,367 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 11:47:59,367 DEBUG SenderThread:141 [sender.py:send():379] send: files
49
+ 2024-04-11 11:47:59,367 INFO SenderThread:141 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 11:47:59,586 INFO wandb-upload_0:141 [upload_job.py:push():131] Uploaded file /tmp/tmpxpmf6ih7wandb/oe84il23-wandb-metadata.json
51
+ 2024-04-11 11:48:00,207 INFO Thread-12 :141 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/wandb-metadata.json
52
+ 2024-04-11 11:48:00,391 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 11:48:00,391 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 11:48:00,394 DEBUG SenderThread:141 [sender.py:send():379] send: telemetry
55
+ 2024-04-11 11:48:00,394 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
56
+ 2024-04-11 11:48:00,400 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
57
+ 2024-04-11 11:48:00,400 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
58
+ 2024-04-11 11:48:00,528 DEBUG SenderThread:141 [sender.py:send():379] send: config
59
+ 2024-04-11 11:48:00,530 DEBUG SenderThread:141 [sender.py:send():379] send: metric
60
+ 2024-04-11 11:48:00,530 DEBUG SenderThread:141 [sender.py:send():379] send: telemetry
61
+ 2024-04-11 11:48:00,530 DEBUG SenderThread:141 [sender.py:send():379] send: metric
62
+ 2024-04-11 11:48:00,530 WARNING SenderThread:141 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
63
+ 2024-04-11 11:48:00,530 DEBUG SenderThread:141 [sender.py:send():379] send: telemetry
64
+ 2024-04-11 11:48:01,208 INFO Thread-12 :141 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
65
+ 2024-04-11 11:48:01,208 INFO Thread-12 :141 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/requirements.txt
66
+ 2024-04-11 11:48:03,209 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
67
+ 2024-04-11 11:48:05,002 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 11:48:07,210 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
69
+ 2024-04-11 11:48:10,002 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-11 11:48:15,008 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 11:48:15,213 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/config.yaml
72
+ 2024-04-11 11:48:15,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
73
+ 2024-04-11 11:48:15,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
74
+ 2024-04-11 11:48:15,394 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
75
+ 2024-04-11 11:48:20,477 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-11 11:48:25,217 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
77
+ 2024-04-11 11:48:26,358 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
78
+ 2024-04-11 11:48:30,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
79
+ 2024-04-11 11:48:30,394 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
80
+ 2024-04-11 11:48:30,394 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
81
+ 2024-04-11 11:48:31,480 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-11 11:48:36,481 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 11:48:41,907 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
84
+ 2024-04-11 11:48:43,224 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
85
+ 2024-04-11 11:48:44,320 DEBUG SystemMonitor:141 [system_monitor.py:_start():172] Starting system metrics aggregation loop
86
+ 2024-04-11 11:48:44,321 DEBUG SenderThread:141 [sender.py:send():379] send: stats
87
+ 2024-04-11 11:48:45,395 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
88
+ 2024-04-11 11:48:45,396 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
89
+ 2024-04-11 11:48:45,396 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
90
+ 2024-04-11 11:48:47,439 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
91
+ 2024-04-11 11:48:52,440 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-04-11 11:48:57,229 INFO Thread-12 :141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
93
+ 2024-04-11 11:48:57,934 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-04-11 11:49:00,394 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
95
+ 2024-04-11 11:49:00,394 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
96
+ 2024-04-11 11:49:00,397 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
97
+ 2024-04-11 11:49:03,490 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-04-11 11:49:08,492 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-11 11:49:09,236 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: pause
100
+ 2024-04-11 11:49:09,236 INFO HandlerThread:141 [handler.py:handle_request_pause():708] stopping system metrics thread
101
+ 2024-04-11 11:49:09,236 INFO HandlerThread:141 [system_monitor.py:finish():203] Stopping system monitor
102
+ 2024-04-11 11:49:09,237 DEBUG SystemMonitor:141 [system_monitor.py:_start():179] Finished system metrics aggregation loop
103
+ 2024-04-11 11:49:09,237 DEBUG SystemMonitor:141 [system_monitor.py:_start():183] Publishing last batch of metrics
104
+ 2024-04-11 11:49:09,238 INFO HandlerThread:141 [interfaces.py:finish():202] Joined cpu monitor
105
+ 2024-04-11 11:49:09,238 INFO HandlerThread:141 [interfaces.py:finish():202] Joined disk monitor
106
+ 2024-04-11 11:49:09,248 INFO HandlerThread:141 [interfaces.py:finish():202] Joined gpu monitor
107
+ 2024-04-11 11:49:09,248 INFO HandlerThread:141 [interfaces.py:finish():202] Joined memory monitor
108
+ 2024-04-11 11:49:09,248 INFO HandlerThread:141 [interfaces.py:finish():202] Joined network monitor
109
+ 2024-04-11 11:49:09,249 DEBUG SenderThread:141 [sender.py:send():379] send: stats
110
+ 2024-04-11 11:49:14,250 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
111
+ 2024-04-11 11:49:15,392 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
112
+ 2024-04-11 11:49:15,393 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
113
+ 2024-04-11 11:49:15,396 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
114
+ 2024-04-11 11:49:19,497 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-04-11 11:49:24,498 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 11:49:29,499 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
117
+ 2024-04-11 11:49:30,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
118
+ 2024-04-11 11:49:30,393 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
119
+ 2024-04-11 11:49:30,433 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
120
+ 2024-04-11 11:49:35,446 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
121
+ 2024-04-11 11:49:40,447 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-11 11:49:45,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
123
+ 2024-04-11 11:49:45,393 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
124
+ 2024-04-11 11:49:45,434 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
125
+ 2024-04-11 11:49:45,466 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-04-11 11:49:50,467 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-04-11 11:49:55,468 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-04-11 11:50:00,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
129
+ 2024-04-11 11:50:00,393 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
130
+ 2024-04-11 11:50:00,434 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
131
+ 2024-04-11 11:50:01,453 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
132
+ 2024-04-11 11:50:06,455 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
133
+ 2024-04-11 11:50:11,455 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-04-11 11:50:15,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
135
+ 2024-04-11 11:50:15,394 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
136
+ 2024-04-11 11:50:15,434 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
137
+ 2024-04-11 11:50:16,514 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-04-11 11:50:21,515 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-04-11 11:50:26,516 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-04-11 11:50:30,393 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: stop_status
141
+ 2024-04-11 11:50:30,394 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: stop_status
142
+ 2024-04-11 11:50:30,434 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
143
+ 2024-04-11 11:50:32,471 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
144
+ 2024-04-11 11:50:37,037 DEBUG SenderThread:141 [sender.py:send():379] send: exit
145
+ 2024-04-11 11:50:37,037 INFO SenderThread:141 [sender.py:send_exit():586] handling exit code: 0
146
+ 2024-04-11 11:50:37,037 INFO SenderThread:141 [sender.py:send_exit():588] handling runtime: 84
147
+ 2024-04-11 11:50:37,039 INFO SenderThread:141 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
148
+ 2024-04-11 11:50:37,039 INFO SenderThread:141 [sender.py:send_exit():594] send defer
149
+ 2024-04-11 11:50:37,039 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
150
+ 2024-04-11 11:50:37,040 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 0
151
+ 2024-04-11 11:50:37,040 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
152
+ 2024-04-11 11:50:37,040 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 0
153
+ 2024-04-11 11:50:37,040 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 1
154
+ 2024-04-11 11:50:37,040 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
155
+ 2024-04-11 11:50:37,040 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 1
156
+ 2024-04-11 11:50:37,040 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
157
+ 2024-04-11 11:50:37,041 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 1
158
+ 2024-04-11 11:50:37,041 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 2
159
+ 2024-04-11 11:50:37,041 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
160
+ 2024-04-11 11:50:37,041 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 2
161
+ 2024-04-11 11:50:37,041 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
162
+ 2024-04-11 11:50:37,041 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 2
163
+ 2024-04-11 11:50:37,041 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 3
164
+ 2024-04-11 11:50:37,041 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
165
+ 2024-04-11 11:50:37,041 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 3
166
+ 2024-04-11 11:50:37,041 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
167
+ 2024-04-11 11:50:37,041 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 3
168
+ 2024-04-11 11:50:37,042 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 4
169
+ 2024-04-11 11:50:37,042 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
170
+ 2024-04-11 11:50:37,042 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 4
171
+ 2024-04-11 11:50:37,042 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
172
+ 2024-04-11 11:50:37,042 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 4
173
+ 2024-04-11 11:50:37,042 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 5
174
+ 2024-04-11 11:50:37,042 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
175
+ 2024-04-11 11:50:37,042 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 5
176
+ 2024-04-11 11:50:37,042 DEBUG SenderThread:141 [sender.py:send():379] send: summary
177
+ 2024-04-11 11:50:37,043 INFO SenderThread:141 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
178
+ 2024-04-11 11:50:37,043 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
179
+ 2024-04-11 11:50:37,043 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 5
180
+ 2024-04-11 11:50:37,043 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 6
181
+ 2024-04-11 11:50:37,043 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
182
+ 2024-04-11 11:50:37,043 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 6
183
+ 2024-04-11 11:50:37,043 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
184
+ 2024-04-11 11:50:37,044 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 6
185
+ 2024-04-11 11:50:37,044 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 7
186
+ 2024-04-11 11:50:37,044 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: status_report
187
+ 2024-04-11 11:50:37,044 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
188
+ 2024-04-11 11:50:37,044 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 7
189
+ 2024-04-11 11:50:37,044 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
190
+ 2024-04-11 11:50:37,044 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 7
191
+ 2024-04-11 11:50:37,263 INFO Thread-12 :141 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/wandb-summary.json
192
+ 2024-04-11 11:50:38,037 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: poll_exit
193
+ 2024-04-11 11:50:38,759 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 8
194
+ 2024-04-11 11:50:38,760 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: poll_exit
195
+ 2024-04-11 11:50:38,760 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
196
+ 2024-04-11 11:50:38,760 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 8
197
+ 2024-04-11 11:50:38,760 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
198
+ 2024-04-11 11:50:38,760 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 8
199
+ 2024-04-11 11:50:38,760 INFO SenderThread:141 [job_builder.py:build():318] Attempting to build job artifact
200
+ 2024-04-11 11:50:38,762 INFO SenderThread:141 [job_builder.py:_get_source_type():466] no source found
201
+ 2024-04-11 11:50:38,762 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 9
202
+ 2024-04-11 11:50:38,762 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
203
+ 2024-04-11 11:50:38,762 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 9
204
+ 2024-04-11 11:50:38,763 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
205
+ 2024-04-11 11:50:38,763 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 9
206
+ 2024-04-11 11:50:38,763 INFO SenderThread:141 [dir_watcher.py:finish():358] shutting down directory watcher
207
+ 2024-04-11 11:50:39,039 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: poll_exit
208
+ 2024-04-11 11:50:39,264 INFO SenderThread:141 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
209
+ 2024-04-11 11:50:39,264 INFO SenderThread:141 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_114743-9c27btud/files
210
+ 2024-04-11 11:50:39,265 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/wandb-metadata.json wandb-metadata.json
211
+ 2024-04-11 11:50:39,265 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/conda-environment.yaml conda-environment.yaml
212
+ 2024-04-11 11:50:39,265 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log output.log
213
+ 2024-04-11 11:50:39,265 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/wandb-summary.json wandb-summary.json
214
+ 2024-04-11 11:50:39,265 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/config.yaml config.yaml
215
+ 2024-04-11 11:50:39,271 INFO SenderThread:141 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_114743-9c27btud/files/requirements.txt requirements.txt
216
+ 2024-04-11 11:50:39,275 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 10
217
+ 2024-04-11 11:50:39,275 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: poll_exit
218
+ 2024-04-11 11:50:39,276 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
219
+ 2024-04-11 11:50:39,276 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 10
220
+ 2024-04-11 11:50:39,279 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
221
+ 2024-04-11 11:50:39,279 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 10
222
+ 2024-04-11 11:50:39,279 INFO SenderThread:141 [file_pusher.py:finish():172] shutting down file pusher
223
+ 2024-04-11 11:50:39,480 INFO wandb-upload_1:141 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_114743-9c27btud/files/wandb-summary.json
224
+ 2024-04-11 11:50:39,526 INFO wandb-upload_2:141 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_114743-9c27btud/files/config.yaml
225
+ 2024-04-11 11:50:39,538 INFO wandb-upload_3:141 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_114743-9c27btud/files/requirements.txt
226
+ 2024-04-11 11:50:39,542 INFO wandb-upload_0:141 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_114743-9c27btud/files/output.log
227
+ 2024-04-11 11:50:39,743 INFO Thread-11 (_thread_body):141 [sender.py:transition_state():614] send defer: 11
228
+ 2024-04-11 11:50:39,743 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
229
+ 2024-04-11 11:50:39,743 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 11
230
+ 2024-04-11 11:50:39,744 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
231
+ 2024-04-11 11:50:39,744 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 11
232
+ 2024-04-11 11:50:39,744 INFO SenderThread:141 [file_pusher.py:join():178] waiting for file pusher
233
+ 2024-04-11 11:50:39,744 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 12
234
+ 2024-04-11 11:50:39,744 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
235
+ 2024-04-11 11:50:39,745 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 12
236
+ 2024-04-11 11:50:39,745 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
237
+ 2024-04-11 11:50:39,745 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 12
238
+ 2024-04-11 11:50:39,745 INFO SenderThread:141 [file_stream.py:finish():614] file stream finish called
239
+ 2024-04-11 11:50:39,819 INFO SenderThread:141 [file_stream.py:finish():618] file stream finish is done
240
+ 2024-04-11 11:50:39,819 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 13
241
+ 2024-04-11 11:50:39,819 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
242
+ 2024-04-11 11:50:39,819 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 13
243
+ 2024-04-11 11:50:39,820 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
244
+ 2024-04-11 11:50:39,820 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 13
245
+ 2024-04-11 11:50:39,820 INFO SenderThread:141 [sender.py:transition_state():614] send defer: 14
246
+ 2024-04-11 11:50:39,820 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: defer
247
+ 2024-04-11 11:50:39,820 INFO HandlerThread:141 [handler.py:handle_request_defer():172] handle defer: 14
248
+ 2024-04-11 11:50:39,820 DEBUG SenderThread:141 [sender.py:send():379] send: final
249
+ 2024-04-11 11:50:39,821 DEBUG SenderThread:141 [sender.py:send():379] send: footer
250
+ 2024-04-11 11:50:39,821 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: defer
251
+ 2024-04-11 11:50:39,821 INFO SenderThread:141 [sender.py:send_request_defer():610] handle sender defer: 14
252
+ 2024-04-11 11:50:39,822 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: poll_exit
253
+ 2024-04-11 11:50:39,822 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: poll_exit
254
+ 2024-04-11 11:50:39,823 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: poll_exit
255
+ 2024-04-11 11:50:39,823 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: poll_exit
256
+ 2024-04-11 11:50:39,823 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: server_info
257
+ 2024-04-11 11:50:39,824 DEBUG SenderThread:141 [sender.py:send_request():406] send_request: server_info
258
+ 2024-04-11 11:50:39,827 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: get_summary
259
+ 2024-04-11 11:50:39,827 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: sampled_history
260
+ 2024-04-11 11:50:39,828 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: internal_messages
261
+ 2024-04-11 11:50:39,836 INFO MainThread:141 [wandb_run.py:_footer_history_summary_info():3920] rendering history
262
+ 2024-04-11 11:50:39,836 INFO MainThread:141 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
263
+ 2024-04-11 11:50:39,837 INFO MainThread:141 [wandb_run.py:_footer_sync_info():3879] logging synced files
264
+ 2024-04-11 11:50:39,837 DEBUG HandlerThread:141 [handler.py:handle_request():146] handle_request: shutdown
265
+ 2024-04-11 11:50:39,837 INFO HandlerThread:141 [handler.py:finish():866] shutting down handler
266
+ 2024-04-11 11:50:40,801 WARNING StreamThr :141 [internal.py:is_dead():414] Internal process exiting, parent pid 94 disappeared
267
+ 2024-04-11 11:50:40,801 ERROR StreamThr :141 [internal.py:wandb_internal():152] Internal process shutdown.
268
+ 2024-04-11 11:50:40,824 INFO WriterThread:141 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_114743-9c27btud/run-9c27btud.wandb
269
+ 2024-04-11 11:50:40,836 INFO SenderThread:141 [sender.py:finish():1546] shutting down sender
270
+ 2024-04-11 11:50:40,836 INFO SenderThread:141 [file_pusher.py:finish():172] shutting down file pusher
271
+ 2024-04-11 11:50:40,837 INFO SenderThread:141 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_114743-9c27btud/logs/debug.log ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Configure stats pid to 94
3
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 11:47:43,612 INFO MainThread:94 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_114743-9c27btud/logs/debug.log
11
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_114743-9c27btud/logs/debug-internal.log
12
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c8d7316f7f0>
13
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:init():617] starting backend
17
+ 2024-04-11 11:47:43,613 INFO MainThread:94 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-11 11:47:43,615 INFO MainThread:94 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-11 11:47:43,619 INFO MainThread:94 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-11 11:47:43,630 INFO MainThread:94 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-11 11:47:44,042 INFO MainThread:94 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-11 11:47:44,046 INFO MainThread:94 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-11 11:47:44,210 INFO MainThread:94 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-11 11:47:44,297 INFO MainThread:94 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-04-11 11:47:44,297 INFO MainThread:94 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-11 11:48:00,392 INFO MainThread:94 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-11 11:48:00,392 INFO MainThread:94 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-11 11:48:00,392 INFO MainThread:94 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-11 11:48:00,392 INFO MainThread:94 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-11 11:48:00,394 INFO MainThread:94 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-11 11:48:00,401 INFO MainThread:94 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_11-38-55_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-11 11:49:09,235 INFO MainThread:94 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-11 11:49:09,235 INFO MainThread:94 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_114743-9c27btud/run-9c27btud.wandb ADDED
Binary file (23.8 kB). View file
 
wandb/run-20240411_115958-sm65qpc7/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_115958-sm65qpc7/files/config.yaml ADDED
@@ -0,0 +1,701 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712836798.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ vocab_size:
78
+ desc: null
79
+ value: 250880
80
+ hidden_size:
81
+ desc: null
82
+ value: 1536
83
+ n_layer:
84
+ desc: null
85
+ value: 24
86
+ n_head:
87
+ desc: null
88
+ value: 16
89
+ layer_norm_epsilon:
90
+ desc: null
91
+ value: 1.0e-05
92
+ initializer_range:
93
+ desc: null
94
+ value: 0.02
95
+ use_cache:
96
+ desc: null
97
+ value: false
98
+ pretraining_tp:
99
+ desc: null
100
+ value: 1
101
+ apply_residual_connection_post_layernorm:
102
+ desc: null
103
+ value: false
104
+ hidden_dropout:
105
+ desc: null
106
+ value: 0.0
107
+ attention_dropout:
108
+ desc: null
109
+ value: 0.0
110
+ bos_token_id:
111
+ desc: null
112
+ value: 1
113
+ eos_token_id:
114
+ desc: null
115
+ value: 2
116
+ slow_but_exact:
117
+ desc: null
118
+ value: false
119
+ return_dict:
120
+ desc: null
121
+ value: true
122
+ output_hidden_states:
123
+ desc: null
124
+ value: false
125
+ output_attentions:
126
+ desc: null
127
+ value: false
128
+ torchscript:
129
+ desc: null
130
+ value: false
131
+ torch_dtype:
132
+ desc: null
133
+ value: null
134
+ use_bfloat16:
135
+ desc: null
136
+ value: false
137
+ tf_legacy_loss:
138
+ desc: null
139
+ value: false
140
+ pruned_heads:
141
+ desc: null
142
+ value: {}
143
+ tie_word_embeddings:
144
+ desc: null
145
+ value: true
146
+ chunk_size_feed_forward:
147
+ desc: null
148
+ value: 0
149
+ is_encoder_decoder:
150
+ desc: null
151
+ value: false
152
+ is_decoder:
153
+ desc: null
154
+ value: false
155
+ cross_attention_hidden_size:
156
+ desc: null
157
+ value: null
158
+ add_cross_attention:
159
+ desc: null
160
+ value: false
161
+ tie_encoder_decoder:
162
+ desc: null
163
+ value: false
164
+ max_length:
165
+ desc: null
166
+ value: 20
167
+ min_length:
168
+ desc: null
169
+ value: 0
170
+ do_sample:
171
+ desc: null
172
+ value: false
173
+ early_stopping:
174
+ desc: null
175
+ value: false
176
+ num_beams:
177
+ desc: null
178
+ value: 1
179
+ num_beam_groups:
180
+ desc: null
181
+ value: 1
182
+ diversity_penalty:
183
+ desc: null
184
+ value: 0.0
185
+ temperature:
186
+ desc: null
187
+ value: 1.0
188
+ top_k:
189
+ desc: null
190
+ value: 50
191
+ top_p:
192
+ desc: null
193
+ value: 1.0
194
+ typical_p:
195
+ desc: null
196
+ value: 1.0
197
+ repetition_penalty:
198
+ desc: null
199
+ value: 1.0
200
+ length_penalty:
201
+ desc: null
202
+ value: 1.0
203
+ no_repeat_ngram_size:
204
+ desc: null
205
+ value: 0
206
+ encoder_no_repeat_ngram_size:
207
+ desc: null
208
+ value: 0
209
+ bad_words_ids:
210
+ desc: null
211
+ value: null
212
+ num_return_sequences:
213
+ desc: null
214
+ value: 1
215
+ output_scores:
216
+ desc: null
217
+ value: false
218
+ return_dict_in_generate:
219
+ desc: null
220
+ value: false
221
+ forced_bos_token_id:
222
+ desc: null
223
+ value: null
224
+ forced_eos_token_id:
225
+ desc: null
226
+ value: null
227
+ remove_invalid_values:
228
+ desc: null
229
+ value: false
230
+ exponential_decay_length_penalty:
231
+ desc: null
232
+ value: null
233
+ suppress_tokens:
234
+ desc: null
235
+ value: null
236
+ begin_suppress_tokens:
237
+ desc: null
238
+ value: null
239
+ architectures:
240
+ desc: null
241
+ value:
242
+ - BloomForCausalLM
243
+ finetuning_task:
244
+ desc: null
245
+ value: null
246
+ id2label:
247
+ desc: null
248
+ value:
249
+ '0': LABEL_0
250
+ '1': LABEL_1
251
+ label2id:
252
+ desc: null
253
+ value:
254
+ LABEL_0: 0
255
+ LABEL_1: 1
256
+ tokenizer_class:
257
+ desc: null
258
+ value: null
259
+ prefix:
260
+ desc: null
261
+ value: null
262
+ pad_token_id:
263
+ desc: null
264
+ value: 3
265
+ sep_token_id:
266
+ desc: null
267
+ value: null
268
+ decoder_start_token_id:
269
+ desc: null
270
+ value: null
271
+ task_specific_params:
272
+ desc: null
273
+ value: null
274
+ problem_type:
275
+ desc: null
276
+ value: null
277
+ _name_or_path:
278
+ desc: null
279
+ value: bigscience/bloomz-1b1
280
+ transformers_version:
281
+ desc: null
282
+ value: 4.39.3
283
+ attention_softmax_in_fp32:
284
+ desc: null
285
+ value: true
286
+ bias_dropout_fusion:
287
+ desc: null
288
+ value: true
289
+ unk_token_id:
290
+ desc: null
291
+ value: 0
292
+ masked_softmax_fusion:
293
+ desc: null
294
+ value: true
295
+ model_type:
296
+ desc: null
297
+ value: bloom
298
+ n_inner:
299
+ desc: null
300
+ value: null
301
+ offset_alibi:
302
+ desc: null
303
+ value: 100
304
+ seq_length:
305
+ desc: null
306
+ value: 2048
307
+ skip_bias_add:
308
+ desc: null
309
+ value: true
310
+ skip_bias_add_qkv:
311
+ desc: null
312
+ value: false
313
+ quantization_config:
314
+ desc: null
315
+ value:
316
+ quant_method: QuantizationMethod.BITS_AND_BYTES
317
+ _load_in_8bit: false
318
+ _load_in_4bit: true
319
+ llm_int8_threshold: 6.0
320
+ llm_int8_skip_modules: null
321
+ llm_int8_enable_fp32_cpu_offload: false
322
+ llm_int8_has_fp16_weight: false
323
+ bnb_4bit_quant_type: nf4
324
+ bnb_4bit_use_double_quant: false
325
+ bnb_4bit_compute_dtype: float16
326
+ bnb_4bit_quant_storage: uint8
327
+ load_in_4bit: true
328
+ load_in_8bit: false
329
+ output_dir:
330
+ desc: null
331
+ value: /kaggle/working/
332
+ overwrite_output_dir:
333
+ desc: null
334
+ value: false
335
+ do_train:
336
+ desc: null
337
+ value: false
338
+ do_eval:
339
+ desc: null
340
+ value: false
341
+ do_predict:
342
+ desc: null
343
+ value: false
344
+ evaluation_strategy:
345
+ desc: null
346
+ value: 'no'
347
+ prediction_loss_only:
348
+ desc: null
349
+ value: false
350
+ per_device_train_batch_size:
351
+ desc: null
352
+ value: 8
353
+ per_device_eval_batch_size:
354
+ desc: null
355
+ value: 8
356
+ per_gpu_train_batch_size:
357
+ desc: null
358
+ value: null
359
+ per_gpu_eval_batch_size:
360
+ desc: null
361
+ value: null
362
+ gradient_accumulation_steps:
363
+ desc: null
364
+ value: 1
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: null
368
+ eval_delay:
369
+ desc: null
370
+ value: 0
371
+ learning_rate:
372
+ desc: null
373
+ value: 3.0e-05
374
+ weight_decay:
375
+ desc: null
376
+ value: 0.001
377
+ adam_beta1:
378
+ desc: null
379
+ value: 0.9
380
+ adam_beta2:
381
+ desc: null
382
+ value: 0.999
383
+ adam_epsilon:
384
+ desc: null
385
+ value: 1.0e-08
386
+ max_grad_norm:
387
+ desc: null
388
+ value: 0.3
389
+ num_train_epochs:
390
+ desc: null
391
+ value: 5
392
+ max_steps:
393
+ desc: null
394
+ value: 20000
395
+ lr_scheduler_type:
396
+ desc: null
397
+ value: cosine
398
+ lr_scheduler_kwargs:
399
+ desc: null
400
+ value: {}
401
+ warmup_ratio:
402
+ desc: null
403
+ value: 0.03
404
+ warmup_steps:
405
+ desc: null
406
+ value: 0
407
+ log_level:
408
+ desc: null
409
+ value: passive
410
+ log_level_replica:
411
+ desc: null
412
+ value: warning
413
+ log_on_each_node:
414
+ desc: null
415
+ value: true
416
+ logging_dir:
417
+ desc: null
418
+ value: /kaggle/working/runs/Apr11_12-04-45_55cd788bb41a
419
+ logging_strategy:
420
+ desc: null
421
+ value: steps
422
+ logging_first_step:
423
+ desc: null
424
+ value: false
425
+ logging_steps:
426
+ desc: null
427
+ value: 20
428
+ logging_nan_inf_filter:
429
+ desc: null
430
+ value: true
431
+ save_strategy:
432
+ desc: null
433
+ value: steps
434
+ save_steps:
435
+ desc: null
436
+ value: 20
437
+ save_total_limit:
438
+ desc: null
439
+ value: 1
440
+ save_safetensors:
441
+ desc: null
442
+ value: true
443
+ save_on_each_node:
444
+ desc: null
445
+ value: false
446
+ save_only_model:
447
+ desc: null
448
+ value: false
449
+ no_cuda:
450
+ desc: null
451
+ value: false
452
+ use_cpu:
453
+ desc: null
454
+ value: false
455
+ use_mps_device:
456
+ desc: null
457
+ value: false
458
+ seed:
459
+ desc: null
460
+ value: 42
461
+ data_seed:
462
+ desc: null
463
+ value: null
464
+ jit_mode_eval:
465
+ desc: null
466
+ value: false
467
+ use_ipex:
468
+ desc: null
469
+ value: false
470
+ bf16:
471
+ desc: null
472
+ value: false
473
+ fp16:
474
+ desc: null
475
+ value: false
476
+ fp16_opt_level:
477
+ desc: null
478
+ value: O1
479
+ half_precision_backend:
480
+ desc: null
481
+ value: auto
482
+ bf16_full_eval:
483
+ desc: null
484
+ value: false
485
+ fp16_full_eval:
486
+ desc: null
487
+ value: false
488
+ tf32:
489
+ desc: null
490
+ value: null
491
+ local_rank:
492
+ desc: null
493
+ value: 0
494
+ ddp_backend:
495
+ desc: null
496
+ value: null
497
+ tpu_num_cores:
498
+ desc: null
499
+ value: null
500
+ tpu_metrics_debug:
501
+ desc: null
502
+ value: false
503
+ debug:
504
+ desc: null
505
+ value: []
506
+ dataloader_drop_last:
507
+ desc: null
508
+ value: false
509
+ eval_steps:
510
+ desc: null
511
+ value: null
512
+ dataloader_num_workers:
513
+ desc: null
514
+ value: 8
515
+ dataloader_prefetch_factor:
516
+ desc: null
517
+ value: null
518
+ past_index:
519
+ desc: null
520
+ value: -1
521
+ run_name:
522
+ desc: null
523
+ value: /kaggle/working/
524
+ disable_tqdm:
525
+ desc: null
526
+ value: false
527
+ remove_unused_columns:
528
+ desc: null
529
+ value: true
530
+ label_names:
531
+ desc: null
532
+ value: null
533
+ load_best_model_at_end:
534
+ desc: null
535
+ value: false
536
+ metric_for_best_model:
537
+ desc: null
538
+ value: null
539
+ greater_is_better:
540
+ desc: null
541
+ value: null
542
+ ignore_data_skip:
543
+ desc: null
544
+ value: false
545
+ fsdp:
546
+ desc: null
547
+ value: []
548
+ fsdp_min_num_params:
549
+ desc: null
550
+ value: 0
551
+ fsdp_config:
552
+ desc: null
553
+ value:
554
+ min_num_params: 0
555
+ xla: false
556
+ xla_fsdp_v2: false
557
+ xla_fsdp_grad_ckpt: false
558
+ fsdp_transformer_layer_cls_to_wrap:
559
+ desc: null
560
+ value: null
561
+ accelerator_config:
562
+ desc: null
563
+ value:
564
+ split_batches: false
565
+ dispatch_batches: null
566
+ even_batches: true
567
+ use_seedable_sampler: true
568
+ deepspeed:
569
+ desc: null
570
+ value: null
571
+ label_smoothing_factor:
572
+ desc: null
573
+ value: 0.0
574
+ optim:
575
+ desc: null
576
+ value: paged_adamw_32bit
577
+ optim_args:
578
+ desc: null
579
+ value: null
580
+ adafactor:
581
+ desc: null
582
+ value: false
583
+ group_by_length:
584
+ desc: null
585
+ value: false
586
+ length_column_name:
587
+ desc: null
588
+ value: length
589
+ report_to:
590
+ desc: null
591
+ value:
592
+ - tensorboard
593
+ - wandb
594
+ ddp_find_unused_parameters:
595
+ desc: null
596
+ value: null
597
+ ddp_bucket_cap_mb:
598
+ desc: null
599
+ value: null
600
+ ddp_broadcast_buffers:
601
+ desc: null
602
+ value: null
603
+ dataloader_pin_memory:
604
+ desc: null
605
+ value: true
606
+ dataloader_persistent_workers:
607
+ desc: null
608
+ value: false
609
+ skip_memory_metrics:
610
+ desc: null
611
+ value: true
612
+ use_legacy_prediction_loop:
613
+ desc: null
614
+ value: false
615
+ push_to_hub:
616
+ desc: null
617
+ value: false
618
+ resume_from_checkpoint:
619
+ desc: null
620
+ value: null
621
+ hub_model_id:
622
+ desc: null
623
+ value: null
624
+ hub_strategy:
625
+ desc: null
626
+ value: every_save
627
+ hub_token:
628
+ desc: null
629
+ value: <HUB_TOKEN>
630
+ hub_private_repo:
631
+ desc: null
632
+ value: false
633
+ hub_always_push:
634
+ desc: null
635
+ value: false
636
+ gradient_checkpointing:
637
+ desc: null
638
+ value: true
639
+ gradient_checkpointing_kwargs:
640
+ desc: null
641
+ value: null
642
+ include_inputs_for_metrics:
643
+ desc: null
644
+ value: false
645
+ fp16_backend:
646
+ desc: null
647
+ value: auto
648
+ push_to_hub_model_id:
649
+ desc: null
650
+ value: null
651
+ push_to_hub_organization:
652
+ desc: null
653
+ value: null
654
+ push_to_hub_token:
655
+ desc: null
656
+ value: <PUSH_TO_HUB_TOKEN>
657
+ mp_parameters:
658
+ desc: null
659
+ value: ''
660
+ auto_find_batch_size:
661
+ desc: null
662
+ value: true
663
+ full_determinism:
664
+ desc: null
665
+ value: false
666
+ torchdynamo:
667
+ desc: null
668
+ value: null
669
+ ray_scope:
670
+ desc: null
671
+ value: last
672
+ ddp_timeout:
673
+ desc: null
674
+ value: 1800
675
+ torch_compile:
676
+ desc: null
677
+ value: false
678
+ torch_compile_backend:
679
+ desc: null
680
+ value: null
681
+ torch_compile_mode:
682
+ desc: null
683
+ value: null
684
+ dispatch_batches:
685
+ desc: null
686
+ value: null
687
+ split_batches:
688
+ desc: null
689
+ value: null
690
+ include_tokens_per_second:
691
+ desc: null
692
+ value: false
693
+ include_num_input_tokens_seen:
694
+ desc: null
695
+ value: false
696
+ neftune_noise_alpha:
697
+ desc: null
698
+ value: null
699
+ optim_target_modules:
700
+ desc: null
701
+ value: null
wandb/run-20240411_115958-sm65qpc7/files/output.log ADDED
@@ -0,0 +1,275 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
38
+ warnings.warn('Was asked to gather along dimension 0, but all '
39
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
40
+ To disable this warning, you can either:
41
+ - Avoid using `tokenizers` before the fork if possible
42
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
44
+ To disable this warning, you can either:
45
+ - Avoid using `tokenizers` before the fork if possible
46
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
47
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
48
+ To disable this warning, you can either:
49
+ - Avoid using `tokenizers` before the fork if possible
50
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
51
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
52
+ To disable this warning, you can either:
53
+ - Avoid using `tokenizers` before the fork if possible
54
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
55
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
56
+ To disable this warning, you can either:
57
+ - Avoid using `tokenizers` before the fork if possible
58
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
59
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
60
+ To disable this warning, you can either:
61
+ - Avoid using `tokenizers` before the fork if possible
62
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
63
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
64
+ To disable this warning, you can either:
65
+ - Avoid using `tokenizers` before the fork if possible
66
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
67
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
68
+ To disable this warning, you can either:
69
+ - Avoid using `tokenizers` before the fork if possible
70
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
71
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
72
+ To disable this warning, you can either:
73
+ - Avoid using `tokenizers` before the fork if possible
74
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
75
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
76
+ To disable this warning, you can either:
77
+ - Avoid using `tokenizers` before the fork if possible
78
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
79
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
80
+ To disable this warning, you can either:
81
+ - Avoid using `tokenizers` before the fork if possible
82
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
83
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
84
+ To disable this warning, you can either:
85
+ - Avoid using `tokenizers` before the fork if possible
86
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
87
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
88
+ To disable this warning, you can either:
89
+ - Avoid using `tokenizers` before the fork if possible
90
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
91
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
92
+ To disable this warning, you can either:
93
+ - Avoid using `tokenizers` before the fork if possible
94
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
95
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
96
+ To disable this warning, you can either:
97
+ - Avoid using `tokenizers` before the fork if possible
98
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
99
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
100
+ To disable this warning, you can either:
101
+ - Avoid using `tokenizers` before the fork if possible
102
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
104
+ To disable this warning, you can either:
105
+ - Avoid using `tokenizers` before the fork if possible
106
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
107
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
108
+ To disable this warning, you can either:
109
+ - Avoid using `tokenizers` before the fork if possible
110
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
111
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
112
+ To disable this warning, you can either:
113
+ - Avoid using `tokenizers` before the fork if possible
114
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
115
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
116
+ To disable this warning, you can either:
117
+ - Avoid using `tokenizers` before the fork if possible
118
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
119
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
120
+ To disable this warning, you can either:
121
+ - Avoid using `tokenizers` before the fork if possible
122
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
123
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
124
+ To disable this warning, you can either:
125
+ - Avoid using `tokenizers` before the fork if possible
126
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
127
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
128
+ To disable this warning, you can either:
129
+ - Avoid using `tokenizers` before the fork if possible
130
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
131
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
132
+ To disable this warning, you can either:
133
+ - Avoid using `tokenizers` before the fork if possible
134
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
135
+ /opt/conda/lib/python3.10/site-packages/trl/trainer/sft_trainer.py:317: UserWarning: You passed a tokenizer with `padding_side` not equal to `right` to the SFTTrainer. This might lead to some unexpected behaviour due to overflow issues when training a model in half-precision. You might consider adding `tokenizer.padding_side = 'right'` to your code.
136
+ warnings.warn(
137
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
138
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
139
+ warnings.warn(
140
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
141
+ To disable this warning, you can either:
142
+ - Avoid using `tokenizers` before the fork if possible
143
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
144
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
145
+ To disable this warning, you can either:
146
+ - Avoid using `tokenizers` before the fork if possible
147
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
148
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
149
+ To disable this warning, you can either:
150
+ - Avoid using `tokenizers` before the fork if possible
151
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
152
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
153
+ To disable this warning, you can either:
154
+ - Avoid using `tokenizers` before the fork if possible
155
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
156
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
157
+ To disable this warning, you can either:
158
+ - Avoid using `tokenizers` before the fork if possible
159
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
160
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
161
+ To disable this warning, you can either:
162
+ - Avoid using `tokenizers` before the fork if possible
163
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
164
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
165
+ To disable this warning, you can either:
166
+ - Avoid using `tokenizers` before the fork if possible
167
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
168
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
169
+ To disable this warning, you can either:
170
+ - Avoid using `tokenizers` before the fork if possible
171
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
172
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
173
+ To disable this warning, you can either:
174
+ - Avoid using `tokenizers` before the fork if possible
175
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
176
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
177
+ To disable this warning, you can either:
178
+ - Avoid using `tokenizers` before the fork if possible
179
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
180
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
181
+ To disable this warning, you can either:
182
+ - Avoid using `tokenizers` before the fork if possible
183
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
184
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
185
+ To disable this warning, you can either:
186
+ - Avoid using `tokenizers` before the fork if possible
187
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
188
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
189
+ To disable this warning, you can either:
190
+ - Avoid using `tokenizers` before the fork if possible
191
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
192
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
193
+ To disable this warning, you can either:
194
+ - Avoid using `tokenizers` before the fork if possible
195
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
196
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
197
+ To disable this warning, you can either:
198
+ - Avoid using `tokenizers` before the fork if possible
199
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
200
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
201
+ To disable this warning, you can either:
202
+ - Avoid using `tokenizers` before the fork if possible
203
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
204
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
205
+ To disable this warning, you can either:
206
+ - Avoid using `tokenizers` before the fork if possible
207
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
208
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
209
+ To disable this warning, you can either:
210
+ - Avoid using `tokenizers` before the fork if possible
211
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
212
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
213
+ To disable this warning, you can either:
214
+ - Avoid using `tokenizers` before the fork if possible
215
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
216
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
217
+ To disable this warning, you can either:
218
+ - Avoid using `tokenizers` before the fork if possible
219
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
220
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
221
+ To disable this warning, you can either:
222
+ - Avoid using `tokenizers` before the fork if possible
223
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
224
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
225
+ To disable this warning, you can either:
226
+ - Avoid using `tokenizers` before the fork if possible
227
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
228
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
229
+ To disable this warning, you can either:
230
+ - Avoid using `tokenizers` before the fork if possible
231
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
232
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
233
+ To disable this warning, you can either:
234
+ - Avoid using `tokenizers` before the fork if possible
235
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
236
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
237
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
238
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
239
+ warnings.warn(
240
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
241
+ warnings.warn('Was asked to gather along dimension 0, but all '
242
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
243
+ warnings.warn(_create_warning_msg(
244
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
245
+ To disable this warning, you can either:
246
+ - Avoid using `tokenizers` before the fork if possible
247
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
248
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
249
+ To disable this warning, you can either:
250
+ - Avoid using `tokenizers` before the fork if possible
251
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
252
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
253
+ To disable this warning, you can either:
254
+ - Avoid using `tokenizers` before the fork if possible
255
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
256
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
257
+ To disable this warning, you can either:
258
+ - Avoid using `tokenizers` before the fork if possible
259
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
260
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
261
+ To disable this warning, you can either:
262
+ - Avoid using `tokenizers` before the fork if possible
263
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
264
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
265
+ To disable this warning, you can either:
266
+ - Avoid using `tokenizers` before the fork if possible
267
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
268
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
269
+ To disable this warning, you can either:
270
+ - Avoid using `tokenizers` before the fork if possible
271
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
272
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
273
+ To disable this warning, you can either:
274
+ - Avoid using `tokenizers` before the fork if possible
275
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
wandb/run-20240411_115958-sm65qpc7/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_115958-sm65qpc7/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T11:59:59.292356",
5
+ "startedAt": "2024-04-11T11:59:58.685383",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "55cd788bb41a",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.18,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.18,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.18,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.18,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.18,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5568.725933074951
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 6.2818, "train/grad_norm": 47.7755126953125, "train/learning_rate": 1e-06, "train/epoch": 0.0, "train/global_step": 20, "_timestamp": 1712837347.0713778, "_runtime": 548.3782157897949, "_step": 0, "_wandb": {"runtime": 375}}
wandb/run-20240411_115958-sm65qpc7/logs/debug-internal.log ADDED
@@ -0,0 +1,483 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 11:59:58,693 INFO StreamThr :472 [internal.py:wandb_internal():86] W&B internal server running at pid: 472, started at: 2024-04-11 11:59:58.693009
2
+ 2024-04-11 11:59:58,697 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 11:59:59,019 INFO WriterThread:472 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/run-sm65qpc7.wandb
4
+ 2024-04-11 11:59:59,020 DEBUG SenderThread:472 [sender.py:send():379] send: header
5
+ 2024-04-11 11:59:59,022 DEBUG SenderThread:472 [sender.py:send():379] send: run
6
+ 2024-04-11 11:59:59,164 INFO SenderThread:472 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files
7
+ 2024-04-11 11:59:59,164 INFO SenderThread:472 [sender.py:_start_run_threads():1124] run started: sm65qpc7 with start time 1712836798.693162
8
+ 2024-04-11 11:59:59,173 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 11:59:59,174 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 11:59:59,266 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 11:59:59,277 DEBUG HandlerThread:472 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 11:59:59,277 DEBUG HandlerThread:472 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 11:59:59,277 INFO HandlerThread:472 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 11:59:59,277 INFO SystemMonitor:472 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 11:59:59,278 INFO SystemMonitor:472 [interfaces.py:start():190] Started cpu monitoring
16
+ 2024-04-11 11:59:59,278 INFO HandlerThread:472 [system_monitor.py:probe():214] Collecting system info
17
+ 2024-04-11 11:59:59,279 INFO SystemMonitor:472 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-11 11:59:59,280 INFO SystemMonitor:472 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 11:59:59,281 INFO SystemMonitor:472 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 11:59:59,282 INFO SystemMonitor:472 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 11:59:59,292 DEBUG HandlerThread:472 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 11:59:59,295 DEBUG HandlerThread:472 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 11:59:59,295 DEBUG HandlerThread:472 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 11:59:59,295 DEBUG HandlerThread:472 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T11:59:59.292356', 'startedAt': '2024-04-11T11:59:58.685383', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '55cd788bb41a', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.18, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5568.725933074951}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 11:59:59,295 INFO HandlerThread:472 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 11:59:59,295 INFO HandlerThread:472 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 11:59:59,295 DEBUG HandlerThread:472 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 12:00:00,167 INFO Thread-12 :472 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/conda-environment.yaml
29
+ 2024-04-11 12:00:14,311 ERROR HandlerThread:472 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 12:00:14,312 DEBUG HandlerThread:472 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 12:00:14,312 INFO HandlerThread:472 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 12:00:14,317 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 12:00:14,317 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 12:00:14,317 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 12:00:14,317 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 12:00:14,318 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 12:00:14,318 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 12:00:14,318 DEBUG SenderThread:472 [sender.py:send():379] send: files
49
+ 2024-04-11 12:00:14,318 INFO SenderThread:472 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 12:00:14,510 INFO wandb-upload_0:472 [upload_job.py:push():131] Uploaded file /tmp/tmpduacgb07wandb/hbppccgg-wandb-metadata.json
51
+ 2024-04-11 12:00:15,169 INFO Thread-12 :472 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-metadata.json
52
+ 2024-04-11 12:00:15,245 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 12:00:15,245 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 12:00:15,248 DEBUG SenderThread:472 [sender.py:send():379] send: telemetry
55
+ 2024-04-11 12:00:15,258 DEBUG SenderThread:472 [sender.py:send():379] send: config
56
+ 2024-04-11 12:00:15,260 DEBUG SenderThread:472 [sender.py:send():379] send: metric
57
+ 2024-04-11 12:00:15,260 DEBUG SenderThread:472 [sender.py:send():379] send: telemetry
58
+ 2024-04-11 12:00:15,260 DEBUG SenderThread:472 [sender.py:send():379] send: metric
59
+ 2024-04-11 12:00:15,261 WARNING SenderThread:472 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-11 12:00:15,261 DEBUG SenderThread:472 [sender.py:send():379] send: telemetry
61
+ 2024-04-11 12:00:15,262 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
62
+ 2024-04-11 12:00:15,262 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
63
+ 2024-04-11 12:00:15,266 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
64
+ 2024-04-11 12:00:16,170 INFO Thread-12 :472 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
65
+ 2024-04-11 12:00:16,170 INFO Thread-12 :472 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/requirements.txt
66
+ 2024-04-11 12:00:18,171 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
67
+ 2024-04-11 12:00:19,818 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 12:00:20,171 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
69
+ 2024-04-11 12:00:22,172 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
70
+ 2024-04-11 12:00:24,819 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 12:00:29,825 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
72
+ 2024-04-11 12:00:30,176 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/config.yaml
73
+ 2024-04-11 12:00:30,247 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
74
+ 2024-04-11 12:00:30,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
75
+ 2024-04-11 12:00:30,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
76
+ 2024-04-11 12:00:35,316 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-04-11 12:00:38,179 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
78
+ 2024-04-11 12:00:40,779 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-04-11 12:00:45,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
80
+ 2024-04-11 12:00:45,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
81
+ 2024-04-11 12:00:45,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
82
+ 2024-04-11 12:00:46,376 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 12:00:51,377 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
84
+ 2024-04-11 12:00:57,292 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
85
+ 2024-04-11 12:00:58,187 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
86
+ 2024-04-11 12:00:59,282 DEBUG SystemMonitor:472 [system_monitor.py:_start():172] Starting system metrics aggregation loop
87
+ 2024-04-11 12:00:59,284 DEBUG SenderThread:472 [sender.py:send():379] send: stats
88
+ 2024-04-11 12:01:00,246 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-04-11 12:01:00,247 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
90
+ 2024-04-11 12:01:00,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-04-11 12:01:02,314 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-04-11 12:01:07,315 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-04-11 12:01:12,403 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-04-11 12:01:14,193 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
95
+ 2024-04-11 12:01:15,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
96
+ 2024-04-11 12:01:15,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
97
+ 2024-04-11 12:01:15,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-04-11 12:01:18,373 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-11 12:01:23,374 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-04-11 12:01:24,481 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: pause
101
+ 2024-04-11 12:01:24,481 INFO HandlerThread:472 [handler.py:handle_request_pause():708] stopping system metrics thread
102
+ 2024-04-11 12:01:24,481 INFO HandlerThread:472 [system_monitor.py:finish():203] Stopping system monitor
103
+ 2024-04-11 12:01:24,481 DEBUG SystemMonitor:472 [system_monitor.py:_start():179] Finished system metrics aggregation loop
104
+ 2024-04-11 12:01:24,482 DEBUG SystemMonitor:472 [system_monitor.py:_start():183] Publishing last batch of metrics
105
+ 2024-04-11 12:01:24,482 INFO HandlerThread:472 [interfaces.py:finish():202] Joined cpu monitor
106
+ 2024-04-11 12:01:24,483 INFO HandlerThread:472 [interfaces.py:finish():202] Joined disk monitor
107
+ 2024-04-11 12:01:24,494 INFO HandlerThread:472 [interfaces.py:finish():202] Joined gpu monitor
108
+ 2024-04-11 12:01:24,494 INFO HandlerThread:472 [interfaces.py:finish():202] Joined memory monitor
109
+ 2024-04-11 12:01:24,494 INFO HandlerThread:472 [interfaces.py:finish():202] Joined network monitor
110
+ 2024-04-11 12:01:24,494 DEBUG SenderThread:472 [sender.py:send():379] send: stats
111
+ 2024-04-11 12:01:28,495 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-04-11 12:01:30,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
113
+ 2024-04-11 12:01:30,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
114
+ 2024-04-11 12:01:30,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
115
+ 2024-04-11 12:01:34,384 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 12:01:39,385 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
117
+ 2024-04-11 12:01:44,386 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
118
+ 2024-04-11 12:01:45,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
119
+ 2024-04-11 12:01:45,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
120
+ 2024-04-11 12:01:45,286 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
121
+ 2024-04-11 12:01:50,287 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-11 12:01:55,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
123
+ 2024-04-11 12:02:00,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
124
+ 2024-04-11 12:02:00,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
125
+ 2024-04-11 12:02:00,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
126
+ 2024-04-11 12:02:00,374 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-04-11 12:02:05,375 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-04-11 12:02:10,376 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-04-11 12:02:15,248 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
130
+ 2024-04-11 12:02:15,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
131
+ 2024-04-11 12:02:15,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
132
+ 2024-04-11 12:02:16,306 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
133
+ 2024-04-11 12:02:21,307 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-04-11 12:02:26,307 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-04-11 12:02:30,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
136
+ 2024-04-11 12:02:30,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
137
+ 2024-04-11 12:02:30,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
138
+ 2024-04-11 12:02:31,328 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-04-11 12:02:36,329 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-04-11 12:02:41,330 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
141
+ 2024-04-11 12:02:45,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
142
+ 2024-04-11 12:02:45,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
143
+ 2024-04-11 12:02:45,290 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
144
+ 2024-04-11 12:02:46,368 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
145
+ 2024-04-11 12:02:51,370 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
146
+ 2024-04-11 12:02:56,370 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
147
+ 2024-04-11 12:03:00,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
148
+ 2024-04-11 12:03:00,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
149
+ 2024-04-11 12:03:00,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
150
+ 2024-04-11 12:03:02,305 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
151
+ 2024-04-11 12:03:07,305 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
152
+ 2024-04-11 12:03:12,306 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
153
+ 2024-04-11 12:03:15,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
154
+ 2024-04-11 12:03:15,249 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
155
+ 2024-04-11 12:03:15,290 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
156
+ 2024-04-11 12:03:17,314 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
157
+ 2024-04-11 12:03:22,315 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
158
+ 2024-04-11 12:03:27,317 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
159
+ 2024-04-11 12:03:30,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
160
+ 2024-04-11 12:03:30,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
161
+ 2024-04-11 12:03:30,290 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
162
+ 2024-04-11 12:03:32,344 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
163
+ 2024-04-11 12:03:37,345 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
164
+ 2024-04-11 12:03:42,345 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
165
+ 2024-04-11 12:03:45,249 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
166
+ 2024-04-11 12:03:45,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
167
+ 2024-04-11 12:03:45,290 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
168
+ 2024-04-11 12:03:48,331 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
169
+ 2024-04-11 12:03:53,332 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
170
+ 2024-04-11 12:03:58,333 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
171
+ 2024-04-11 12:04:00,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
172
+ 2024-04-11 12:04:00,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
173
+ 2024-04-11 12:04:00,290 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
174
+ 2024-04-11 12:04:04,304 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
175
+ 2024-04-11 12:04:09,305 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
176
+ 2024-04-11 12:04:14,306 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
177
+ 2024-04-11 12:04:15,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
178
+ 2024-04-11 12:04:15,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
179
+ 2024-04-11 12:04:15,291 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
180
+ 2024-04-11 12:04:20,293 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
181
+ 2024-04-11 12:04:25,294 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
182
+ 2024-04-11 12:04:30,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
183
+ 2024-04-11 12:04:30,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
184
+ 2024-04-11 12:04:30,291 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
185
+ 2024-04-11 12:04:30,318 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-04-11 12:04:35,320 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
187
+ 2024-04-11 12:04:40,320 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
188
+ 2024-04-11 12:04:45,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
189
+ 2024-04-11 12:04:45,250 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
190
+ 2024-04-11 12:04:45,291 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
191
+ 2024-04-11 12:04:45,793 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: resume
192
+ 2024-04-11 12:04:45,793 INFO HandlerThread:472 [handler.py:handle_request_resume():699] starting system metrics thread
193
+ 2024-04-11 12:04:45,793 INFO HandlerThread:472 [system_monitor.py:start():194] Starting system monitor
194
+ 2024-04-11 12:04:45,793 INFO SystemMonitor:472 [system_monitor.py:_start():158] Starting system asset monitoring threads
195
+ 2024-04-11 12:04:45,794 INFO SystemMonitor:472 [interfaces.py:start():190] Started cpu monitoring
196
+ 2024-04-11 12:04:45,794 INFO SystemMonitor:472 [interfaces.py:start():190] Started disk monitoring
197
+ 2024-04-11 12:04:45,795 INFO SystemMonitor:472 [interfaces.py:start():190] Started gpu monitoring
198
+ 2024-04-11 12:04:45,795 INFO SystemMonitor:472 [interfaces.py:start():190] Started memory monitoring
199
+ 2024-04-11 12:04:45,797 INFO SystemMonitor:472 [interfaces.py:start():190] Started network monitoring
200
+ 2024-04-11 12:04:46,308 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
201
+ 2024-04-11 12:04:51,309 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
202
+ 2024-04-11 12:04:56,310 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
203
+ 2024-04-11 12:05:00,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
204
+ 2024-04-11 12:05:00,251 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
205
+ 2024-04-11 12:05:00,253 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
206
+ 2024-04-11 12:05:01,348 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
207
+ 2024-04-11 12:05:06,349 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
208
+ 2024-04-11 12:05:11,349 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
209
+ 2024-04-11 12:05:15,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
210
+ 2024-04-11 12:05:15,251 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
211
+ 2024-04-11 12:05:15,291 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
212
+ 2024-04-11 12:05:17,338 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
213
+ 2024-04-11 12:05:22,339 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
214
+ 2024-04-11 12:05:27,340 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
215
+ 2024-04-11 12:05:30,250 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
216
+ 2024-04-11 12:05:30,251 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
217
+ 2024-04-11 12:05:30,291 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
218
+ 2024-04-11 12:05:32,357 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-04-11 12:05:37,358 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
220
+ 2024-04-11 12:05:42,359 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
221
+ 2024-04-11 12:05:45,251 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
222
+ 2024-04-11 12:05:45,251 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
223
+ 2024-04-11 12:05:45,292 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
224
+ 2024-04-11 12:05:45,797 DEBUG SystemMonitor:472 [system_monitor.py:_start():172] Starting system metrics aggregation loop
225
+ 2024-04-11 12:05:45,798 DEBUG SenderThread:472 [sender.py:send():379] send: stats
226
+ 2024-04-11 12:05:47,799 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
227
+ 2024-04-11 12:05:52,800 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
228
+ 2024-04-11 12:05:57,801 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
229
+ 2024-04-11 12:06:00,251 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
230
+ 2024-04-11 12:06:00,251 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
231
+ 2024-04-11 12:06:00,274 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
232
+ 2024-04-11 12:06:03,275 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
233
+ 2024-04-11 12:06:08,276 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
234
+ 2024-04-11 12:06:13,277 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
235
+ 2024-04-11 12:06:15,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
236
+ 2024-04-11 12:06:15,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
237
+ 2024-04-11 12:06:15,257 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
238
+ 2024-04-11 12:06:15,801 DEBUG SenderThread:472 [sender.py:send():379] send: stats
239
+ 2024-04-11 12:06:18,802 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
240
+ 2024-04-11 12:06:23,803 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
241
+ 2024-04-11 12:06:28,804 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
242
+ 2024-04-11 12:06:30,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
243
+ 2024-04-11 12:06:30,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
244
+ 2024-04-11 12:06:30,257 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
245
+ 2024-04-11 12:06:34,358 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
246
+ 2024-04-11 12:06:39,359 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-04-11 12:06:44,360 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
248
+ 2024-04-11 12:06:45,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
249
+ 2024-04-11 12:06:45,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
250
+ 2024-04-11 12:06:45,257 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
251
+ 2024-04-11 12:06:45,802 DEBUG SenderThread:472 [sender.py:send():379] send: stats
252
+ 2024-04-11 12:06:49,803 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
253
+ 2024-04-11 12:06:54,804 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
254
+ 2024-04-11 12:06:59,805 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
255
+ 2024-04-11 12:07:00,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
256
+ 2024-04-11 12:07:00,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
257
+ 2024-04-11 12:07:00,257 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
258
+ 2024-04-11 12:07:05,347 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
259
+ 2024-04-11 12:07:10,348 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
260
+ 2024-04-11 12:07:15,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
261
+ 2024-04-11 12:07:15,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
262
+ 2024-04-11 12:07:15,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
263
+ 2024-04-11 12:07:15,804 DEBUG SenderThread:472 [sender.py:send():379] send: stats
264
+ 2024-04-11 12:07:15,806 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
265
+ 2024-04-11 12:07:20,805 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
266
+ 2024-04-11 12:07:25,806 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
267
+ 2024-04-11 12:07:30,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
268
+ 2024-04-11 12:07:30,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
269
+ 2024-04-11 12:07:30,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
270
+ 2024-04-11 12:07:31,382 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
271
+ 2024-04-11 12:07:36,383 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
272
+ 2024-04-11 12:07:41,384 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
273
+ 2024-04-11 12:07:45,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
274
+ 2024-04-11 12:07:45,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
275
+ 2024-04-11 12:07:45,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
276
+ 2024-04-11 12:07:45,804 DEBUG SenderThread:472 [sender.py:send():379] send: stats
277
+ 2024-04-11 12:07:46,805 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
278
+ 2024-04-11 12:07:51,806 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
279
+ 2024-04-11 12:07:56,807 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
280
+ 2024-04-11 12:08:00,252 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
281
+ 2024-04-11 12:08:00,257 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
282
+ 2024-04-11 12:08:00,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
283
+ 2024-04-11 12:08:02,384 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
284
+ 2024-04-11 12:08:07,471 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
285
+ 2024-04-11 12:08:07,549 DEBUG SenderThread:472 [sender.py:send():379] send: config
286
+ 2024-04-11 12:08:07,552 DEBUG SenderThread:472 [sender.py:send():379] send: metric
287
+ 2024-04-11 12:08:07,552 DEBUG SenderThread:472 [sender.py:send():379] send: metric
288
+ 2024-04-11 12:08:07,552 WARNING SenderThread:472 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
289
+ 2024-04-11 12:08:08,360 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
290
+ 2024-04-11 12:08:10,361 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
291
+ 2024-04-11 12:08:12,915 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
292
+ 2024-04-11 12:08:15,253 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
293
+ 2024-04-11 12:08:15,258 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
294
+ 2024-04-11 12:08:15,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
295
+ 2024-04-11 12:08:15,805 DEBUG SenderThread:472 [sender.py:send():379] send: stats
296
+ 2024-04-11 12:08:18,812 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
297
+ 2024-04-11 12:08:19,364 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/config.yaml
298
+ 2024-04-11 12:08:23,957 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
299
+ 2024-04-11 12:08:26,367 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
300
+ 2024-04-11 12:08:29,061 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
301
+ 2024-04-11 12:08:30,253 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
302
+ 2024-04-11 12:08:30,258 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
303
+ 2024-04-11 12:08:30,258 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
304
+ 2024-04-11 12:08:34,293 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
305
+ 2024-04-11 12:08:40,133 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
306
+ 2024-04-11 12:08:40,372 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
307
+ 2024-04-11 12:08:45,134 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
308
+ 2024-04-11 12:08:45,254 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
309
+ 2024-04-11 12:08:45,258 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
310
+ 2024-04-11 12:08:45,259 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
311
+ 2024-04-11 12:08:45,807 DEBUG SenderThread:472 [sender.py:send():379] send: stats
312
+ 2024-04-11 12:08:50,808 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
313
+ 2024-04-11 12:08:55,809 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
314
+ 2024-04-11 12:09:00,254 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
315
+ 2024-04-11 12:09:00,260 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
316
+ 2024-04-11 12:09:00,261 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
317
+ 2024-04-11 12:09:01,307 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
318
+ 2024-04-11 12:09:06,307 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-04-11 12:09:07,071 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: partial_history
320
+ 2024-04-11 12:09:07,073 DEBUG SenderThread:472 [sender.py:send():379] send: metric
321
+ 2024-04-11 12:09:07,074 DEBUG SenderThread:472 [sender.py:send():379] send: metric
322
+ 2024-04-11 12:09:07,074 DEBUG SenderThread:472 [sender.py:send():379] send: metric
323
+ 2024-04-11 12:09:07,074 DEBUG SenderThread:472 [sender.py:send():379] send: metric
324
+ 2024-04-11 12:09:07,074 DEBUG SenderThread:472 [sender.py:send():379] send: history
325
+ 2024-04-11 12:09:07,074 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: summary_record
326
+ 2024-04-11 12:09:07,074 INFO SenderThread:472 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
327
+ 2024-04-11 12:09:07,383 INFO Thread-12 :472 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json
328
+ 2024-04-11 12:09:08,383 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
329
+ 2024-04-11 12:09:11,449 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
330
+ 2024-04-11 12:09:12,385 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
331
+ 2024-04-11 12:09:14,386 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
332
+ 2024-04-11 12:09:15,253 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
333
+ 2024-04-11 12:09:15,258 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
334
+ 2024-04-11 12:09:15,259 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
335
+ 2024-04-11 12:09:15,808 DEBUG SenderThread:472 [sender.py:send():379] send: stats
336
+ 2024-04-11 12:09:16,808 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
337
+ 2024-04-11 12:09:21,814 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
338
+ 2024-04-11 12:09:22,389 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/config.yaml
339
+ 2024-04-11 12:09:26,926 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
340
+ 2024-04-11 12:09:30,254 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
341
+ 2024-04-11 12:09:30,258 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: stop_status
342
+ 2024-04-11 12:09:30,259 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: stop_status
343
+ 2024-04-11 12:09:32,310 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
344
+ 2024-04-11 12:09:34,393 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
345
+ 2024-04-11 12:09:36,289 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: pause
346
+ 2024-04-11 12:09:36,290 INFO HandlerThread:472 [handler.py:handle_request_pause():708] stopping system metrics thread
347
+ 2024-04-11 12:09:36,290 INFO HandlerThread:472 [system_monitor.py:finish():203] Stopping system monitor
348
+ 2024-04-11 12:09:36,290 INFO HandlerThread:472 [interfaces.py:finish():202] Joined cpu monitor
349
+ 2024-04-11 12:09:36,291 INFO HandlerThread:472 [interfaces.py:finish():202] Joined disk monitor
350
+ 2024-04-11 12:09:36,291 DEBUG SystemMonitor:472 [system_monitor.py:_start():179] Finished system metrics aggregation loop
351
+ 2024-04-11 12:09:36,291 DEBUG SystemMonitor:472 [system_monitor.py:_start():183] Publishing last batch of metrics
352
+ 2024-04-11 12:09:36,302 INFO HandlerThread:472 [interfaces.py:finish():202] Joined gpu monitor
353
+ 2024-04-11 12:09:36,302 INFO HandlerThread:472 [interfaces.py:finish():202] Joined memory monitor
354
+ 2024-04-11 12:09:36,302 INFO HandlerThread:472 [interfaces.py:finish():202] Joined network monitor
355
+ 2024-04-11 12:09:36,303 DEBUG SenderThread:472 [sender.py:send():379] send: stats
356
+ 2024-04-11 12:09:36,394 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
357
+ 2024-04-11 12:09:38,304 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
358
+ 2024-04-11 12:09:40,061 DEBUG SenderThread:472 [sender.py:send():379] send: exit
359
+ 2024-04-11 12:09:40,061 INFO SenderThread:472 [sender.py:send_exit():586] handling exit code: 0
360
+ 2024-04-11 12:09:40,061 INFO SenderThread:472 [sender.py:send_exit():588] handling runtime: 375
361
+ 2024-04-11 12:09:40,061 INFO SenderThread:472 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
362
+ 2024-04-11 12:09:40,062 INFO SenderThread:472 [sender.py:send_exit():594] send defer
363
+ 2024-04-11 12:09:40,062 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
364
+ 2024-04-11 12:09:40,062 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 0
365
+ 2024-04-11 12:09:40,062 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
366
+ 2024-04-11 12:09:40,062 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 0
367
+ 2024-04-11 12:09:40,062 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 1
368
+ 2024-04-11 12:09:40,063 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
369
+ 2024-04-11 12:09:40,063 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 1
370
+ 2024-04-11 12:09:40,063 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
371
+ 2024-04-11 12:09:40,063 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 1
372
+ 2024-04-11 12:09:40,063 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 2
373
+ 2024-04-11 12:09:40,063 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
374
+ 2024-04-11 12:09:40,063 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 2
375
+ 2024-04-11 12:09:40,063 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
376
+ 2024-04-11 12:09:40,063 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 2
377
+ 2024-04-11 12:09:40,063 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 3
378
+ 2024-04-11 12:09:40,063 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
379
+ 2024-04-11 12:09:40,064 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 3
380
+ 2024-04-11 12:09:40,064 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
381
+ 2024-04-11 12:09:40,064 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 3
382
+ 2024-04-11 12:09:40,064 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 4
383
+ 2024-04-11 12:09:40,064 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
384
+ 2024-04-11 12:09:40,064 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 4
385
+ 2024-04-11 12:09:40,064 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
386
+ 2024-04-11 12:09:40,064 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 4
387
+ 2024-04-11 12:09:40,064 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 5
388
+ 2024-04-11 12:09:40,064 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
389
+ 2024-04-11 12:09:40,064 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 5
390
+ 2024-04-11 12:09:40,065 DEBUG SenderThread:472 [sender.py:send():379] send: summary
391
+ 2024-04-11 12:09:40,065 INFO SenderThread:472 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
392
+ 2024-04-11 12:09:40,065 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
393
+ 2024-04-11 12:09:40,065 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 5
394
+ 2024-04-11 12:09:40,065 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 6
395
+ 2024-04-11 12:09:40,066 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
396
+ 2024-04-11 12:09:40,066 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 6
397
+ 2024-04-11 12:09:40,066 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
398
+ 2024-04-11 12:09:40,066 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 6
399
+ 2024-04-11 12:09:40,066 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 7
400
+ 2024-04-11 12:09:40,066 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: status_report
401
+ 2024-04-11 12:09:40,066 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
402
+ 2024-04-11 12:09:40,066 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 7
403
+ 2024-04-11 12:09:40,067 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
404
+ 2024-04-11 12:09:40,067 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 7
405
+ 2024-04-11 12:09:40,395 INFO Thread-12 :472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json
406
+ 2024-04-11 12:09:41,061 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: poll_exit
407
+ 2024-04-11 12:09:41,992 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 8
408
+ 2024-04-11 12:09:41,992 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: poll_exit
409
+ 2024-04-11 12:09:41,992 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
410
+ 2024-04-11 12:09:41,993 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 8
411
+ 2024-04-11 12:09:41,993 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
412
+ 2024-04-11 12:09:41,993 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 8
413
+ 2024-04-11 12:09:41,993 INFO SenderThread:472 [job_builder.py:build():318] Attempting to build job artifact
414
+ 2024-04-11 12:09:42,034 INFO SenderThread:472 [job_builder.py:_get_source_type():466] no source found
415
+ 2024-04-11 12:09:42,034 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 9
416
+ 2024-04-11 12:09:42,034 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
417
+ 2024-04-11 12:09:42,034 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 9
418
+ 2024-04-11 12:09:42,035 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
419
+ 2024-04-11 12:09:42,035 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 9
420
+ 2024-04-11 12:09:42,035 INFO SenderThread:472 [dir_watcher.py:finish():358] shutting down directory watcher
421
+ 2024-04-11 12:09:42,061 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: poll_exit
422
+ 2024-04-11 12:09:42,397 INFO SenderThread:472 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
423
+ 2024-04-11 12:09:42,397 INFO SenderThread:472 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files
424
+ 2024-04-11 12:09:42,397 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-metadata.json wandb-metadata.json
425
+ 2024-04-11 12:09:42,397 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/conda-environment.yaml conda-environment.yaml
426
+ 2024-04-11 12:09:42,398 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log output.log
427
+ 2024-04-11 12:09:42,398 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json wandb-summary.json
428
+ 2024-04-11 12:09:42,402 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/config.yaml config.yaml
429
+ 2024-04-11 12:09:42,406 INFO SenderThread:472 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/requirements.txt requirements.txt
430
+ 2024-04-11 12:09:42,411 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 10
431
+ 2024-04-11 12:09:42,411 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: poll_exit
432
+ 2024-04-11 12:09:42,411 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
433
+ 2024-04-11 12:09:42,412 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 10
434
+ 2024-04-11 12:09:42,415 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
435
+ 2024-04-11 12:09:42,415 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 10
436
+ 2024-04-11 12:09:42,415 INFO SenderThread:472 [file_pusher.py:finish():172] shutting down file pusher
437
+ 2024-04-11 12:09:42,628 INFO wandb-upload_0:472 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/output.log
438
+ 2024-04-11 12:09:42,629 INFO wandb-upload_2:472 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/config.yaml
439
+ 2024-04-11 12:09:42,632 INFO wandb-upload_1:472 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/wandb-summary.json
440
+ 2024-04-11 12:09:42,702 INFO wandb-upload_3:472 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_115958-sm65qpc7/files/requirements.txt
441
+ 2024-04-11 12:09:42,903 INFO Thread-11 (_thread_body):472 [sender.py:transition_state():614] send defer: 11
442
+ 2024-04-11 12:09:42,903 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
443
+ 2024-04-11 12:09:42,903 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 11
444
+ 2024-04-11 12:09:42,904 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
445
+ 2024-04-11 12:09:42,904 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 11
446
+ 2024-04-11 12:09:42,904 INFO SenderThread:472 [file_pusher.py:join():178] waiting for file pusher
447
+ 2024-04-11 12:09:42,904 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 12
448
+ 2024-04-11 12:09:42,904 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
449
+ 2024-04-11 12:09:42,904 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 12
450
+ 2024-04-11 12:09:42,905 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
451
+ 2024-04-11 12:09:42,905 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 12
452
+ 2024-04-11 12:09:42,905 INFO SenderThread:472 [file_stream.py:finish():614] file stream finish called
453
+ 2024-04-11 12:09:42,989 INFO SenderThread:472 [file_stream.py:finish():618] file stream finish is done
454
+ 2024-04-11 12:09:42,989 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 13
455
+ 2024-04-11 12:09:42,990 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
456
+ 2024-04-11 12:09:42,990 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 13
457
+ 2024-04-11 12:09:42,990 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
458
+ 2024-04-11 12:09:42,990 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 13
459
+ 2024-04-11 12:09:42,990 INFO SenderThread:472 [sender.py:transition_state():614] send defer: 14
460
+ 2024-04-11 12:09:42,991 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: defer
461
+ 2024-04-11 12:09:42,991 INFO HandlerThread:472 [handler.py:handle_request_defer():172] handle defer: 14
462
+ 2024-04-11 12:09:42,991 DEBUG SenderThread:472 [sender.py:send():379] send: final
463
+ 2024-04-11 12:09:42,991 DEBUG SenderThread:472 [sender.py:send():379] send: footer
464
+ 2024-04-11 12:09:42,991 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: defer
465
+ 2024-04-11 12:09:42,991 INFO SenderThread:472 [sender.py:send_request_defer():610] handle sender defer: 14
466
+ 2024-04-11 12:09:42,992 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: poll_exit
467
+ 2024-04-11 12:09:42,992 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: poll_exit
468
+ 2024-04-11 12:09:42,993 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: poll_exit
469
+ 2024-04-11 12:09:42,993 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: poll_exit
470
+ 2024-04-11 12:09:42,994 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: server_info
471
+ 2024-04-11 12:09:42,994 DEBUG SenderThread:472 [sender.py:send_request():406] send_request: server_info
472
+ 2024-04-11 12:09:42,998 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: get_summary
473
+ 2024-04-11 12:09:42,999 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: sampled_history
474
+ 2024-04-11 12:09:42,999 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: internal_messages
475
+ 2024-04-11 12:09:43,015 INFO MainThread:472 [wandb_run.py:_footer_history_summary_info():3920] rendering history
476
+ 2024-04-11 12:09:43,015 INFO MainThread:472 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
477
+ 2024-04-11 12:09:43,015 INFO MainThread:472 [wandb_run.py:_footer_sync_info():3879] logging synced files
478
+ 2024-04-11 12:09:43,016 DEBUG HandlerThread:472 [handler.py:handle_request():146] handle_request: shutdown
479
+ 2024-04-11 12:09:43,016 INFO HandlerThread:472 [handler.py:finish():866] shutting down handler
480
+ 2024-04-11 12:09:43,994 INFO WriterThread:472 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_115958-sm65qpc7/run-sm65qpc7.wandb
481
+ 2024-04-11 12:09:44,015 INFO SenderThread:472 [sender.py:finish():1546] shutting down sender
482
+ 2024-04-11 12:09:44,015 INFO SenderThread:472 [file_pusher.py:finish():172] shutting down file pusher
483
+ 2024-04-11 12:09:44,015 INFO SenderThread:472 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_115958-sm65qpc7/logs/debug.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 11:59:58,687 INFO MainThread:427 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 11:59:58,687 INFO MainThread:427 [wandb_setup.py:_flush():76] Configure stats pid to 427
3
+ 2024-04-11 11:59:58,687 INFO MainThread:427 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 11:59:58,687 INFO MainThread:427 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 11:59:58,687 INFO MainThread:427 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_115958-sm65qpc7/logs/debug.log
10
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_115958-sm65qpc7/logs/debug-internal.log
11
+ 2024-04-11 11:59:58,688 INFO MainThread:427 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7a45231678e0>
12
+ 2024-04-11 11:59:58,689 INFO MainThread:427 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 11:59:58,689 INFO MainThread:427 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 11:59:58,689 INFO MainThread:427 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 11:59:58,689 INFO MainThread:427 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 11:59:58,691 INFO MainThread:427 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 11:59:58,692 INFO MainThread:427 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 11:59:58,706 INFO MainThread:427 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 11:59:59,018 INFO MainThread:427 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 11:59:59,021 INFO MainThread:427 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 11:59:59,173 INFO MainThread:427 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 11:59:59,260 INFO MainThread:427 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 11:59:59,260 INFO MainThread:427 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 12:00:15,246 INFO MainThread:427 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 12:00:15,246 INFO MainThread:427 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 12:00:15,246 INFO MainThread:427 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 12:00:15,246 INFO MainThread:427 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 12:00:15,247 INFO MainThread:427 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 12:00:15,253 INFO MainThread:427 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_11-52-16_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-11 12:01:24,480 INFO MainThread:427 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-11 12:01:24,480 INFO MainThread:427 [wandb_init.py:_pause_backend():438] pausing backend
34
+ 2024-04-11 12:04:45,792 INFO MainThread:427 [wandb_init.py:_resume_backend():443] resuming backend
35
+ 2024-04-11 12:08:07,539 INFO MainThread:427 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_12-04-45_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
36
+ 2024-04-11 12:09:36,289 INFO MainThread:427 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-04-11 12:09:36,289 INFO MainThread:427 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_115958-sm65qpc7/run-sm65qpc7.wandb ADDED
Binary file (53.9 kB). View file
 
wandb/run-20240411_121051-1qmakf52/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_121051-1qmakf52/files/config.yaml ADDED
@@ -0,0 +1,701 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712837451.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ vocab_size:
78
+ desc: null
79
+ value: 250880
80
+ hidden_size:
81
+ desc: null
82
+ value: 1536
83
+ n_layer:
84
+ desc: null
85
+ value: 24
86
+ n_head:
87
+ desc: null
88
+ value: 16
89
+ layer_norm_epsilon:
90
+ desc: null
91
+ value: 1.0e-05
92
+ initializer_range:
93
+ desc: null
94
+ value: 0.02
95
+ use_cache:
96
+ desc: null
97
+ value: false
98
+ pretraining_tp:
99
+ desc: null
100
+ value: 1
101
+ apply_residual_connection_post_layernorm:
102
+ desc: null
103
+ value: false
104
+ hidden_dropout:
105
+ desc: null
106
+ value: 0.0
107
+ attention_dropout:
108
+ desc: null
109
+ value: 0.0
110
+ bos_token_id:
111
+ desc: null
112
+ value: 1
113
+ eos_token_id:
114
+ desc: null
115
+ value: 2
116
+ slow_but_exact:
117
+ desc: null
118
+ value: false
119
+ return_dict:
120
+ desc: null
121
+ value: true
122
+ output_hidden_states:
123
+ desc: null
124
+ value: false
125
+ output_attentions:
126
+ desc: null
127
+ value: false
128
+ torchscript:
129
+ desc: null
130
+ value: false
131
+ torch_dtype:
132
+ desc: null
133
+ value: null
134
+ use_bfloat16:
135
+ desc: null
136
+ value: false
137
+ tf_legacy_loss:
138
+ desc: null
139
+ value: false
140
+ pruned_heads:
141
+ desc: null
142
+ value: {}
143
+ tie_word_embeddings:
144
+ desc: null
145
+ value: true
146
+ chunk_size_feed_forward:
147
+ desc: null
148
+ value: 0
149
+ is_encoder_decoder:
150
+ desc: null
151
+ value: false
152
+ is_decoder:
153
+ desc: null
154
+ value: false
155
+ cross_attention_hidden_size:
156
+ desc: null
157
+ value: null
158
+ add_cross_attention:
159
+ desc: null
160
+ value: false
161
+ tie_encoder_decoder:
162
+ desc: null
163
+ value: false
164
+ max_length:
165
+ desc: null
166
+ value: 20
167
+ min_length:
168
+ desc: null
169
+ value: 0
170
+ do_sample:
171
+ desc: null
172
+ value: false
173
+ early_stopping:
174
+ desc: null
175
+ value: false
176
+ num_beams:
177
+ desc: null
178
+ value: 1
179
+ num_beam_groups:
180
+ desc: null
181
+ value: 1
182
+ diversity_penalty:
183
+ desc: null
184
+ value: 0.0
185
+ temperature:
186
+ desc: null
187
+ value: 1.0
188
+ top_k:
189
+ desc: null
190
+ value: 50
191
+ top_p:
192
+ desc: null
193
+ value: 1.0
194
+ typical_p:
195
+ desc: null
196
+ value: 1.0
197
+ repetition_penalty:
198
+ desc: null
199
+ value: 1.0
200
+ length_penalty:
201
+ desc: null
202
+ value: 1.0
203
+ no_repeat_ngram_size:
204
+ desc: null
205
+ value: 0
206
+ encoder_no_repeat_ngram_size:
207
+ desc: null
208
+ value: 0
209
+ bad_words_ids:
210
+ desc: null
211
+ value: null
212
+ num_return_sequences:
213
+ desc: null
214
+ value: 1
215
+ output_scores:
216
+ desc: null
217
+ value: false
218
+ return_dict_in_generate:
219
+ desc: null
220
+ value: false
221
+ forced_bos_token_id:
222
+ desc: null
223
+ value: null
224
+ forced_eos_token_id:
225
+ desc: null
226
+ value: null
227
+ remove_invalid_values:
228
+ desc: null
229
+ value: false
230
+ exponential_decay_length_penalty:
231
+ desc: null
232
+ value: null
233
+ suppress_tokens:
234
+ desc: null
235
+ value: null
236
+ begin_suppress_tokens:
237
+ desc: null
238
+ value: null
239
+ architectures:
240
+ desc: null
241
+ value:
242
+ - BloomForCausalLM
243
+ finetuning_task:
244
+ desc: null
245
+ value: null
246
+ id2label:
247
+ desc: null
248
+ value:
249
+ '0': LABEL_0
250
+ '1': LABEL_1
251
+ label2id:
252
+ desc: null
253
+ value:
254
+ LABEL_0: 0
255
+ LABEL_1: 1
256
+ tokenizer_class:
257
+ desc: null
258
+ value: null
259
+ prefix:
260
+ desc: null
261
+ value: null
262
+ pad_token_id:
263
+ desc: null
264
+ value: 3
265
+ sep_token_id:
266
+ desc: null
267
+ value: null
268
+ decoder_start_token_id:
269
+ desc: null
270
+ value: null
271
+ task_specific_params:
272
+ desc: null
273
+ value: null
274
+ problem_type:
275
+ desc: null
276
+ value: null
277
+ _name_or_path:
278
+ desc: null
279
+ value: bigscience/bloomz-1b1
280
+ transformers_version:
281
+ desc: null
282
+ value: 4.39.3
283
+ attention_softmax_in_fp32:
284
+ desc: null
285
+ value: true
286
+ bias_dropout_fusion:
287
+ desc: null
288
+ value: true
289
+ unk_token_id:
290
+ desc: null
291
+ value: 0
292
+ masked_softmax_fusion:
293
+ desc: null
294
+ value: true
295
+ model_type:
296
+ desc: null
297
+ value: bloom
298
+ n_inner:
299
+ desc: null
300
+ value: null
301
+ offset_alibi:
302
+ desc: null
303
+ value: 100
304
+ seq_length:
305
+ desc: null
306
+ value: 2048
307
+ skip_bias_add:
308
+ desc: null
309
+ value: true
310
+ skip_bias_add_qkv:
311
+ desc: null
312
+ value: false
313
+ quantization_config:
314
+ desc: null
315
+ value:
316
+ quant_method: QuantizationMethod.BITS_AND_BYTES
317
+ _load_in_8bit: false
318
+ _load_in_4bit: true
319
+ llm_int8_threshold: 6.0
320
+ llm_int8_skip_modules: null
321
+ llm_int8_enable_fp32_cpu_offload: false
322
+ llm_int8_has_fp16_weight: false
323
+ bnb_4bit_quant_type: nf4
324
+ bnb_4bit_use_double_quant: false
325
+ bnb_4bit_compute_dtype: float16
326
+ bnb_4bit_quant_storage: uint8
327
+ load_in_4bit: true
328
+ load_in_8bit: false
329
+ output_dir:
330
+ desc: null
331
+ value: /kaggle/working/
332
+ overwrite_output_dir:
333
+ desc: null
334
+ value: false
335
+ do_train:
336
+ desc: null
337
+ value: false
338
+ do_eval:
339
+ desc: null
340
+ value: false
341
+ do_predict:
342
+ desc: null
343
+ value: false
344
+ evaluation_strategy:
345
+ desc: null
346
+ value: 'no'
347
+ prediction_loss_only:
348
+ desc: null
349
+ value: false
350
+ per_device_train_batch_size:
351
+ desc: null
352
+ value: 8
353
+ per_device_eval_batch_size:
354
+ desc: null
355
+ value: 8
356
+ per_gpu_train_batch_size:
357
+ desc: null
358
+ value: null
359
+ per_gpu_eval_batch_size:
360
+ desc: null
361
+ value: null
362
+ gradient_accumulation_steps:
363
+ desc: null
364
+ value: 1
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: null
368
+ eval_delay:
369
+ desc: null
370
+ value: 0
371
+ learning_rate:
372
+ desc: null
373
+ value: 3.0e-05
374
+ weight_decay:
375
+ desc: null
376
+ value: 0.001
377
+ adam_beta1:
378
+ desc: null
379
+ value: 0.9
380
+ adam_beta2:
381
+ desc: null
382
+ value: 0.999
383
+ adam_epsilon:
384
+ desc: null
385
+ value: 1.0e-08
386
+ max_grad_norm:
387
+ desc: null
388
+ value: 0.3
389
+ num_train_epochs:
390
+ desc: null
391
+ value: 5
392
+ max_steps:
393
+ desc: null
394
+ value: 18000
395
+ lr_scheduler_type:
396
+ desc: null
397
+ value: cosine
398
+ lr_scheduler_kwargs:
399
+ desc: null
400
+ value: {}
401
+ warmup_ratio:
402
+ desc: null
403
+ value: 0.03
404
+ warmup_steps:
405
+ desc: null
406
+ value: 0
407
+ log_level:
408
+ desc: null
409
+ value: passive
410
+ log_level_replica:
411
+ desc: null
412
+ value: warning
413
+ log_on_each_node:
414
+ desc: null
415
+ value: true
416
+ logging_dir:
417
+ desc: null
418
+ value: /kaggle/working/runs/Apr11_12-10-45_55cd788bb41a
419
+ logging_strategy:
420
+ desc: null
421
+ value: steps
422
+ logging_first_step:
423
+ desc: null
424
+ value: false
425
+ logging_steps:
426
+ desc: null
427
+ value: 20
428
+ logging_nan_inf_filter:
429
+ desc: null
430
+ value: true
431
+ save_strategy:
432
+ desc: null
433
+ value: steps
434
+ save_steps:
435
+ desc: null
436
+ value: 20
437
+ save_total_limit:
438
+ desc: null
439
+ value: 1
440
+ save_safetensors:
441
+ desc: null
442
+ value: true
443
+ save_on_each_node:
444
+ desc: null
445
+ value: false
446
+ save_only_model:
447
+ desc: null
448
+ value: false
449
+ no_cuda:
450
+ desc: null
451
+ value: false
452
+ use_cpu:
453
+ desc: null
454
+ value: false
455
+ use_mps_device:
456
+ desc: null
457
+ value: false
458
+ seed:
459
+ desc: null
460
+ value: 42
461
+ data_seed:
462
+ desc: null
463
+ value: null
464
+ jit_mode_eval:
465
+ desc: null
466
+ value: false
467
+ use_ipex:
468
+ desc: null
469
+ value: false
470
+ bf16:
471
+ desc: null
472
+ value: false
473
+ fp16:
474
+ desc: null
475
+ value: false
476
+ fp16_opt_level:
477
+ desc: null
478
+ value: O1
479
+ half_precision_backend:
480
+ desc: null
481
+ value: auto
482
+ bf16_full_eval:
483
+ desc: null
484
+ value: false
485
+ fp16_full_eval:
486
+ desc: null
487
+ value: false
488
+ tf32:
489
+ desc: null
490
+ value: null
491
+ local_rank:
492
+ desc: null
493
+ value: 0
494
+ ddp_backend:
495
+ desc: null
496
+ value: null
497
+ tpu_num_cores:
498
+ desc: null
499
+ value: null
500
+ tpu_metrics_debug:
501
+ desc: null
502
+ value: false
503
+ debug:
504
+ desc: null
505
+ value: []
506
+ dataloader_drop_last:
507
+ desc: null
508
+ value: false
509
+ eval_steps:
510
+ desc: null
511
+ value: null
512
+ dataloader_num_workers:
513
+ desc: null
514
+ value: 8
515
+ dataloader_prefetch_factor:
516
+ desc: null
517
+ value: null
518
+ past_index:
519
+ desc: null
520
+ value: -1
521
+ run_name:
522
+ desc: null
523
+ value: /kaggle/working/
524
+ disable_tqdm:
525
+ desc: null
526
+ value: false
527
+ remove_unused_columns:
528
+ desc: null
529
+ value: true
530
+ label_names:
531
+ desc: null
532
+ value: null
533
+ load_best_model_at_end:
534
+ desc: null
535
+ value: false
536
+ metric_for_best_model:
537
+ desc: null
538
+ value: null
539
+ greater_is_better:
540
+ desc: null
541
+ value: null
542
+ ignore_data_skip:
543
+ desc: null
544
+ value: false
545
+ fsdp:
546
+ desc: null
547
+ value: []
548
+ fsdp_min_num_params:
549
+ desc: null
550
+ value: 0
551
+ fsdp_config:
552
+ desc: null
553
+ value:
554
+ min_num_params: 0
555
+ xla: false
556
+ xla_fsdp_v2: false
557
+ xla_fsdp_grad_ckpt: false
558
+ fsdp_transformer_layer_cls_to_wrap:
559
+ desc: null
560
+ value: null
561
+ accelerator_config:
562
+ desc: null
563
+ value:
564
+ split_batches: false
565
+ dispatch_batches: null
566
+ even_batches: true
567
+ use_seedable_sampler: true
568
+ deepspeed:
569
+ desc: null
570
+ value: null
571
+ label_smoothing_factor:
572
+ desc: null
573
+ value: 0.0
574
+ optim:
575
+ desc: null
576
+ value: paged_adamw_32bit
577
+ optim_args:
578
+ desc: null
579
+ value: null
580
+ adafactor:
581
+ desc: null
582
+ value: false
583
+ group_by_length:
584
+ desc: null
585
+ value: false
586
+ length_column_name:
587
+ desc: null
588
+ value: length
589
+ report_to:
590
+ desc: null
591
+ value:
592
+ - tensorboard
593
+ - wandb
594
+ ddp_find_unused_parameters:
595
+ desc: null
596
+ value: null
597
+ ddp_bucket_cap_mb:
598
+ desc: null
599
+ value: null
600
+ ddp_broadcast_buffers:
601
+ desc: null
602
+ value: null
603
+ dataloader_pin_memory:
604
+ desc: null
605
+ value: true
606
+ dataloader_persistent_workers:
607
+ desc: null
608
+ value: false
609
+ skip_memory_metrics:
610
+ desc: null
611
+ value: true
612
+ use_legacy_prediction_loop:
613
+ desc: null
614
+ value: false
615
+ push_to_hub:
616
+ desc: null
617
+ value: false
618
+ resume_from_checkpoint:
619
+ desc: null
620
+ value: null
621
+ hub_model_id:
622
+ desc: null
623
+ value: null
624
+ hub_strategy:
625
+ desc: null
626
+ value: every_save
627
+ hub_token:
628
+ desc: null
629
+ value: <HUB_TOKEN>
630
+ hub_private_repo:
631
+ desc: null
632
+ value: false
633
+ hub_always_push:
634
+ desc: null
635
+ value: false
636
+ gradient_checkpointing:
637
+ desc: null
638
+ value: true
639
+ gradient_checkpointing_kwargs:
640
+ desc: null
641
+ value: null
642
+ include_inputs_for_metrics:
643
+ desc: null
644
+ value: false
645
+ fp16_backend:
646
+ desc: null
647
+ value: auto
648
+ push_to_hub_model_id:
649
+ desc: null
650
+ value: null
651
+ push_to_hub_organization:
652
+ desc: null
653
+ value: null
654
+ push_to_hub_token:
655
+ desc: null
656
+ value: <PUSH_TO_HUB_TOKEN>
657
+ mp_parameters:
658
+ desc: null
659
+ value: ''
660
+ auto_find_batch_size:
661
+ desc: null
662
+ value: true
663
+ full_determinism:
664
+ desc: null
665
+ value: false
666
+ torchdynamo:
667
+ desc: null
668
+ value: null
669
+ ray_scope:
670
+ desc: null
671
+ value: last
672
+ ddp_timeout:
673
+ desc: null
674
+ value: 1800
675
+ torch_compile:
676
+ desc: null
677
+ value: false
678
+ torch_compile_backend:
679
+ desc: null
680
+ value: null
681
+ torch_compile_mode:
682
+ desc: null
683
+ value: null
684
+ dispatch_batches:
685
+ desc: null
686
+ value: null
687
+ split_batches:
688
+ desc: null
689
+ value: null
690
+ include_tokens_per_second:
691
+ desc: null
692
+ value: false
693
+ include_num_input_tokens_seen:
694
+ desc: null
695
+ value: false
696
+ neftune_noise_alpha:
697
+ desc: null
698
+ value: null
699
+ optim_target_modules:
700
+ desc: null
701
+ value: null
wandb/run-20240411_121051-1qmakf52/files/output.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
6
+ warnings.warn('Was asked to gather along dimension 0, but all '
7
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
8
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
9
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
10
+ warnings.warn(
11
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
12
+ warnings.warn('Was asked to gather along dimension 0, but all '
13
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
14
+ warnings.warn(_create_warning_msg(
wandb/run-20240411_121051-1qmakf52/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_121051-1qmakf52/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T12:10:52.265357",
5
+ "startedAt": "2024-04-11T12:10:51.527492",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "55cd788bb41a",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.18,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.18,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.18,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.18,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.18,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5568.848327636719
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 6.2815, "train/grad_norm": 51.30440139770508, "train/learning_rate": 1.111111111111111e-06, "train/epoch": 0.0, "train/global_step": 20, "_timestamp": 1712837529.2155979, "_runtime": 77.68191194534302, "_step": 0, "_wandb": {"runtime": 117}}
wandb/run-20240411_121051-1qmakf52/logs/debug-internal.log ADDED
@@ -0,0 +1,269 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 12:10:51,534 INFO StreamThr :1038 [internal.py:wandb_internal():86] W&B internal server running at pid: 1038, started at: 2024-04-11 12:10:51.533844
2
+ 2024-04-11 12:10:51,536 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 12:10:51,980 INFO WriterThread:1038 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_121051-1qmakf52/run-1qmakf52.wandb
4
+ 2024-04-11 12:10:51,980 DEBUG SenderThread:1038 [sender.py:send():379] send: header
5
+ 2024-04-11 12:10:51,983 DEBUG SenderThread:1038 [sender.py:send():379] send: run
6
+ 2024-04-11 12:10:52,140 INFO SenderThread:1038 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files
7
+ 2024-04-11 12:10:52,140 INFO SenderThread:1038 [sender.py:_start_run_threads():1124] run started: 1qmakf52 with start time 1712837451.533686
8
+ 2024-04-11 12:10:52,148 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 12:10:52,148 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 12:10:52,239 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 12:10:52,251 DEBUG HandlerThread:1038 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 12:10:52,251 DEBUG HandlerThread:1038 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 12:10:52,251 INFO HandlerThread:1038 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 12:10:52,251 INFO SystemMonitor:1038 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 12:10:52,251 INFO HandlerThread:1038 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-11 12:10:52,252 INFO SystemMonitor:1038 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-11 12:10:52,252 INFO SystemMonitor:1038 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-11 12:10:52,253 INFO SystemMonitor:1038 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 12:10:52,255 INFO SystemMonitor:1038 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 12:10:52,255 INFO SystemMonitor:1038 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 12:10:52,265 DEBUG HandlerThread:1038 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 12:10:52,267 DEBUG HandlerThread:1038 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 12:10:52,267 DEBUG HandlerThread:1038 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 12:10:52,267 DEBUG HandlerThread:1038 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T12:10:52.265357', 'startedAt': '2024-04-11T12:10:51.527492', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '55cd788bb41a', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.18, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5568.848327636719}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 12:10:52,267 INFO HandlerThread:1038 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 12:10:52,267 INFO HandlerThread:1038 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 12:10:52,267 DEBUG HandlerThread:1038 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 12:10:53,142 INFO Thread-12 :1038 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/conda-environment.yaml
29
+ 2024-04-11 12:11:07,284 ERROR HandlerThread:1038 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 12:11:07,285 DEBUG HandlerThread:1038 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 12:11:07,286 INFO HandlerThread:1038 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 12:11:07,291 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 12:11:07,292 DEBUG SenderThread:1038 [sender.py:send():379] send: files
49
+ 2024-04-11 12:11:07,292 INFO SenderThread:1038 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 12:11:07,506 INFO wandb-upload_0:1038 [upload_job.py:push():131] Uploaded file /tmp/tmpl_gzxit6wandb/fhtouylb-wandb-metadata.json
51
+ 2024-04-11 12:11:08,146 INFO Thread-12 :1038 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-metadata.json
52
+ 2024-04-11 12:11:08,211 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 12:11:08,211 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 12:11:08,214 DEBUG SenderThread:1038 [sender.py:send():379] send: telemetry
55
+ 2024-04-11 12:11:08,214 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
56
+ 2024-04-11 12:11:08,215 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
57
+ 2024-04-11 12:11:08,226 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
58
+ 2024-04-11 12:11:08,352 DEBUG SenderThread:1038 [sender.py:send():379] send: config
59
+ 2024-04-11 12:11:08,355 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
60
+ 2024-04-11 12:11:08,355 DEBUG SenderThread:1038 [sender.py:send():379] send: telemetry
61
+ 2024-04-11 12:11:08,355 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
62
+ 2024-04-11 12:11:08,355 WARNING SenderThread:1038 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
63
+ 2024-04-11 12:11:08,355 DEBUG SenderThread:1038 [sender.py:send():379] send: telemetry
64
+ 2024-04-11 12:11:09,146 INFO Thread-12 :1038 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
65
+ 2024-04-11 12:11:09,147 INFO Thread-12 :1038 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/requirements.txt
66
+ 2024-04-11 12:11:11,147 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
67
+ 2024-04-11 12:11:12,609 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 12:11:13,148 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
69
+ 2024-04-11 12:11:17,610 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-11 12:11:22,615 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 12:11:23,152 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/config.yaml
72
+ 2024-04-11 12:11:23,214 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
73
+ 2024-04-11 12:11:23,214 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
74
+ 2024-04-11 12:11:23,215 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
75
+ 2024-04-11 12:11:28,309 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-11 12:11:33,310 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-04-11 12:11:38,214 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
78
+ 2024-04-11 12:11:38,215 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
79
+ 2024-04-11 12:11:38,215 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
80
+ 2024-04-11 12:11:38,331 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-11 12:11:43,331 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-11 12:11:48,332 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 12:11:52,255 DEBUG SystemMonitor:1038 [system_monitor.py:_start():172] Starting system metrics aggregation loop
84
+ 2024-04-11 12:11:52,257 DEBUG SenderThread:1038 [sender.py:send():379] send: stats
85
+ 2024-04-11 12:11:53,213 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
86
+ 2024-04-11 12:11:53,213 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
87
+ 2024-04-11 12:11:53,242 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
88
+ 2024-04-11 12:11:54,243 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-04-11 12:11:59,244 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
90
+ 2024-04-11 12:12:04,245 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
91
+ 2024-04-11 12:12:08,220 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
92
+ 2024-04-11 12:12:08,220 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
93
+ 2024-04-11 12:12:08,261 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
94
+ 2024-04-11 12:12:09,216 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: partial_history
95
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
96
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
97
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
98
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send():379] send: metric
99
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send():379] send: history
100
+ 2024-04-11 12:12:09,218 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: summary_record
101
+ 2024-04-11 12:12:09,220 INFO SenderThread:1038 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
102
+ 2024-04-11 12:12:09,257 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
103
+ 2024-04-11 12:12:10,169 INFO Thread-12 :1038 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json
104
+ 2024-04-11 12:12:11,170 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
105
+ 2024-04-11 12:12:14,258 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-04-11 12:12:17,172 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
107
+ 2024-04-11 12:12:20,205 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-04-11 12:12:22,257 DEBUG SenderThread:1038 [sender.py:send():379] send: stats
109
+ 2024-04-11 12:12:23,213 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
110
+ 2024-04-11 12:12:23,213 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
111
+ 2024-04-11 12:12:23,216 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
112
+ 2024-04-11 12:12:25,254 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-04-11 12:12:26,176 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/config.yaml
114
+ 2024-04-11 12:12:30,356 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-04-11 12:12:35,357 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 12:12:38,213 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
117
+ 2024-04-11 12:12:38,214 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
118
+ 2024-04-11 12:12:38,217 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-04-11 12:12:39,181 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
120
+ 2024-04-11 12:12:41,321 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
121
+ 2024-04-11 12:12:46,322 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-11 12:12:49,689 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: pause
123
+ 2024-04-11 12:12:49,689 INFO HandlerThread:1038 [handler.py:handle_request_pause():708] stopping system metrics thread
124
+ 2024-04-11 12:12:49,689 INFO HandlerThread:1038 [system_monitor.py:finish():203] Stopping system monitor
125
+ 2024-04-11 12:12:49,689 INFO HandlerThread:1038 [interfaces.py:finish():202] Joined cpu monitor
126
+ 2024-04-11 12:12:49,690 INFO HandlerThread:1038 [interfaces.py:finish():202] Joined disk monitor
127
+ 2024-04-11 12:12:49,690 DEBUG SystemMonitor:1038 [system_monitor.py:_start():179] Finished system metrics aggregation loop
128
+ 2024-04-11 12:12:49,690 DEBUG SystemMonitor:1038 [system_monitor.py:_start():183] Publishing last batch of metrics
129
+ 2024-04-11 12:12:49,701 INFO HandlerThread:1038 [interfaces.py:finish():202] Joined gpu monitor
130
+ 2024-04-11 12:12:49,702 INFO HandlerThread:1038 [interfaces.py:finish():202] Joined memory monitor
131
+ 2024-04-11 12:12:49,702 INFO HandlerThread:1038 [interfaces.py:finish():202] Joined network monitor
132
+ 2024-04-11 12:12:49,702 DEBUG SenderThread:1038 [sender.py:send():379] send: stats
133
+ 2024-04-11 12:12:51,703 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-04-11 12:12:53,213 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
135
+ 2024-04-11 12:12:53,213 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
136
+ 2024-04-11 12:12:53,216 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
137
+ 2024-04-11 12:12:57,342 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-04-11 12:13:02,343 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-04-11 12:13:07,344 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-04-11 12:13:08,213 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: stop_status
141
+ 2024-04-11 12:13:08,213 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: stop_status
142
+ 2024-04-11 12:13:08,254 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
143
+ 2024-04-11 12:13:13,313 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
144
+ 2024-04-11 12:13:14,058 DEBUG SenderThread:1038 [sender.py:send():379] send: exit
145
+ 2024-04-11 12:13:14,058 INFO SenderThread:1038 [sender.py:send_exit():586] handling exit code: 0
146
+ 2024-04-11 12:13:14,058 INFO SenderThread:1038 [sender.py:send_exit():588] handling runtime: 117
147
+ 2024-04-11 12:13:14,061 INFO SenderThread:1038 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
148
+ 2024-04-11 12:13:14,061 INFO SenderThread:1038 [sender.py:send_exit():594] send defer
149
+ 2024-04-11 12:13:14,062 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
150
+ 2024-04-11 12:13:14,062 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 0
151
+ 2024-04-11 12:13:14,062 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
152
+ 2024-04-11 12:13:14,062 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 0
153
+ 2024-04-11 12:13:14,062 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 1
154
+ 2024-04-11 12:13:14,062 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
155
+ 2024-04-11 12:13:14,062 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 1
156
+ 2024-04-11 12:13:14,062 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
157
+ 2024-04-11 12:13:14,062 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 1
158
+ 2024-04-11 12:13:14,063 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 2
159
+ 2024-04-11 12:13:14,063 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
160
+ 2024-04-11 12:13:14,063 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 2
161
+ 2024-04-11 12:13:14,063 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
162
+ 2024-04-11 12:13:14,063 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 2
163
+ 2024-04-11 12:13:14,063 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 3
164
+ 2024-04-11 12:13:14,063 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
165
+ 2024-04-11 12:13:14,063 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 3
166
+ 2024-04-11 12:13:14,063 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
167
+ 2024-04-11 12:13:14,063 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 3
168
+ 2024-04-11 12:13:14,063 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 4
169
+ 2024-04-11 12:13:14,064 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
170
+ 2024-04-11 12:13:14,064 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 4
171
+ 2024-04-11 12:13:14,064 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
172
+ 2024-04-11 12:13:14,064 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 4
173
+ 2024-04-11 12:13:14,064 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 5
174
+ 2024-04-11 12:13:14,064 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
175
+ 2024-04-11 12:13:14,064 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 5
176
+ 2024-04-11 12:13:14,064 DEBUG SenderThread:1038 [sender.py:send():379] send: summary
177
+ 2024-04-11 12:13:14,065 INFO SenderThread:1038 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
178
+ 2024-04-11 12:13:14,065 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
179
+ 2024-04-11 12:13:14,065 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 5
180
+ 2024-04-11 12:13:14,065 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 6
181
+ 2024-04-11 12:13:14,065 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
182
+ 2024-04-11 12:13:14,065 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 6
183
+ 2024-04-11 12:13:14,065 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
184
+ 2024-04-11 12:13:14,065 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 6
185
+ 2024-04-11 12:13:14,066 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 7
186
+ 2024-04-11 12:13:14,066 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: status_report
187
+ 2024-04-11 12:13:14,066 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
188
+ 2024-04-11 12:13:14,066 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 7
189
+ 2024-04-11 12:13:14,066 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
190
+ 2024-04-11 12:13:14,066 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 7
191
+ 2024-04-11 12:13:14,193 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json
192
+ 2024-04-11 12:13:14,525 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 8
193
+ 2024-04-11 12:13:14,525 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
194
+ 2024-04-11 12:13:14,525 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 8
195
+ 2024-04-11 12:13:14,526 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
196
+ 2024-04-11 12:13:14,526 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 8
197
+ 2024-04-11 12:13:14,526 INFO SenderThread:1038 [job_builder.py:build():318] Attempting to build job artifact
198
+ 2024-04-11 12:13:14,528 INFO SenderThread:1038 [job_builder.py:_get_source_type():466] no source found
199
+ 2024-04-11 12:13:14,528 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 9
200
+ 2024-04-11 12:13:14,528 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
201
+ 2024-04-11 12:13:14,528 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 9
202
+ 2024-04-11 12:13:14,528 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
203
+ 2024-04-11 12:13:14,528 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 9
204
+ 2024-04-11 12:13:14,528 INFO SenderThread:1038 [dir_watcher.py:finish():358] shutting down directory watcher
205
+ 2024-04-11 12:13:15,058 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: poll_exit
206
+ 2024-04-11 12:13:15,193 INFO Thread-12 :1038 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
207
+ 2024-04-11 12:13:15,194 INFO SenderThread:1038 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files
208
+ 2024-04-11 12:13:15,194 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-metadata.json wandb-metadata.json
209
+ 2024-04-11 12:13:15,195 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/conda-environment.yaml conda-environment.yaml
210
+ 2024-04-11 12:13:15,195 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log output.log
211
+ 2024-04-11 12:13:15,195 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json wandb-summary.json
212
+ 2024-04-11 12:13:15,198 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/config.yaml config.yaml
213
+ 2024-04-11 12:13:15,198 INFO SenderThread:1038 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/requirements.txt requirements.txt
214
+ 2024-04-11 12:13:15,201 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 10
215
+ 2024-04-11 12:13:15,202 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: poll_exit
216
+ 2024-04-11 12:13:15,208 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
217
+ 2024-04-11 12:13:15,208 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 10
218
+ 2024-04-11 12:13:15,209 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
219
+ 2024-04-11 12:13:15,209 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 10
220
+ 2024-04-11 12:13:15,209 INFO SenderThread:1038 [file_pusher.py:finish():172] shutting down file pusher
221
+ 2024-04-11 12:13:15,357 INFO wandb-upload_1:1038 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/wandb-summary.json
222
+ 2024-04-11 12:13:15,407 INFO wandb-upload_0:1038 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/output.log
223
+ 2024-04-11 12:13:15,429 INFO wandb-upload_3:1038 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/requirements.txt
224
+ 2024-04-11 12:13:15,430 INFO wandb-upload_2:1038 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240411_121051-1qmakf52/files/config.yaml
225
+ 2024-04-11 12:13:15,630 INFO Thread-11 (_thread_body):1038 [sender.py:transition_state():614] send defer: 11
226
+ 2024-04-11 12:13:15,631 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
227
+ 2024-04-11 12:13:15,631 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 11
228
+ 2024-04-11 12:13:15,631 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
229
+ 2024-04-11 12:13:15,631 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 11
230
+ 2024-04-11 12:13:15,632 INFO SenderThread:1038 [file_pusher.py:join():178] waiting for file pusher
231
+ 2024-04-11 12:13:15,632 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 12
232
+ 2024-04-11 12:13:15,632 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
233
+ 2024-04-11 12:13:15,632 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 12
234
+ 2024-04-11 12:13:15,632 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
235
+ 2024-04-11 12:13:15,632 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 12
236
+ 2024-04-11 12:13:15,632 INFO SenderThread:1038 [file_stream.py:finish():614] file stream finish called
237
+ 2024-04-11 12:13:15,691 INFO SenderThread:1038 [file_stream.py:finish():618] file stream finish is done
238
+ 2024-04-11 12:13:15,691 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 13
239
+ 2024-04-11 12:13:15,691 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
240
+ 2024-04-11 12:13:15,691 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 13
241
+ 2024-04-11 12:13:15,691 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
242
+ 2024-04-11 12:13:15,692 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 13
243
+ 2024-04-11 12:13:15,692 INFO SenderThread:1038 [sender.py:transition_state():614] send defer: 14
244
+ 2024-04-11 12:13:15,692 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: defer
245
+ 2024-04-11 12:13:15,692 INFO HandlerThread:1038 [handler.py:handle_request_defer():172] handle defer: 14
246
+ 2024-04-11 12:13:15,692 DEBUG SenderThread:1038 [sender.py:send():379] send: final
247
+ 2024-04-11 12:13:15,692 DEBUG SenderThread:1038 [sender.py:send():379] send: footer
248
+ 2024-04-11 12:13:15,693 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: defer
249
+ 2024-04-11 12:13:15,693 INFO SenderThread:1038 [sender.py:send_request_defer():610] handle sender defer: 14
250
+ 2024-04-11 12:13:15,694 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: poll_exit
251
+ 2024-04-11 12:13:15,694 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: poll_exit
252
+ 2024-04-11 12:13:15,694 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: poll_exit
253
+ 2024-04-11 12:13:15,695 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: poll_exit
254
+ 2024-04-11 12:13:15,695 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: server_info
255
+ 2024-04-11 12:13:15,696 DEBUG SenderThread:1038 [sender.py:send_request():406] send_request: server_info
256
+ 2024-04-11 12:13:15,699 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: get_summary
257
+ 2024-04-11 12:13:15,699 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: sampled_history
258
+ 2024-04-11 12:13:15,700 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: internal_messages
259
+ 2024-04-11 12:13:15,727 INFO MainThread:1038 [wandb_run.py:_footer_history_summary_info():3920] rendering history
260
+ 2024-04-11 12:13:15,727 INFO MainThread:1038 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
261
+ 2024-04-11 12:13:15,727 INFO MainThread:1038 [wandb_run.py:_footer_sync_info():3879] logging synced files
262
+ 2024-04-11 12:13:15,728 DEBUG HandlerThread:1038 [handler.py:handle_request():146] handle_request: shutdown
263
+ 2024-04-11 12:13:15,728 INFO HandlerThread:1038 [handler.py:finish():866] shutting down handler
264
+ 2024-04-11 12:13:16,688 WARNING StreamThr :1038 [internal.py:is_dead():414] Internal process exiting, parent pid 999 disappeared
265
+ 2024-04-11 12:13:16,688 ERROR StreamThr :1038 [internal.py:wandb_internal():152] Internal process shutdown.
266
+ 2024-04-11 12:13:16,696 INFO WriterThread:1038 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240411_121051-1qmakf52/run-1qmakf52.wandb
267
+ 2024-04-11 12:13:16,727 INFO SenderThread:1038 [sender.py:finish():1546] shutting down sender
268
+ 2024-04-11 12:13:16,727 INFO SenderThread:1038 [file_pusher.py:finish():172] shutting down file pusher
269
+ 2024-04-11 12:13:16,727 INFO SenderThread:1038 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240411_121051-1qmakf52/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Configure stats pid to 999
3
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 12:10:51,529 INFO MainThread:999 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_121051-1qmakf52/logs/debug.log
10
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_121051-1qmakf52/logs/debug-internal.log
11
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7d327a0d55d0>
12
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 12:10:51,530 INFO MainThread:999 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 12:10:51,532 INFO MainThread:999 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 12:10:51,533 INFO MainThread:999 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 12:10:51,545 INFO MainThread:999 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 12:10:51,979 INFO MainThread:999 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 12:10:51,982 INFO MainThread:999 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 12:10:52,147 INFO MainThread:999 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 12:10:52,234 INFO MainThread:999 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 12:10:52,234 INFO MainThread:999 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 12:11:08,212 INFO MainThread:999 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 12:11:08,212 INFO MainThread:999 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 12:11:08,212 INFO MainThread:999 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 12:11:08,213 INFO MainThread:999 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 12:11:08,214 INFO MainThread:999 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 12:11:08,220 INFO MainThread:999 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 18000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_12-10-45_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-11 12:12:49,688 INFO MainThread:999 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-11 12:12:49,688 INFO MainThread:999 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240411_121051-1qmakf52/run-1qmakf52.wandb ADDED
Binary file (13.6 kB). View file
 
wandb/run-20240411_122441-3wiow4fo/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240411_122441-3wiow4fo/files/config.yaml ADDED
@@ -0,0 +1,690 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712838281.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ - 62
48
+ 4: 3.10.13
49
+ 5: 0.16.5
50
+ 6: 4.39.3
51
+ 8:
52
+ - 1
53
+ - 2
54
+ - 5
55
+ 9:
56
+ 1: transformers_trainer
57
+ 13: linux-x86_64
58
+ m:
59
+ - 1: train/global_step
60
+ 6:
61
+ - 3
62
+ - 1: train/epoch
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ vocab_size:
67
+ desc: null
68
+ value: 250880
69
+ hidden_size:
70
+ desc: null
71
+ value: 1024
72
+ n_layer:
73
+ desc: null
74
+ value: 24
75
+ n_head:
76
+ desc: null
77
+ value: 16
78
+ layer_norm_epsilon:
79
+ desc: null
80
+ value: 1.0e-05
81
+ initializer_range:
82
+ desc: null
83
+ value: 0.02
84
+ use_cache:
85
+ desc: null
86
+ value: false
87
+ pretraining_tp:
88
+ desc: null
89
+ value: 1
90
+ apply_residual_connection_post_layernorm:
91
+ desc: null
92
+ value: false
93
+ hidden_dropout:
94
+ desc: null
95
+ value: 0.0
96
+ attention_dropout:
97
+ desc: null
98
+ value: 0.0
99
+ bos_token_id:
100
+ desc: null
101
+ value: 1
102
+ eos_token_id:
103
+ desc: null
104
+ value: 2
105
+ slow_but_exact:
106
+ desc: null
107
+ value: false
108
+ return_dict:
109
+ desc: null
110
+ value: true
111
+ output_hidden_states:
112
+ desc: null
113
+ value: false
114
+ output_attentions:
115
+ desc: null
116
+ value: false
117
+ torchscript:
118
+ desc: null
119
+ value: false
120
+ torch_dtype:
121
+ desc: null
122
+ value: null
123
+ use_bfloat16:
124
+ desc: null
125
+ value: false
126
+ tf_legacy_loss:
127
+ desc: null
128
+ value: false
129
+ pruned_heads:
130
+ desc: null
131
+ value: {}
132
+ tie_word_embeddings:
133
+ desc: null
134
+ value: true
135
+ chunk_size_feed_forward:
136
+ desc: null
137
+ value: 0
138
+ is_encoder_decoder:
139
+ desc: null
140
+ value: false
141
+ is_decoder:
142
+ desc: null
143
+ value: false
144
+ cross_attention_hidden_size:
145
+ desc: null
146
+ value: null
147
+ add_cross_attention:
148
+ desc: null
149
+ value: false
150
+ tie_encoder_decoder:
151
+ desc: null
152
+ value: false
153
+ max_length:
154
+ desc: null
155
+ value: 20
156
+ min_length:
157
+ desc: null
158
+ value: 0
159
+ do_sample:
160
+ desc: null
161
+ value: false
162
+ early_stopping:
163
+ desc: null
164
+ value: false
165
+ num_beams:
166
+ desc: null
167
+ value: 1
168
+ num_beam_groups:
169
+ desc: null
170
+ value: 1
171
+ diversity_penalty:
172
+ desc: null
173
+ value: 0.0
174
+ temperature:
175
+ desc: null
176
+ value: 1.0
177
+ top_k:
178
+ desc: null
179
+ value: 50
180
+ top_p:
181
+ desc: null
182
+ value: 1.0
183
+ typical_p:
184
+ desc: null
185
+ value: 1.0
186
+ repetition_penalty:
187
+ desc: null
188
+ value: 1.0
189
+ length_penalty:
190
+ desc: null
191
+ value: 1.0
192
+ no_repeat_ngram_size:
193
+ desc: null
194
+ value: 0
195
+ encoder_no_repeat_ngram_size:
196
+ desc: null
197
+ value: 0
198
+ bad_words_ids:
199
+ desc: null
200
+ value: null
201
+ num_return_sequences:
202
+ desc: null
203
+ value: 1
204
+ output_scores:
205
+ desc: null
206
+ value: false
207
+ return_dict_in_generate:
208
+ desc: null
209
+ value: false
210
+ forced_bos_token_id:
211
+ desc: null
212
+ value: null
213
+ forced_eos_token_id:
214
+ desc: null
215
+ value: null
216
+ remove_invalid_values:
217
+ desc: null
218
+ value: false
219
+ exponential_decay_length_penalty:
220
+ desc: null
221
+ value: null
222
+ suppress_tokens:
223
+ desc: null
224
+ value: null
225
+ begin_suppress_tokens:
226
+ desc: null
227
+ value: null
228
+ architectures:
229
+ desc: null
230
+ value:
231
+ - BloomForCausalLM
232
+ finetuning_task:
233
+ desc: null
234
+ value: null
235
+ id2label:
236
+ desc: null
237
+ value:
238
+ '0': LABEL_0
239
+ '1': LABEL_1
240
+ label2id:
241
+ desc: null
242
+ value:
243
+ LABEL_0: 0
244
+ LABEL_1: 1
245
+ tokenizer_class:
246
+ desc: null
247
+ value: null
248
+ prefix:
249
+ desc: null
250
+ value: null
251
+ pad_token_id:
252
+ desc: null
253
+ value: 3
254
+ sep_token_id:
255
+ desc: null
256
+ value: null
257
+ decoder_start_token_id:
258
+ desc: null
259
+ value: null
260
+ task_specific_params:
261
+ desc: null
262
+ value: null
263
+ problem_type:
264
+ desc: null
265
+ value: null
266
+ _name_or_path:
267
+ desc: null
268
+ value: bigscience/bloomz-560m
269
+ transformers_version:
270
+ desc: null
271
+ value: 4.39.3
272
+ attention_softmax_in_fp32:
273
+ desc: null
274
+ value: true
275
+ bias_dropout_fusion:
276
+ desc: null
277
+ value: true
278
+ unk_token_id:
279
+ desc: null
280
+ value: 0
281
+ masked_softmax_fusion:
282
+ desc: null
283
+ value: true
284
+ model_type:
285
+ desc: null
286
+ value: bloom
287
+ n_inner:
288
+ desc: null
289
+ value: null
290
+ offset_alibi:
291
+ desc: null
292
+ value: 100
293
+ seq_length:
294
+ desc: null
295
+ value: 2048
296
+ skip_bias_add:
297
+ desc: null
298
+ value: true
299
+ skip_bias_add_qkv:
300
+ desc: null
301
+ value: false
302
+ quantization_config:
303
+ desc: null
304
+ value:
305
+ quant_method: QuantizationMethod.BITS_AND_BYTES
306
+ _load_in_8bit: false
307
+ _load_in_4bit: true
308
+ llm_int8_threshold: 6.0
309
+ llm_int8_skip_modules: null
310
+ llm_int8_enable_fp32_cpu_offload: false
311
+ llm_int8_has_fp16_weight: false
312
+ bnb_4bit_quant_type: nf4
313
+ bnb_4bit_use_double_quant: false
314
+ bnb_4bit_compute_dtype: float16
315
+ bnb_4bit_quant_storage: uint8
316
+ load_in_4bit: true
317
+ load_in_8bit: false
318
+ output_dir:
319
+ desc: null
320
+ value: /kaggle/working/
321
+ overwrite_output_dir:
322
+ desc: null
323
+ value: false
324
+ do_train:
325
+ desc: null
326
+ value: false
327
+ do_eval:
328
+ desc: null
329
+ value: false
330
+ do_predict:
331
+ desc: null
332
+ value: false
333
+ evaluation_strategy:
334
+ desc: null
335
+ value: 'no'
336
+ prediction_loss_only:
337
+ desc: null
338
+ value: false
339
+ per_device_train_batch_size:
340
+ desc: null
341
+ value: 8
342
+ per_device_eval_batch_size:
343
+ desc: null
344
+ value: 8
345
+ per_gpu_train_batch_size:
346
+ desc: null
347
+ value: null
348
+ per_gpu_eval_batch_size:
349
+ desc: null
350
+ value: null
351
+ gradient_accumulation_steps:
352
+ desc: null
353
+ value: 1
354
+ eval_accumulation_steps:
355
+ desc: null
356
+ value: null
357
+ eval_delay:
358
+ desc: null
359
+ value: 0
360
+ learning_rate:
361
+ desc: null
362
+ value: 3.0e-05
363
+ weight_decay:
364
+ desc: null
365
+ value: 0.0001
366
+ adam_beta1:
367
+ desc: null
368
+ value: 0.9
369
+ adam_beta2:
370
+ desc: null
371
+ value: 0.999
372
+ adam_epsilon:
373
+ desc: null
374
+ value: 1.0e-08
375
+ max_grad_norm:
376
+ desc: null
377
+ value: 0.3
378
+ num_train_epochs:
379
+ desc: null
380
+ value: 5
381
+ max_steps:
382
+ desc: null
383
+ value: 10
384
+ lr_scheduler_type:
385
+ desc: null
386
+ value: cosine
387
+ lr_scheduler_kwargs:
388
+ desc: null
389
+ value: {}
390
+ warmup_ratio:
391
+ desc: null
392
+ value: 0.03
393
+ warmup_steps:
394
+ desc: null
395
+ value: 0
396
+ log_level:
397
+ desc: null
398
+ value: passive
399
+ log_level_replica:
400
+ desc: null
401
+ value: warning
402
+ log_on_each_node:
403
+ desc: null
404
+ value: true
405
+ logging_dir:
406
+ desc: null
407
+ value: /kaggle/working/runs/Apr11_12-17-04_55cd788bb41a
408
+ logging_strategy:
409
+ desc: null
410
+ value: steps
411
+ logging_first_step:
412
+ desc: null
413
+ value: false
414
+ logging_steps:
415
+ desc: null
416
+ value: 20
417
+ logging_nan_inf_filter:
418
+ desc: null
419
+ value: true
420
+ save_strategy:
421
+ desc: null
422
+ value: steps
423
+ save_steps:
424
+ desc: null
425
+ value: 20
426
+ save_total_limit:
427
+ desc: null
428
+ value: 1
429
+ save_safetensors:
430
+ desc: null
431
+ value: true
432
+ save_on_each_node:
433
+ desc: null
434
+ value: false
435
+ save_only_model:
436
+ desc: null
437
+ value: false
438
+ no_cuda:
439
+ desc: null
440
+ value: false
441
+ use_cpu:
442
+ desc: null
443
+ value: false
444
+ use_mps_device:
445
+ desc: null
446
+ value: false
447
+ seed:
448
+ desc: null
449
+ value: 42
450
+ data_seed:
451
+ desc: null
452
+ value: null
453
+ jit_mode_eval:
454
+ desc: null
455
+ value: false
456
+ use_ipex:
457
+ desc: null
458
+ value: false
459
+ bf16:
460
+ desc: null
461
+ value: false
462
+ fp16:
463
+ desc: null
464
+ value: false
465
+ fp16_opt_level:
466
+ desc: null
467
+ value: O1
468
+ half_precision_backend:
469
+ desc: null
470
+ value: auto
471
+ bf16_full_eval:
472
+ desc: null
473
+ value: false
474
+ fp16_full_eval:
475
+ desc: null
476
+ value: false
477
+ tf32:
478
+ desc: null
479
+ value: null
480
+ local_rank:
481
+ desc: null
482
+ value: 0
483
+ ddp_backend:
484
+ desc: null
485
+ value: null
486
+ tpu_num_cores:
487
+ desc: null
488
+ value: null
489
+ tpu_metrics_debug:
490
+ desc: null
491
+ value: false
492
+ debug:
493
+ desc: null
494
+ value: []
495
+ dataloader_drop_last:
496
+ desc: null
497
+ value: false
498
+ eval_steps:
499
+ desc: null
500
+ value: null
501
+ dataloader_num_workers:
502
+ desc: null
503
+ value: 8
504
+ dataloader_prefetch_factor:
505
+ desc: null
506
+ value: null
507
+ past_index:
508
+ desc: null
509
+ value: -1
510
+ run_name:
511
+ desc: null
512
+ value: /kaggle/working/
513
+ disable_tqdm:
514
+ desc: null
515
+ value: false
516
+ remove_unused_columns:
517
+ desc: null
518
+ value: true
519
+ label_names:
520
+ desc: null
521
+ value: null
522
+ load_best_model_at_end:
523
+ desc: null
524
+ value: false
525
+ metric_for_best_model:
526
+ desc: null
527
+ value: null
528
+ greater_is_better:
529
+ desc: null
530
+ value: null
531
+ ignore_data_skip:
532
+ desc: null
533
+ value: false
534
+ fsdp:
535
+ desc: null
536
+ value: []
537
+ fsdp_min_num_params:
538
+ desc: null
539
+ value: 0
540
+ fsdp_config:
541
+ desc: null
542
+ value:
543
+ min_num_params: 0
544
+ xla: false
545
+ xla_fsdp_v2: false
546
+ xla_fsdp_grad_ckpt: false
547
+ fsdp_transformer_layer_cls_to_wrap:
548
+ desc: null
549
+ value: null
550
+ accelerator_config:
551
+ desc: null
552
+ value:
553
+ split_batches: false
554
+ dispatch_batches: null
555
+ even_batches: true
556
+ use_seedable_sampler: true
557
+ deepspeed:
558
+ desc: null
559
+ value: null
560
+ label_smoothing_factor:
561
+ desc: null
562
+ value: 0.0
563
+ optim:
564
+ desc: null
565
+ value: paged_adamw_32bit
566
+ optim_args:
567
+ desc: null
568
+ value: null
569
+ adafactor:
570
+ desc: null
571
+ value: false
572
+ group_by_length:
573
+ desc: null
574
+ value: false
575
+ length_column_name:
576
+ desc: null
577
+ value: length
578
+ report_to:
579
+ desc: null
580
+ value:
581
+ - tensorboard
582
+ - wandb
583
+ ddp_find_unused_parameters:
584
+ desc: null
585
+ value: null
586
+ ddp_bucket_cap_mb:
587
+ desc: null
588
+ value: null
589
+ ddp_broadcast_buffers:
590
+ desc: null
591
+ value: null
592
+ dataloader_pin_memory:
593
+ desc: null
594
+ value: true
595
+ dataloader_persistent_workers:
596
+ desc: null
597
+ value: false
598
+ skip_memory_metrics:
599
+ desc: null
600
+ value: true
601
+ use_legacy_prediction_loop:
602
+ desc: null
603
+ value: false
604
+ push_to_hub:
605
+ desc: null
606
+ value: false
607
+ resume_from_checkpoint:
608
+ desc: null
609
+ value: null
610
+ hub_model_id:
611
+ desc: null
612
+ value: null
613
+ hub_strategy:
614
+ desc: null
615
+ value: every_save
616
+ hub_token:
617
+ desc: null
618
+ value: <HUB_TOKEN>
619
+ hub_private_repo:
620
+ desc: null
621
+ value: false
622
+ hub_always_push:
623
+ desc: null
624
+ value: false
625
+ gradient_checkpointing:
626
+ desc: null
627
+ value: true
628
+ gradient_checkpointing_kwargs:
629
+ desc: null
630
+ value: null
631
+ include_inputs_for_metrics:
632
+ desc: null
633
+ value: false
634
+ fp16_backend:
635
+ desc: null
636
+ value: auto
637
+ push_to_hub_model_id:
638
+ desc: null
639
+ value: null
640
+ push_to_hub_organization:
641
+ desc: null
642
+ value: null
643
+ push_to_hub_token:
644
+ desc: null
645
+ value: <PUSH_TO_HUB_TOKEN>
646
+ mp_parameters:
647
+ desc: null
648
+ value: ''
649
+ auto_find_batch_size:
650
+ desc: null
651
+ value: true
652
+ full_determinism:
653
+ desc: null
654
+ value: false
655
+ torchdynamo:
656
+ desc: null
657
+ value: null
658
+ ray_scope:
659
+ desc: null
660
+ value: last
661
+ ddp_timeout:
662
+ desc: null
663
+ value: 1800
664
+ torch_compile:
665
+ desc: null
666
+ value: false
667
+ torch_compile_backend:
668
+ desc: null
669
+ value: null
670
+ torch_compile_mode:
671
+ desc: null
672
+ value: null
673
+ dispatch_batches:
674
+ desc: null
675
+ value: null
676
+ split_batches:
677
+ desc: null
678
+ value: null
679
+ include_tokens_per_second:
680
+ desc: null
681
+ value: false
682
+ include_num_input_tokens_seen:
683
+ desc: null
684
+ value: false
685
+ neftune_noise_alpha:
686
+ desc: null
687
+ value: null
688
+ optim_target_modules:
689
+ desc: null
690
+ value: null
wandb/run-20240411_122441-3wiow4fo/files/output.log ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
38
+ warnings.warn('Was asked to gather along dimension 0, but all '
39
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
40
+ To disable this warning, you can either:
41
+ - Avoid using `tokenizers` before the fork if possible
42
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
44
+ To disable this warning, you can either:
45
+ - Avoid using `tokenizers` before the fork if possible
46
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
47
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
48
+ To disable this warning, you can either:
49
+ - Avoid using `tokenizers` before the fork if possible
50
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
51
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
52
+ To disable this warning, you can either:
53
+ - Avoid using `tokenizers` before the fork if possible
54
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
55
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
56
+ To disable this warning, you can either:
57
+ - Avoid using `tokenizers` before the fork if possible
58
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
59
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
60
+ To disable this warning, you can either:
61
+ - Avoid using `tokenizers` before the fork if possible
62
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
63
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
64
+ To disable this warning, you can either:
65
+ - Avoid using `tokenizers` before the fork if possible
66
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
67
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
68
+ To disable this warning, you can either:
69
+ - Avoid using `tokenizers` before the fork if possible
70
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
71
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
72
+ To disable this warning, you can either:
73
+ - Avoid using `tokenizers` before the fork if possible
74
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
75
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
76
+ To disable this warning, you can either:
77
+ - Avoid using `tokenizers` before the fork if possible
78
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
79
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
80
+ To disable this warning, you can either:
81
+ - Avoid using `tokenizers` before the fork if possible
82
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
83
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
84
+ To disable this warning, you can either:
85
+ - Avoid using `tokenizers` before the fork if possible
86
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
87
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
88
+ To disable this warning, you can either:
89
+ - Avoid using `tokenizers` before the fork if possible
90
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
91
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
92
+ To disable this warning, you can either:
93
+ - Avoid using `tokenizers` before the fork if possible
94
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
95
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
96
+ To disable this warning, you can either:
97
+ - Avoid using `tokenizers` before the fork if possible
98
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
99
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
100
+ To disable this warning, you can either:
101
+ - Avoid using `tokenizers` before the fork if possible
102
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
wandb/run-20240411_122441-3wiow4fo/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240411_122441-3wiow4fo/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-11T12:24:41.715654",
5
+ "startedAt": "2024-04-11T12:24:41.032425",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "55cd788bb41a",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.18,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.18,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.18,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.18,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.18,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5571.194152832031
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240411_122441-3wiow4fo/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train_runtime": 12.6279, "train_samples_per_second": 3.168, "train_steps_per_second": 0.792, "total_flos": 19241066102784.0, "train_loss": 5.502705001831055, "train/epoch": 0.0, "train/global_step": 10, "_timestamp": 1712838333.068445, "_runtime": 52.02996587753296, "_step": 0}
wandb/run-20240411_122441-3wiow4fo/logs/debug-internal.log ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 12:24:41,039 INFO StreamThr :1468 [internal.py:wandb_internal():86] W&B internal server running at pid: 1468, started at: 2024-04-11 12:24:41.038680
2
+ 2024-04-11 12:24:41,041 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-11 12:24:41,424 INFO WriterThread:1468 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/run-3wiow4fo.wandb
4
+ 2024-04-11 12:24:41,425 DEBUG SenderThread:1468 [sender.py:send():379] send: header
5
+ 2024-04-11 12:24:41,433 DEBUG SenderThread:1468 [sender.py:send():379] send: run
6
+ 2024-04-11 12:24:41,575 INFO SenderThread:1468 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files
7
+ 2024-04-11 12:24:41,575 INFO SenderThread:1468 [sender.py:_start_run_threads():1124] run started: 3wiow4fo with start time 1712838281.038479
8
+ 2024-04-11 12:24:41,584 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-11 12:24:41,584 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-11 12:24:41,681 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-11 12:24:41,695 DEBUG HandlerThread:1468 [system_info.py:__init__():26] System info init
12
+ 2024-04-11 12:24:41,695 DEBUG HandlerThread:1468 [system_info.py:__init__():41] System info init done
13
+ 2024-04-11 12:24:41,695 INFO HandlerThread:1468 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-11 12:24:41,695 INFO SystemMonitor:1468 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-11 12:24:41,696 INFO HandlerThread:1468 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-11 12:24:41,696 INFO SystemMonitor:1468 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-11 12:24:41,697 INFO SystemMonitor:1468 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-11 12:24:41,699 INFO SystemMonitor:1468 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-11 12:24:41,700 INFO SystemMonitor:1468 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-11 12:24:41,701 INFO SystemMonitor:1468 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-11 12:24:41,715 DEBUG HandlerThread:1468 [system_info.py:probe():150] Probing system
22
+ 2024-04-11 12:24:41,717 DEBUG HandlerThread:1468 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-11 12:24:41,717 DEBUG HandlerThread:1468 [system_info.py:probe():198] Probing system done
24
+ 2024-04-11 12:24:41,717 DEBUG HandlerThread:1468 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-11T12:24:41.715654', 'startedAt': '2024-04-11T12:24:41.032425', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '55cd788bb41a', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.18, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}, {'current': 2000.18, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5571.194152832031}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-11 12:24:41,717 INFO HandlerThread:1468 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-11 12:24:41,718 INFO HandlerThread:1468 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-11 12:24:41,718 DEBUG HandlerThread:1468 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-11 12:24:42,577 INFO Thread-12 :1468 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/conda-environment.yaml
29
+ 2024-04-11 12:24:56,733 ERROR HandlerThread:1468 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-11 12:24:56,734 DEBUG HandlerThread:1468 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-11 12:24:56,734 INFO HandlerThread:1468 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-11 12:24:56,739 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-11 12:24:56,739 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-11 12:24:56,739 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-11 12:24:56,739 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-11 12:24:56,740 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-11 12:24:56,740 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-11 12:24:56,740 DEBUG SenderThread:1468 [sender.py:send():379] send: files
49
+ 2024-04-11 12:24:56,740 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-11 12:24:56,944 INFO wandb-upload_0:1468 [upload_job.py:push():131] Uploaded file /tmp/tmp67awq2ybwandb/dyzxjezo-wandb-metadata.json
51
+ 2024-04-11 12:24:57,580 INFO Thread-12 :1468 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/wandb-metadata.json
52
+ 2024-04-11 12:24:57,688 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-11 12:24:57,689 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-11 12:24:57,691 DEBUG SenderThread:1468 [sender.py:send():379] send: telemetry
55
+ 2024-04-11 12:24:57,701 DEBUG SenderThread:1468 [sender.py:send():379] send: config
56
+ 2024-04-11 12:24:57,703 DEBUG SenderThread:1468 [sender.py:send():379] send: metric
57
+ 2024-04-11 12:24:57,704 DEBUG SenderThread:1468 [sender.py:send():379] send: telemetry
58
+ 2024-04-11 12:24:57,704 DEBUG SenderThread:1468 [sender.py:send():379] send: metric
59
+ 2024-04-11 12:24:57,704 WARNING SenderThread:1468 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-11 12:24:57,704 DEBUG SenderThread:1468 [sender.py:send():379] send: telemetry
61
+ 2024-04-11 12:24:57,704 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
62
+ 2024-04-11 12:24:57,705 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
63
+ 2024-04-11 12:24:57,706 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-04-11 12:24:58,581 INFO Thread-12 :1468 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
65
+ 2024-04-11 12:24:58,581 INFO Thread-12 :1468 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/requirements.txt
66
+ 2024-04-11 12:25:00,581 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
67
+ 2024-04-11 12:25:01,973 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-11 12:25:02,582 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
69
+ 2024-04-11 12:25:06,974 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-11 12:25:11,989 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-04-11 12:25:12,586 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/config.yaml
72
+ 2024-04-11 12:25:12,691 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
73
+ 2024-04-11 12:25:12,691 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
74
+ 2024-04-11 12:25:12,692 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
75
+ 2024-04-11 12:25:14,587 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
76
+ 2024-04-11 12:25:17,774 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-04-11 12:25:22,589 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
78
+ 2024-04-11 12:25:22,814 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-04-11 12:25:27,691 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
80
+ 2024-04-11 12:25:27,691 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
81
+ 2024-04-11 12:25:27,691 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
82
+ 2024-04-11 12:25:28,752 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
83
+ 2024-04-11 12:25:33,068 DEBUG SenderThread:1468 [sender.py:send():379] send: telemetry
84
+ 2024-04-11 12:25:33,108 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: partial_history
85
+ 2024-04-11 12:25:33,110 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
86
+ 2024-04-11 12:25:33,112 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
87
+ 2024-04-11 12:25:33,112 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
88
+ 2024-04-11 12:25:33,112 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
89
+ 2024-04-11 12:25:33,113 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
90
+ 2024-04-11 12:25:33,113 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
91
+ 2024-04-11 12:25:33,113 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
92
+ 2024-04-11 12:25:33,113 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
93
+ 2024-04-11 12:25:33,114 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
94
+ 2024-04-11 12:25:33,114 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
95
+ 2024-04-11 12:25:33,114 DEBUG SenderThread:1468 [sender.py:send():379] send: metric
96
+ 2024-04-11 12:25:33,114 DEBUG SenderThread:1468 [sender.py:send():379] send: history
97
+ 2024-04-11 12:25:33,114 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: summary_record
98
+ 2024-04-11 12:25:33,114 INFO SenderThread:1468 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
99
+ 2024-04-11 12:25:33,594 INFO Thread-12 :1468 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/wandb-summary.json
100
+ 2024-04-11 12:25:34,249 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
101
+ 2024-04-11 12:25:34,594 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/output.log
102
+ 2024-04-11 12:25:39,250 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
103
+ 2024-04-11 12:25:41,701 DEBUG SystemMonitor:1468 [system_monitor.py:_start():172] Starting system metrics aggregation loop
104
+ 2024-04-11 12:25:41,703 DEBUG SenderThread:1468 [sender.py:send():379] send: stats
105
+ 2024-04-11 12:25:42,689 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
106
+ 2024-04-11 12:25:42,690 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
107
+ 2024-04-11 12:25:42,690 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
108
+ 2024-04-11 12:25:44,768 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
109
+ 2024-04-11 12:25:45,599 INFO Thread-12 :1468 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240411_122441-3wiow4fo/files/config.yaml
110
+ 2024-04-11 12:25:49,891 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
111
+ 2024-04-11 12:25:54,892 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-04-11 12:25:57,689 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
113
+ 2024-04-11 12:25:57,690 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
114
+ 2024-04-11 12:25:57,726 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
115
+ 2024-04-11 12:26:00,727 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-04-11 12:26:05,729 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
117
+ 2024-04-11 12:26:10,730 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
118
+ 2024-04-11 12:26:11,704 DEBUG SenderThread:1468 [sender.py:send():379] send: stats
119
+ 2024-04-11 12:26:12,690 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
120
+ 2024-04-11 12:26:12,690 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
121
+ 2024-04-11 12:26:12,731 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
122
+ 2024-04-11 12:26:15,819 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
123
+ 2024-04-11 12:26:20,820 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
124
+ 2024-04-11 12:26:25,820 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: status_report
125
+ 2024-04-11 12:26:27,690 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: stop_status
126
+ 2024-04-11 12:26:27,691 DEBUG SenderThread:1468 [sender.py:send_request():406] send_request: stop_status
127
+ 2024-04-11 12:26:27,726 DEBUG HandlerThread:1468 [handler.py:handle_request():146] handle_request: internal_messages
wandb/run-20240411_122441-3wiow4fo/logs/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Configure stats pid to 1425
3
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240411_122441-3wiow4fo/logs/debug.log
10
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240411_122441-3wiow4fo/logs/debug-internal.log
11
+ 2024-04-11 12:24:41,034 INFO MainThread:1425 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7bd820cc4d00>
12
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():617] starting backend
16
+ 2024-04-11 12:24:41,035 INFO MainThread:1425 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-11 12:24:41,036 INFO MainThread:1425 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-11 12:24:41,038 INFO MainThread:1425 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-11 12:24:41,050 INFO MainThread:1425 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-11 12:24:41,423 INFO MainThread:1425 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-11 12:24:41,432 INFO MainThread:1425 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-11 12:24:41,583 INFO MainThread:1425 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-11 12:24:41,671 INFO MainThread:1425 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-11 12:24:41,672 INFO MainThread:1425 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-11 12:24:57,689 INFO MainThread:1425 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-11 12:24:57,689 INFO MainThread:1425 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-11 12:24:57,690 INFO MainThread:1425 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-11 12:24:57,690 INFO MainThread:1425 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-11 12:24:57,691 INFO MainThread:1425 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-11 12:24:57,696 INFO MainThread:1425 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1024, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-560m', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr11_12-17-04_55cd788bb41a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240411_122441-3wiow4fo/run-3wiow4fo.wandb ADDED
Binary file (18.2 kB). View file