Femboyuwu2000 commited on
Commit
c59484b
1 Parent(s): 6db4657

lierotica_llama2

Browse files
Files changed (46) hide show
  1. .gitattributes +1 -0
  2. README.md +5 -3
  3. adapter_config.json +5 -5
  4. runs/Apr10_11-37-24_57bca537cf5f/events.out.tfevents.1712749319.57bca537cf5f.98.0 +3 -0
  5. runs/Apr10_11-44-26_57bca537cf5f/events.out.tfevents.1712749588.57bca537cf5f.248.0 +3 -0
  6. runs/Apr10_11-48-06_57bca537cf5f/events.out.tfevents.1712749816.57bca537cf5f.399.0 +3 -0
  7. runs/Apr10_20-58-11_57bca537cf5f/events.out.tfevents.1712782828.57bca537cf5f.556.0 +3 -0
  8. training_args.bin +1 -1
  9. wandb/debug-internal.log +0 -0
  10. wandb/debug.log +29 -29
  11. wandb/run-20240410_114243-a6hgqntd/files/conda-environment.yaml +0 -0
  12. wandb/run-20240410_114243-a6hgqntd/files/config.yaml +670 -0
  13. wandb/run-20240410_114243-a6hgqntd/files/output.log +36 -0
  14. wandb/run-20240410_114243-a6hgqntd/files/requirements.txt +864 -0
  15. wandb/run-20240410_114243-a6hgqntd/files/wandb-metadata.json +66 -0
  16. wandb/run-20240410_114243-a6hgqntd/files/wandb-summary.json +1 -0
  17. wandb/run-20240410_114243-a6hgqntd/logs/debug-internal.log +225 -0
  18. wandb/run-20240410_114243-a6hgqntd/logs/debug.log +34 -0
  19. wandb/run-20240410_114243-a6hgqntd/run-a6hgqntd.wandb +0 -0
  20. wandb/run-20240410_114629-hdrbjcbp/files/conda-environment.yaml +0 -0
  21. wandb/run-20240410_114629-hdrbjcbp/files/config.yaml +670 -0
  22. wandb/run-20240410_114629-hdrbjcbp/files/output.log +36 -0
  23. wandb/run-20240410_114629-hdrbjcbp/files/requirements.txt +864 -0
  24. wandb/run-20240410_114629-hdrbjcbp/files/wandb-metadata.json +66 -0
  25. wandb/run-20240410_114629-hdrbjcbp/files/wandb-summary.json +1 -0
  26. wandb/run-20240410_114629-hdrbjcbp/logs/debug-internal.log +223 -0
  27. wandb/run-20240410_114629-hdrbjcbp/logs/debug.log +33 -0
  28. wandb/run-20240410_114629-hdrbjcbp/run-hdrbjcbp.wandb +0 -0
  29. wandb/run-20240410_115017-r9p6sok2/files/conda-environment.yaml +0 -0
  30. wandb/run-20240410_115017-r9p6sok2/files/config.yaml +687 -0
  31. wandb/run-20240410_115017-r9p6sok2/files/output.log +234 -0
  32. wandb/run-20240410_115017-r9p6sok2/files/requirements.txt +864 -0
  33. wandb/run-20240410_115017-r9p6sok2/files/wandb-metadata.json +66 -0
  34. wandb/run-20240410_115017-r9p6sok2/files/wandb-summary.json +1 -0
  35. wandb/run-20240410_115017-r9p6sok2/logs/debug-internal.log +0 -0
  36. wandb/run-20240410_115017-r9p6sok2/logs/debug.log +33 -0
  37. wandb/run-20240410_115017-r9p6sok2/run-r9p6sok2.wandb +3 -0
  38. wandb/run-20240410_210030-g0qpa26d/files/conda-environment.yaml +0 -0
  39. wandb/run-20240410_210030-g0qpa26d/files/config.yaml +670 -0
  40. wandb/run-20240410_210030-g0qpa26d/files/output.log +35 -0
  41. wandb/run-20240410_210030-g0qpa26d/files/requirements.txt +864 -0
  42. wandb/run-20240410_210030-g0qpa26d/files/wandb-metadata.json +66 -0
  43. wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json +1 -0
  44. wandb/run-20240410_210030-g0qpa26d/logs/debug-internal.log +137 -0
  45. wandb/run-20240410_210030-g0qpa26d/logs/debug.log +31 -0
  46. wandb/run-20240410_210030-g0qpa26d/run-g0qpa26d.wandb +0 -0
.gitattributes CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  wandb/run-20240408_203054-ldiuneeg/run-ldiuneeg.wandb filter=lfs diff=lfs merge=lfs -text
37
  wandb/run-20240410_010250-ft4a6i5j/run-ft4a6i5j.wandb filter=lfs diff=lfs merge=lfs -text
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  wandb/run-20240408_203054-ldiuneeg/run-ldiuneeg.wandb filter=lfs diff=lfs merge=lfs -text
37
  wandb/run-20240410_010250-ft4a6i5j/run-ft4a6i5j.wandb filter=lfs diff=lfs merge=lfs -text
38
+ wandb/run-20240410_115017-r9p6sok2/run-r9p6sok2.wandb filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -35,14 +35,16 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 7e-06
39
  - train_batch_size: 8
40
  - eval_batch_size: 8
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
- - lr_scheduler_warmup_ratio: 0.03
45
- - training_steps: 15000
46
 
47
  ### Training results
48
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 3e-05
39
  - train_batch_size: 8
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
+ - lr_scheduler_warmup_ratio: 0.05
47
+ - training_steps: 10
48
 
49
  ### Training results
50
 
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_proj",
24
- "v_proj",
25
- "o_proj",
26
- "k_proj",
27
  "up_proj",
 
 
28
  "q_proj",
29
- "down_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
 
23
  "up_proj",
24
+ "gate_proj",
25
+ "down_proj",
26
  "q_proj",
27
+ "k_proj",
28
+ "o_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
runs/Apr10_11-37-24_57bca537cf5f/events.out.tfevents.1712749319.57bca537cf5f.98.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ae715d1c46a5f4083c0ad974b9a22fd1777ad165cea17aacb56a5d03534361
3
+ size 4184
runs/Apr10_11-44-26_57bca537cf5f/events.out.tfevents.1712749588.57bca537cf5f.248.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f0c4b5c57c4c4d0741e7f30b136cec4ba21ddf4b26866f42ded579102098ea8
3
+ size 4184
runs/Apr10_11-48-06_57bca537cf5f/events.out.tfevents.1712749816.57bca537cf5f.399.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e70d5a74b52710c835f413c363e75d887f2640f0092a71d43a1417597f6df7f
3
+ size 26593
runs/Apr10_20-58-11_57bca537cf5f/events.out.tfevents.1712782828.57bca537cf5f.556.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3388146ab5c836b6d12f20b5af24401e2c5b65ade739131f98cf5c55f9155cf4
3
+ size 5708
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:147058c240b6504dffc71c5e3cf4e64c5649a0006b05271d541e3a054520c963
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13609889d8491533770ceab667ce50ea137267ebb81ebecd751651b23e13c7ab
3
  size 4920
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,31 +1,31 @@
1
- 2024-04-10 01:02:50,456 INFO MainThread:483 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Configure stats pid to 483
3
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_010250-ft4a6i5j/logs/debug.log
10
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_010250-ft4a6i5j/logs/debug-internal.log
11
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78e1476822c0>
12
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:init():567] calling init triggers
13
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:init():617] starting backend
16
- 2024-04-10 01:02:50,457 INFO MainThread:483 [wandb_init.py:init():621] setting up manager
17
- 2024-04-10 01:02:50,459 INFO MainThread:483 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-10 01:02:50,460 INFO MainThread:483 [wandb_init.py:init():629] backend started and connected
19
- 2024-04-10 01:02:50,473 INFO MainThread:483 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
- 2024-04-10 01:02:50,810 INFO MainThread:483 [wandb_init.py:init():721] updated telemetry
21
- 2024-04-10 01:02:50,813 INFO MainThread:483 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
- 2024-04-10 01:02:51,049 INFO MainThread:483 [wandb_run.py:_on_init():2344] communicating current version
23
- 2024-04-10 01:02:51,115 INFO MainThread:483 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-10 01:02:51,115 INFO MainThread:483 [wandb_init.py:init():805] starting run threads in backend
26
- 2024-04-10 01:03:07,089 INFO MainThread:483 [wandb_run.py:_console_start():2323] atexit reg
27
- 2024-04-10 01:03:07,089 INFO MainThread:483 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
- 2024-04-10 01:03:07,091 INFO MainThread:483 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
- 2024-04-10 01:03:07,091 INFO MainThread:483 [wandb_run.py:_redirect():2268] Redirects installed.
30
- 2024-04-10 01:03:07,092 INFO MainThread:483 [wandb_init.py:init():848] run started, returning control to user process
31
- 2024-04-10 01:03:07,097 INFO MainThread:483 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 7e-06, 'weight_decay': 0.001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.07, 'num_train_epochs': 5, 'max_steps': 15000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_01-00-47_d91c9dc8354a', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 100, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_32bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
1
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Configure stats pid to 556
3
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug.log
10
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug-internal.log
11
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ce03cb61630>
12
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():617] starting backend
16
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-10 21:00:30,779 INFO MainThread:556 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-10 21:00:30,780 INFO MainThread:556 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-10 21:00:30,794 INFO MainThread:556 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-10 21:00:31,093 INFO MainThread:556 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-10 21:00:31,096 INFO MainThread:556 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-10 21:00:31,291 INFO MainThread:556 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-10 21:00:47,389 INFO MainThread:556 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-10 21:00:47,395 INFO MainThread:556 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_20-58-11_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_114243-a6hgqntd/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_114243-a6hgqntd/files/config.yaml ADDED
@@ -0,0 +1,670 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712749363.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ vocab_size:
62
+ desc: null
63
+ value: 32000
64
+ max_position_embeddings:
65
+ desc: null
66
+ value: 2048
67
+ hidden_size:
68
+ desc: null
69
+ value: 2048
70
+ intermediate_size:
71
+ desc: null
72
+ value: 5632
73
+ num_hidden_layers:
74
+ desc: null
75
+ value: 22
76
+ num_attention_heads:
77
+ desc: null
78
+ value: 32
79
+ num_key_value_heads:
80
+ desc: null
81
+ value: 4
82
+ hidden_act:
83
+ desc: null
84
+ value: silu
85
+ initializer_range:
86
+ desc: null
87
+ value: 0.02
88
+ rms_norm_eps:
89
+ desc: null
90
+ value: 1.0e-05
91
+ pretraining_tp:
92
+ desc: null
93
+ value: 1
94
+ use_cache:
95
+ desc: null
96
+ value: false
97
+ rope_theta:
98
+ desc: null
99
+ value: 10000.0
100
+ rope_scaling:
101
+ desc: null
102
+ value: null
103
+ attention_bias:
104
+ desc: null
105
+ value: false
106
+ attention_dropout:
107
+ desc: null
108
+ value: 0.0
109
+ return_dict:
110
+ desc: null
111
+ value: true
112
+ output_hidden_states:
113
+ desc: null
114
+ value: false
115
+ output_attentions:
116
+ desc: null
117
+ value: false
118
+ torchscript:
119
+ desc: null
120
+ value: false
121
+ torch_dtype:
122
+ desc: null
123
+ value: float32
124
+ use_bfloat16:
125
+ desc: null
126
+ value: false
127
+ tf_legacy_loss:
128
+ desc: null
129
+ value: false
130
+ pruned_heads:
131
+ desc: null
132
+ value: {}
133
+ tie_word_embeddings:
134
+ desc: null
135
+ value: false
136
+ chunk_size_feed_forward:
137
+ desc: null
138
+ value: 0
139
+ is_encoder_decoder:
140
+ desc: null
141
+ value: false
142
+ is_decoder:
143
+ desc: null
144
+ value: false
145
+ cross_attention_hidden_size:
146
+ desc: null
147
+ value: null
148
+ add_cross_attention:
149
+ desc: null
150
+ value: false
151
+ tie_encoder_decoder:
152
+ desc: null
153
+ value: false
154
+ max_length:
155
+ desc: null
156
+ value: 20
157
+ min_length:
158
+ desc: null
159
+ value: 0
160
+ do_sample:
161
+ desc: null
162
+ value: false
163
+ early_stopping:
164
+ desc: null
165
+ value: false
166
+ num_beams:
167
+ desc: null
168
+ value: 1
169
+ num_beam_groups:
170
+ desc: null
171
+ value: 1
172
+ diversity_penalty:
173
+ desc: null
174
+ value: 0.0
175
+ temperature:
176
+ desc: null
177
+ value: 1.0
178
+ top_k:
179
+ desc: null
180
+ value: 50
181
+ top_p:
182
+ desc: null
183
+ value: 1.0
184
+ typical_p:
185
+ desc: null
186
+ value: 1.0
187
+ repetition_penalty:
188
+ desc: null
189
+ value: 1.0
190
+ length_penalty:
191
+ desc: null
192
+ value: 1.0
193
+ no_repeat_ngram_size:
194
+ desc: null
195
+ value: 0
196
+ encoder_no_repeat_ngram_size:
197
+ desc: null
198
+ value: 0
199
+ bad_words_ids:
200
+ desc: null
201
+ value: null
202
+ num_return_sequences:
203
+ desc: null
204
+ value: 1
205
+ output_scores:
206
+ desc: null
207
+ value: false
208
+ return_dict_in_generate:
209
+ desc: null
210
+ value: false
211
+ forced_bos_token_id:
212
+ desc: null
213
+ value: null
214
+ forced_eos_token_id:
215
+ desc: null
216
+ value: null
217
+ remove_invalid_values:
218
+ desc: null
219
+ value: false
220
+ exponential_decay_length_penalty:
221
+ desc: null
222
+ value: null
223
+ suppress_tokens:
224
+ desc: null
225
+ value: null
226
+ begin_suppress_tokens:
227
+ desc: null
228
+ value: null
229
+ architectures:
230
+ desc: null
231
+ value:
232
+ - LlamaForCausalLM
233
+ finetuning_task:
234
+ desc: null
235
+ value: null
236
+ id2label:
237
+ desc: null
238
+ value:
239
+ '0': LABEL_0
240
+ '1': LABEL_1
241
+ label2id:
242
+ desc: null
243
+ value:
244
+ LABEL_0: 0
245
+ LABEL_1: 1
246
+ tokenizer_class:
247
+ desc: null
248
+ value: null
249
+ prefix:
250
+ desc: null
251
+ value: null
252
+ bos_token_id:
253
+ desc: null
254
+ value: 1
255
+ pad_token_id:
256
+ desc: null
257
+ value: null
258
+ eos_token_id:
259
+ desc: null
260
+ value: 2
261
+ sep_token_id:
262
+ desc: null
263
+ value: null
264
+ decoder_start_token_id:
265
+ desc: null
266
+ value: null
267
+ task_specific_params:
268
+ desc: null
269
+ value: null
270
+ problem_type:
271
+ desc: null
272
+ value: null
273
+ _name_or_path:
274
+ desc: null
275
+ value: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
276
+ transformers_version:
277
+ desc: null
278
+ value: 4.39.3
279
+ model_type:
280
+ desc: null
281
+ value: llama
282
+ quantization_config:
283
+ desc: null
284
+ value:
285
+ quant_method: QuantizationMethod.BITS_AND_BYTES
286
+ _load_in_8bit: false
287
+ _load_in_4bit: true
288
+ llm_int8_threshold: 6.0
289
+ llm_int8_skip_modules: null
290
+ llm_int8_enable_fp32_cpu_offload: false
291
+ llm_int8_has_fp16_weight: false
292
+ bnb_4bit_quant_type: nf4
293
+ bnb_4bit_use_double_quant: false
294
+ bnb_4bit_compute_dtype: float16
295
+ bnb_4bit_quant_storage: uint8
296
+ load_in_4bit: true
297
+ load_in_8bit: false
298
+ output_dir:
299
+ desc: null
300
+ value: /kaggle/working/
301
+ overwrite_output_dir:
302
+ desc: null
303
+ value: false
304
+ do_train:
305
+ desc: null
306
+ value: false
307
+ do_eval:
308
+ desc: null
309
+ value: false
310
+ do_predict:
311
+ desc: null
312
+ value: false
313
+ evaluation_strategy:
314
+ desc: null
315
+ value: 'no'
316
+ prediction_loss_only:
317
+ desc: null
318
+ value: false
319
+ per_device_train_batch_size:
320
+ desc: null
321
+ value: 8
322
+ per_device_eval_batch_size:
323
+ desc: null
324
+ value: 8
325
+ per_gpu_train_batch_size:
326
+ desc: null
327
+ value: null
328
+ per_gpu_eval_batch_size:
329
+ desc: null
330
+ value: null
331
+ gradient_accumulation_steps:
332
+ desc: null
333
+ value: 4
334
+ eval_accumulation_steps:
335
+ desc: null
336
+ value: null
337
+ eval_delay:
338
+ desc: null
339
+ value: 0
340
+ learning_rate:
341
+ desc: null
342
+ value: 3.0e-06
343
+ weight_decay:
344
+ desc: null
345
+ value: 0.0001
346
+ adam_beta1:
347
+ desc: null
348
+ value: 0.9
349
+ adam_beta2:
350
+ desc: null
351
+ value: 0.999
352
+ adam_epsilon:
353
+ desc: null
354
+ value: 1.0e-08
355
+ max_grad_norm:
356
+ desc: null
357
+ value: 0.01
358
+ num_train_epochs:
359
+ desc: null
360
+ value: 5
361
+ max_steps:
362
+ desc: null
363
+ value: 15000
364
+ lr_scheduler_type:
365
+ desc: null
366
+ value: cosine
367
+ lr_scheduler_kwargs:
368
+ desc: null
369
+ value: {}
370
+ warmup_ratio:
371
+ desc: null
372
+ value: 0.05
373
+ warmup_steps:
374
+ desc: null
375
+ value: 0
376
+ log_level:
377
+ desc: null
378
+ value: passive
379
+ log_level_replica:
380
+ desc: null
381
+ value: warning
382
+ log_on_each_node:
383
+ desc: null
384
+ value: true
385
+ logging_dir:
386
+ desc: null
387
+ value: /kaggle/working/runs/Apr10_11-37-24_57bca537cf5f
388
+ logging_strategy:
389
+ desc: null
390
+ value: steps
391
+ logging_first_step:
392
+ desc: null
393
+ value: false
394
+ logging_steps:
395
+ desc: null
396
+ value: 30
397
+ logging_nan_inf_filter:
398
+ desc: null
399
+ value: true
400
+ save_strategy:
401
+ desc: null
402
+ value: steps
403
+ save_steps:
404
+ desc: null
405
+ value: 30
406
+ save_total_limit:
407
+ desc: null
408
+ value: 1
409
+ save_safetensors:
410
+ desc: null
411
+ value: true
412
+ save_on_each_node:
413
+ desc: null
414
+ value: false
415
+ save_only_model:
416
+ desc: null
417
+ value: false
418
+ no_cuda:
419
+ desc: null
420
+ value: false
421
+ use_cpu:
422
+ desc: null
423
+ value: false
424
+ use_mps_device:
425
+ desc: null
426
+ value: false
427
+ seed:
428
+ desc: null
429
+ value: 42
430
+ data_seed:
431
+ desc: null
432
+ value: null
433
+ jit_mode_eval:
434
+ desc: null
435
+ value: false
436
+ use_ipex:
437
+ desc: null
438
+ value: false
439
+ bf16:
440
+ desc: null
441
+ value: false
442
+ fp16:
443
+ desc: null
444
+ value: false
445
+ fp16_opt_level:
446
+ desc: null
447
+ value: O1
448
+ half_precision_backend:
449
+ desc: null
450
+ value: auto
451
+ bf16_full_eval:
452
+ desc: null
453
+ value: false
454
+ fp16_full_eval:
455
+ desc: null
456
+ value: false
457
+ tf32:
458
+ desc: null
459
+ value: null
460
+ local_rank:
461
+ desc: null
462
+ value: 0
463
+ ddp_backend:
464
+ desc: null
465
+ value: null
466
+ tpu_num_cores:
467
+ desc: null
468
+ value: null
469
+ tpu_metrics_debug:
470
+ desc: null
471
+ value: false
472
+ debug:
473
+ desc: null
474
+ value: []
475
+ dataloader_drop_last:
476
+ desc: null
477
+ value: false
478
+ eval_steps:
479
+ desc: null
480
+ value: null
481
+ dataloader_num_workers:
482
+ desc: null
483
+ value: 8
484
+ dataloader_prefetch_factor:
485
+ desc: null
486
+ value: null
487
+ past_index:
488
+ desc: null
489
+ value: -1
490
+ run_name:
491
+ desc: null
492
+ value: /kaggle/working/
493
+ disable_tqdm:
494
+ desc: null
495
+ value: false
496
+ remove_unused_columns:
497
+ desc: null
498
+ value: true
499
+ label_names:
500
+ desc: null
501
+ value: null
502
+ load_best_model_at_end:
503
+ desc: null
504
+ value: false
505
+ metric_for_best_model:
506
+ desc: null
507
+ value: null
508
+ greater_is_better:
509
+ desc: null
510
+ value: null
511
+ ignore_data_skip:
512
+ desc: null
513
+ value: false
514
+ fsdp:
515
+ desc: null
516
+ value: []
517
+ fsdp_min_num_params:
518
+ desc: null
519
+ value: 0
520
+ fsdp_config:
521
+ desc: null
522
+ value:
523
+ min_num_params: 0
524
+ xla: false
525
+ xla_fsdp_v2: false
526
+ xla_fsdp_grad_ckpt: false
527
+ fsdp_transformer_layer_cls_to_wrap:
528
+ desc: null
529
+ value: null
530
+ accelerator_config:
531
+ desc: null
532
+ value:
533
+ split_batches: false
534
+ dispatch_batches: null
535
+ even_batches: true
536
+ use_seedable_sampler: true
537
+ deepspeed:
538
+ desc: null
539
+ value: null
540
+ label_smoothing_factor:
541
+ desc: null
542
+ value: 0.0
543
+ optim:
544
+ desc: null
545
+ value: paged_adamw_8bit
546
+ optim_args:
547
+ desc: null
548
+ value: null
549
+ adafactor:
550
+ desc: null
551
+ value: false
552
+ group_by_length:
553
+ desc: null
554
+ value: false
555
+ length_column_name:
556
+ desc: null
557
+ value: length
558
+ report_to:
559
+ desc: null
560
+ value:
561
+ - tensorboard
562
+ - wandb
563
+ ddp_find_unused_parameters:
564
+ desc: null
565
+ value: null
566
+ ddp_bucket_cap_mb:
567
+ desc: null
568
+ value: null
569
+ ddp_broadcast_buffers:
570
+ desc: null
571
+ value: null
572
+ dataloader_pin_memory:
573
+ desc: null
574
+ value: true
575
+ dataloader_persistent_workers:
576
+ desc: null
577
+ value: false
578
+ skip_memory_metrics:
579
+ desc: null
580
+ value: true
581
+ use_legacy_prediction_loop:
582
+ desc: null
583
+ value: false
584
+ push_to_hub:
585
+ desc: null
586
+ value: false
587
+ resume_from_checkpoint:
588
+ desc: null
589
+ value: null
590
+ hub_model_id:
591
+ desc: null
592
+ value: null
593
+ hub_strategy:
594
+ desc: null
595
+ value: every_save
596
+ hub_token:
597
+ desc: null
598
+ value: <HUB_TOKEN>
599
+ hub_private_repo:
600
+ desc: null
601
+ value: false
602
+ hub_always_push:
603
+ desc: null
604
+ value: false
605
+ gradient_checkpointing:
606
+ desc: null
607
+ value: true
608
+ gradient_checkpointing_kwargs:
609
+ desc: null
610
+ value: null
611
+ include_inputs_for_metrics:
612
+ desc: null
613
+ value: false
614
+ fp16_backend:
615
+ desc: null
616
+ value: auto
617
+ push_to_hub_model_id:
618
+ desc: null
619
+ value: null
620
+ push_to_hub_organization:
621
+ desc: null
622
+ value: null
623
+ push_to_hub_token:
624
+ desc: null
625
+ value: <PUSH_TO_HUB_TOKEN>
626
+ mp_parameters:
627
+ desc: null
628
+ value: ''
629
+ auto_find_batch_size:
630
+ desc: null
631
+ value: true
632
+ full_determinism:
633
+ desc: null
634
+ value: false
635
+ torchdynamo:
636
+ desc: null
637
+ value: null
638
+ ray_scope:
639
+ desc: null
640
+ value: last
641
+ ddp_timeout:
642
+ desc: null
643
+ value: 1800
644
+ torch_compile:
645
+ desc: null
646
+ value: false
647
+ torch_compile_backend:
648
+ desc: null
649
+ value: null
650
+ torch_compile_mode:
651
+ desc: null
652
+ value: null
653
+ dispatch_batches:
654
+ desc: null
655
+ value: null
656
+ split_batches:
657
+ desc: null
658
+ value: null
659
+ include_tokens_per_second:
660
+ desc: null
661
+ value: false
662
+ include_num_input_tokens_seen:
663
+ desc: null
664
+ value: false
665
+ neftune_noise_alpha:
666
+ desc: null
667
+ value: null
668
+ optim_target_modules:
669
+ desc: null
670
+ value: null
wandb/run-20240410_114243-a6hgqntd/files/output.log ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
wandb/run-20240410_114243-a6hgqntd/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240410_114243-a6hgqntd/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T11:42:43.806558",
5
+ "startedAt": "2024-04-10T11:42:43.180890",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "57bca537cf5f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5568.826057434082
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240410_114243-a6hgqntd/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 59}}
wandb/run-20240410_114243-a6hgqntd/logs/debug-internal.log ADDED
@@ -0,0 +1,225 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 11:42:43,188 INFO StreamThr :144 [internal.py:wandb_internal():86] W&B internal server running at pid: 144, started at: 2024-04-10 11:42:43.187267
2
+ 2024-04-10 11:42:43,189 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-10 11:42:43,491 INFO WriterThread:144 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/run-a6hgqntd.wandb
4
+ 2024-04-10 11:42:43,492 DEBUG SenderThread:144 [sender.py:send():379] send: header
5
+ 2024-04-10 11:42:43,495 DEBUG SenderThread:144 [sender.py:send():379] send: run
6
+ 2024-04-10 11:42:43,678 INFO SenderThread:144 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files
7
+ 2024-04-10 11:42:43,678 INFO SenderThread:144 [sender.py:_start_run_threads():1124] run started: a6hgqntd with start time 1712749363.188282
8
+ 2024-04-10 11:42:43,686 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-10 11:42:43,686 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-10 11:42:43,780 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-10 11:42:43,792 DEBUG HandlerThread:144 [system_info.py:__init__():26] System info init
12
+ 2024-04-10 11:42:43,792 DEBUG HandlerThread:144 [system_info.py:__init__():41] System info init done
13
+ 2024-04-10 11:42:43,792 INFO HandlerThread:144 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-10 11:42:43,792 INFO SystemMonitor:144 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-10 11:42:43,792 INFO HandlerThread:144 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-10 11:42:43,793 INFO SystemMonitor:144 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-10 11:42:43,793 INFO SystemMonitor:144 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-10 11:42:43,794 INFO SystemMonitor:144 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-10 11:42:43,795 INFO SystemMonitor:144 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-10 11:42:43,796 INFO SystemMonitor:144 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-10 11:42:43,806 DEBUG HandlerThread:144 [system_info.py:probe():150] Probing system
22
+ 2024-04-10 11:42:43,808 DEBUG HandlerThread:144 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-10 11:42:43,808 DEBUG HandlerThread:144 [system_info.py:probe():198] Probing system done
24
+ 2024-04-10 11:42:43,808 DEBUG HandlerThread:144 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T11:42:43.806558', 'startedAt': '2024-04-10T11:42:43.180890', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '57bca537cf5f', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5568.826057434082}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-10 11:42:43,809 INFO HandlerThread:144 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-10 11:42:43,809 INFO HandlerThread:144 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-10 11:42:43,809 DEBUG HandlerThread:144 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-10 11:42:44,682 INFO Thread-12 :144 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/conda-environment.yaml
29
+ 2024-04-10 11:42:58,823 ERROR HandlerThread:144 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-10 11:42:58,825 DEBUG HandlerThread:144 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-10 11:42:58,825 INFO HandlerThread:144 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-10 11:42:58,832 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-10 11:42:58,833 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-10 11:42:58,833 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-10 11:42:58,833 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-10 11:42:58,833 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-10 11:42:58,833 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-10 11:42:58,834 DEBUG SenderThread:144 [sender.py:send():379] send: files
49
+ 2024-04-10 11:42:58,834 INFO SenderThread:144 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-10 11:42:59,058 INFO wandb-upload_0:144 [upload_job.py:push():131] Uploaded file /tmp/tmp3lpmf8m8wandb/q0pbe4nu-wandb-metadata.json
51
+ 2024-04-10 11:42:59,684 INFO Thread-12 :144 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/wandb-metadata.json
52
+ 2024-04-10 11:42:59,776 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-10 11:42:59,776 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-10 11:42:59,780 DEBUG SenderThread:144 [sender.py:send():379] send: telemetry
55
+ 2024-04-10 11:42:59,786 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: stop_status
56
+ 2024-04-10 11:42:59,786 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: stop_status
57
+ 2024-04-10 11:42:59,790 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: internal_messages
58
+ 2024-04-10 11:42:59,894 DEBUG SenderThread:144 [sender.py:send():379] send: config
59
+ 2024-04-10 11:42:59,896 DEBUG SenderThread:144 [sender.py:send():379] send: metric
60
+ 2024-04-10 11:42:59,896 DEBUG SenderThread:144 [sender.py:send():379] send: telemetry
61
+ 2024-04-10 11:42:59,897 DEBUG SenderThread:144 [sender.py:send():379] send: metric
62
+ 2024-04-10 11:42:59,897 WARNING SenderThread:144 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
63
+ 2024-04-10 11:42:59,897 DEBUG SenderThread:144 [sender.py:send():379] send: telemetry
64
+ 2024-04-10 11:43:00,684 INFO Thread-12 :144 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/output.log
65
+ 2024-04-10 11:43:00,685 INFO Thread-12 :144 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/requirements.txt
66
+ 2024-04-10 11:43:02,685 INFO Thread-12 :144 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/output.log
67
+ 2024-04-10 11:43:03,929 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-10 11:43:08,930 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-04-10 11:43:13,936 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-10 11:43:14,690 INFO Thread-12 :144 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/config.yaml
71
+ 2024-04-10 11:43:14,779 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-04-10 11:43:14,779 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: stop_status
73
+ 2024-04-10 11:43:14,780 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: stop_status
74
+ 2024-04-10 11:43:19,838 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-04-10 11:43:24,839 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-10 11:43:29,777 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: stop_status
77
+ 2024-04-10 11:43:29,778 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: stop_status
78
+ 2024-04-10 11:43:29,817 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: internal_messages
79
+ 2024-04-10 11:43:30,827 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-04-10 11:43:35,828 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-10 11:43:40,829 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-10 11:43:43,736 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: pause
83
+ 2024-04-10 11:43:43,736 INFO HandlerThread:144 [handler.py:handle_request_pause():708] stopping system metrics thread
84
+ 2024-04-10 11:43:43,736 INFO HandlerThread:144 [system_monitor.py:finish():203] Stopping system monitor
85
+ 2024-04-10 11:43:43,736 DEBUG SystemMonitor:144 [system_monitor.py:_start():172] Starting system metrics aggregation loop
86
+ 2024-04-10 11:43:43,737 DEBUG SystemMonitor:144 [system_monitor.py:_start():179] Finished system metrics aggregation loop
87
+ 2024-04-10 11:43:43,737 DEBUG SystemMonitor:144 [system_monitor.py:_start():183] Publishing last batch of metrics
88
+ 2024-04-10 11:43:43,738 INFO HandlerThread:144 [interfaces.py:finish():202] Joined cpu monitor
89
+ 2024-04-10 11:43:43,738 INFO HandlerThread:144 [interfaces.py:finish():202] Joined disk monitor
90
+ 2024-04-10 11:43:43,747 INFO HandlerThread:144 [interfaces.py:finish():202] Joined gpu monitor
91
+ 2024-04-10 11:43:43,747 INFO HandlerThread:144 [interfaces.py:finish():202] Joined memory monitor
92
+ 2024-04-10 11:43:43,748 INFO HandlerThread:144 [interfaces.py:finish():202] Joined network monitor
93
+ 2024-04-10 11:43:43,748 DEBUG SenderThread:144 [sender.py:send():379] send: stats
94
+ 2024-04-10 11:43:44,777 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: stop_status
95
+ 2024-04-10 11:43:44,778 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: stop_status
96
+ 2024-04-10 11:43:44,778 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: internal_messages
97
+ 2024-04-10 11:43:46,829 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-04-10 11:43:51,830 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-10 11:43:56,832 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-04-10 11:43:57,053 DEBUG SenderThread:144 [sender.py:send():379] send: exit
101
+ 2024-04-10 11:43:57,053 INFO SenderThread:144 [sender.py:send_exit():586] handling exit code: 0
102
+ 2024-04-10 11:43:57,053 INFO SenderThread:144 [sender.py:send_exit():588] handling runtime: 59
103
+ 2024-04-10 11:43:57,055 INFO SenderThread:144 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
104
+ 2024-04-10 11:43:57,055 INFO SenderThread:144 [sender.py:send_exit():594] send defer
105
+ 2024-04-10 11:43:57,055 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
106
+ 2024-04-10 11:43:57,055 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 0
107
+ 2024-04-10 11:43:57,056 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
108
+ 2024-04-10 11:43:57,056 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 0
109
+ 2024-04-10 11:43:57,056 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 1
110
+ 2024-04-10 11:43:57,056 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
111
+ 2024-04-10 11:43:57,056 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 1
112
+ 2024-04-10 11:43:57,056 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
113
+ 2024-04-10 11:43:57,056 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 1
114
+ 2024-04-10 11:43:57,056 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 2
115
+ 2024-04-10 11:43:57,056 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
116
+ 2024-04-10 11:43:57,056 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 2
117
+ 2024-04-10 11:43:57,057 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
118
+ 2024-04-10 11:43:57,057 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 2
119
+ 2024-04-10 11:43:57,057 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 3
120
+ 2024-04-10 11:43:57,057 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
121
+ 2024-04-10 11:43:57,057 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 3
122
+ 2024-04-10 11:43:57,057 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
123
+ 2024-04-10 11:43:57,057 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 3
124
+ 2024-04-10 11:43:57,057 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 4
125
+ 2024-04-10 11:43:57,057 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
126
+ 2024-04-10 11:43:57,057 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 4
127
+ 2024-04-10 11:43:57,058 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
128
+ 2024-04-10 11:43:57,058 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 4
129
+ 2024-04-10 11:43:57,058 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 5
130
+ 2024-04-10 11:43:57,058 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
131
+ 2024-04-10 11:43:57,058 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 5
132
+ 2024-04-10 11:43:57,058 DEBUG SenderThread:144 [sender.py:send():379] send: summary
133
+ 2024-04-10 11:43:57,058 INFO SenderThread:144 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
134
+ 2024-04-10 11:43:57,059 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
135
+ 2024-04-10 11:43:57,059 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 5
136
+ 2024-04-10 11:43:57,059 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 6
137
+ 2024-04-10 11:43:57,059 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
138
+ 2024-04-10 11:43:57,059 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 6
139
+ 2024-04-10 11:43:57,059 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
140
+ 2024-04-10 11:43:57,059 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 6
141
+ 2024-04-10 11:43:57,059 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 7
142
+ 2024-04-10 11:43:57,059 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: status_report
143
+ 2024-04-10 11:43:57,059 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
144
+ 2024-04-10 11:43:57,059 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 7
145
+ 2024-04-10 11:43:57,060 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
146
+ 2024-04-10 11:43:57,060 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 7
147
+ 2024-04-10 11:43:57,707 INFO Thread-12 :144 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/wandb-summary.json
148
+ 2024-04-10 11:43:57,970 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 8
149
+ 2024-04-10 11:43:57,971 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
150
+ 2024-04-10 11:43:57,971 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 8
151
+ 2024-04-10 11:43:57,971 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
152
+ 2024-04-10 11:43:57,971 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 8
153
+ 2024-04-10 11:43:57,971 INFO SenderThread:144 [job_builder.py:build():318] Attempting to build job artifact
154
+ 2024-04-10 11:43:57,973 INFO SenderThread:144 [job_builder.py:_get_source_type():466] no source found
155
+ 2024-04-10 11:43:57,973 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 9
156
+ 2024-04-10 11:43:57,973 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
157
+ 2024-04-10 11:43:57,973 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 9
158
+ 2024-04-10 11:43:57,974 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
159
+ 2024-04-10 11:43:57,974 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 9
160
+ 2024-04-10 11:43:57,974 INFO SenderThread:144 [dir_watcher.py:finish():358] shutting down directory watcher
161
+ 2024-04-10 11:43:58,053 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: poll_exit
162
+ 2024-04-10 11:43:58,707 INFO Thread-12 :144 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/output.log
163
+ 2024-04-10 11:43:58,708 INFO SenderThread:144 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files
164
+ 2024-04-10 11:43:58,708 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/config.yaml config.yaml
165
+ 2024-04-10 11:43:58,708 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/conda-environment.yaml conda-environment.yaml
166
+ 2024-04-10 11:43:58,711 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/output.log output.log
167
+ 2024-04-10 11:43:58,712 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/wandb-metadata.json wandb-metadata.json
168
+ 2024-04-10 11:43:58,712 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/requirements.txt requirements.txt
169
+ 2024-04-10 11:43:58,712 INFO SenderThread:144 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/wandb-summary.json wandb-summary.json
170
+ 2024-04-10 11:43:58,715 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 10
171
+ 2024-04-10 11:43:58,716 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: poll_exit
172
+ 2024-04-10 11:43:58,718 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
173
+ 2024-04-10 11:43:58,719 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 10
174
+ 2024-04-10 11:43:58,722 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
175
+ 2024-04-10 11:43:58,722 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 10
176
+ 2024-04-10 11:43:58,722 INFO SenderThread:144 [file_pusher.py:finish():172] shutting down file pusher
177
+ 2024-04-10 11:43:58,867 INFO wandb-upload_0:144 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/config.yaml
178
+ 2024-04-10 11:43:58,941 INFO wandb-upload_1:144 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/output.log
179
+ 2024-04-10 11:43:58,948 INFO wandb-upload_3:144 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/wandb-summary.json
180
+ 2024-04-10 11:43:58,986 INFO wandb-upload_2:144 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114243-a6hgqntd/files/requirements.txt
181
+ 2024-04-10 11:43:59,055 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: poll_exit
182
+ 2024-04-10 11:43:59,055 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: poll_exit
183
+ 2024-04-10 11:43:59,186 INFO Thread-11 (_thread_body):144 [sender.py:transition_state():614] send defer: 11
184
+ 2024-04-10 11:43:59,187 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
185
+ 2024-04-10 11:43:59,187 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 11
186
+ 2024-04-10 11:43:59,187 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
187
+ 2024-04-10 11:43:59,187 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 11
188
+ 2024-04-10 11:43:59,187 INFO SenderThread:144 [file_pusher.py:join():178] waiting for file pusher
189
+ 2024-04-10 11:43:59,188 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 12
190
+ 2024-04-10 11:43:59,188 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
191
+ 2024-04-10 11:43:59,188 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 12
192
+ 2024-04-10 11:43:59,188 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
193
+ 2024-04-10 11:43:59,189 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 12
194
+ 2024-04-10 11:43:59,189 INFO SenderThread:144 [file_stream.py:finish():614] file stream finish called
195
+ 2024-04-10 11:43:59,266 INFO SenderThread:144 [file_stream.py:finish():618] file stream finish is done
196
+ 2024-04-10 11:43:59,266 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 13
197
+ 2024-04-10 11:43:59,266 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
198
+ 2024-04-10 11:43:59,266 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 13
199
+ 2024-04-10 11:43:59,266 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
200
+ 2024-04-10 11:43:59,266 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 13
201
+ 2024-04-10 11:43:59,266 INFO SenderThread:144 [sender.py:transition_state():614] send defer: 14
202
+ 2024-04-10 11:43:59,267 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: defer
203
+ 2024-04-10 11:43:59,267 INFO HandlerThread:144 [handler.py:handle_request_defer():172] handle defer: 14
204
+ 2024-04-10 11:43:59,267 DEBUG SenderThread:144 [sender.py:send():379] send: final
205
+ 2024-04-10 11:43:59,267 DEBUG SenderThread:144 [sender.py:send():379] send: footer
206
+ 2024-04-10 11:43:59,267 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: defer
207
+ 2024-04-10 11:43:59,268 INFO SenderThread:144 [sender.py:send_request_defer():610] handle sender defer: 14
208
+ 2024-04-10 11:43:59,268 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: poll_exit
209
+ 2024-04-10 11:43:59,269 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: poll_exit
210
+ 2024-04-10 11:43:59,269 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: poll_exit
211
+ 2024-04-10 11:43:59,270 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: poll_exit
212
+ 2024-04-10 11:43:59,270 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: server_info
213
+ 2024-04-10 11:43:59,271 DEBUG SenderThread:144 [sender.py:send_request():406] send_request: server_info
214
+ 2024-04-10 11:43:59,275 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: get_summary
215
+ 2024-04-10 11:43:59,275 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: sampled_history
216
+ 2024-04-10 11:43:59,276 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: internal_messages
217
+ 2024-04-10 11:43:59,289 INFO MainThread:144 [wandb_run.py:_footer_history_summary_info():3920] rendering history
218
+ 2024-04-10 11:43:59,289 INFO MainThread:144 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
219
+ 2024-04-10 11:43:59,289 INFO MainThread:144 [wandb_run.py:_footer_sync_info():3879] logging synced files
220
+ 2024-04-10 11:43:59,289 DEBUG HandlerThread:144 [handler.py:handle_request():146] handle_request: shutdown
221
+ 2024-04-10 11:43:59,290 INFO HandlerThread:144 [handler.py:finish():866] shutting down handler
222
+ 2024-04-10 11:44:00,271 INFO WriterThread:144 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240410_114243-a6hgqntd/run-a6hgqntd.wandb
223
+ 2024-04-10 11:44:00,289 INFO SenderThread:144 [sender.py:finish():1546] shutting down sender
224
+ 2024-04-10 11:44:00,289 INFO SenderThread:144 [file_pusher.py:finish():172] shutting down file pusher
225
+ 2024-04-10 11:44:00,289 INFO SenderThread:144 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240410_114243-a6hgqntd/logs/debug.log ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Configure stats pid to 98
3
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 11:42:43,182 INFO MainThread:98 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_114243-a6hgqntd/logs/debug.log
11
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_114243-a6hgqntd/logs/debug-internal.log
12
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7a2f8d4c2980>
13
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:init():617] starting backend
17
+ 2024-04-10 11:42:43,183 INFO MainThread:98 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-10 11:42:43,185 INFO MainThread:98 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-10 11:42:43,188 INFO MainThread:98 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-10 11:42:43,199 INFO MainThread:98 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-10 11:42:43,490 INFO MainThread:98 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-10 11:42:43,494 INFO MainThread:98 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-10 11:42:43,685 INFO MainThread:98 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-10 11:42:43,774 INFO MainThread:98 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-04-10 11:42:43,775 INFO MainThread:98 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-10 11:42:59,777 INFO MainThread:98 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-10 11:42:59,777 INFO MainThread:98 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-10 11:42:59,779 INFO MainThread:98 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-10 11:42:59,779 INFO MainThread:98 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-10 11:42:59,780 INFO MainThread:98 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-10 11:42:59,786 INFO MainThread:98 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-06, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 15000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_11-37-24_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 30, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 30, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-10 11:43:43,735 INFO MainThread:98 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-10 11:43:43,736 INFO MainThread:98 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240410_114243-a6hgqntd/run-a6hgqntd.wandb ADDED
Binary file (12.5 kB). View file
 
wandb/run-20240410_114629-hdrbjcbp/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_114629-hdrbjcbp/files/config.yaml ADDED
@@ -0,0 +1,670 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712749589.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ vocab_size:
62
+ desc: null
63
+ value: 32000
64
+ max_position_embeddings:
65
+ desc: null
66
+ value: 2048
67
+ hidden_size:
68
+ desc: null
69
+ value: 2048
70
+ intermediate_size:
71
+ desc: null
72
+ value: 5632
73
+ num_hidden_layers:
74
+ desc: null
75
+ value: 22
76
+ num_attention_heads:
77
+ desc: null
78
+ value: 32
79
+ num_key_value_heads:
80
+ desc: null
81
+ value: 4
82
+ hidden_act:
83
+ desc: null
84
+ value: silu
85
+ initializer_range:
86
+ desc: null
87
+ value: 0.02
88
+ rms_norm_eps:
89
+ desc: null
90
+ value: 1.0e-05
91
+ pretraining_tp:
92
+ desc: null
93
+ value: 1
94
+ use_cache:
95
+ desc: null
96
+ value: false
97
+ rope_theta:
98
+ desc: null
99
+ value: 10000.0
100
+ rope_scaling:
101
+ desc: null
102
+ value: null
103
+ attention_bias:
104
+ desc: null
105
+ value: false
106
+ attention_dropout:
107
+ desc: null
108
+ value: 0.0
109
+ return_dict:
110
+ desc: null
111
+ value: true
112
+ output_hidden_states:
113
+ desc: null
114
+ value: false
115
+ output_attentions:
116
+ desc: null
117
+ value: false
118
+ torchscript:
119
+ desc: null
120
+ value: false
121
+ torch_dtype:
122
+ desc: null
123
+ value: float32
124
+ use_bfloat16:
125
+ desc: null
126
+ value: false
127
+ tf_legacy_loss:
128
+ desc: null
129
+ value: false
130
+ pruned_heads:
131
+ desc: null
132
+ value: {}
133
+ tie_word_embeddings:
134
+ desc: null
135
+ value: false
136
+ chunk_size_feed_forward:
137
+ desc: null
138
+ value: 0
139
+ is_encoder_decoder:
140
+ desc: null
141
+ value: false
142
+ is_decoder:
143
+ desc: null
144
+ value: false
145
+ cross_attention_hidden_size:
146
+ desc: null
147
+ value: null
148
+ add_cross_attention:
149
+ desc: null
150
+ value: false
151
+ tie_encoder_decoder:
152
+ desc: null
153
+ value: false
154
+ max_length:
155
+ desc: null
156
+ value: 20
157
+ min_length:
158
+ desc: null
159
+ value: 0
160
+ do_sample:
161
+ desc: null
162
+ value: false
163
+ early_stopping:
164
+ desc: null
165
+ value: false
166
+ num_beams:
167
+ desc: null
168
+ value: 1
169
+ num_beam_groups:
170
+ desc: null
171
+ value: 1
172
+ diversity_penalty:
173
+ desc: null
174
+ value: 0.0
175
+ temperature:
176
+ desc: null
177
+ value: 1.0
178
+ top_k:
179
+ desc: null
180
+ value: 50
181
+ top_p:
182
+ desc: null
183
+ value: 1.0
184
+ typical_p:
185
+ desc: null
186
+ value: 1.0
187
+ repetition_penalty:
188
+ desc: null
189
+ value: 1.0
190
+ length_penalty:
191
+ desc: null
192
+ value: 1.0
193
+ no_repeat_ngram_size:
194
+ desc: null
195
+ value: 0
196
+ encoder_no_repeat_ngram_size:
197
+ desc: null
198
+ value: 0
199
+ bad_words_ids:
200
+ desc: null
201
+ value: null
202
+ num_return_sequences:
203
+ desc: null
204
+ value: 1
205
+ output_scores:
206
+ desc: null
207
+ value: false
208
+ return_dict_in_generate:
209
+ desc: null
210
+ value: false
211
+ forced_bos_token_id:
212
+ desc: null
213
+ value: null
214
+ forced_eos_token_id:
215
+ desc: null
216
+ value: null
217
+ remove_invalid_values:
218
+ desc: null
219
+ value: false
220
+ exponential_decay_length_penalty:
221
+ desc: null
222
+ value: null
223
+ suppress_tokens:
224
+ desc: null
225
+ value: null
226
+ begin_suppress_tokens:
227
+ desc: null
228
+ value: null
229
+ architectures:
230
+ desc: null
231
+ value:
232
+ - LlamaForCausalLM
233
+ finetuning_task:
234
+ desc: null
235
+ value: null
236
+ id2label:
237
+ desc: null
238
+ value:
239
+ '0': LABEL_0
240
+ '1': LABEL_1
241
+ label2id:
242
+ desc: null
243
+ value:
244
+ LABEL_0: 0
245
+ LABEL_1: 1
246
+ tokenizer_class:
247
+ desc: null
248
+ value: null
249
+ prefix:
250
+ desc: null
251
+ value: null
252
+ bos_token_id:
253
+ desc: null
254
+ value: 1
255
+ pad_token_id:
256
+ desc: null
257
+ value: null
258
+ eos_token_id:
259
+ desc: null
260
+ value: 2
261
+ sep_token_id:
262
+ desc: null
263
+ value: null
264
+ decoder_start_token_id:
265
+ desc: null
266
+ value: null
267
+ task_specific_params:
268
+ desc: null
269
+ value: null
270
+ problem_type:
271
+ desc: null
272
+ value: null
273
+ _name_or_path:
274
+ desc: null
275
+ value: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
276
+ transformers_version:
277
+ desc: null
278
+ value: 4.39.3
279
+ model_type:
280
+ desc: null
281
+ value: llama
282
+ quantization_config:
283
+ desc: null
284
+ value:
285
+ quant_method: QuantizationMethod.BITS_AND_BYTES
286
+ _load_in_8bit: false
287
+ _load_in_4bit: true
288
+ llm_int8_threshold: 6.0
289
+ llm_int8_skip_modules: null
290
+ llm_int8_enable_fp32_cpu_offload: false
291
+ llm_int8_has_fp16_weight: false
292
+ bnb_4bit_quant_type: nf4
293
+ bnb_4bit_use_double_quant: false
294
+ bnb_4bit_compute_dtype: float16
295
+ bnb_4bit_quant_storage: uint8
296
+ load_in_4bit: true
297
+ load_in_8bit: false
298
+ output_dir:
299
+ desc: null
300
+ value: /kaggle/working/
301
+ overwrite_output_dir:
302
+ desc: null
303
+ value: false
304
+ do_train:
305
+ desc: null
306
+ value: false
307
+ do_eval:
308
+ desc: null
309
+ value: false
310
+ do_predict:
311
+ desc: null
312
+ value: false
313
+ evaluation_strategy:
314
+ desc: null
315
+ value: 'no'
316
+ prediction_loss_only:
317
+ desc: null
318
+ value: false
319
+ per_device_train_batch_size:
320
+ desc: null
321
+ value: 8
322
+ per_device_eval_batch_size:
323
+ desc: null
324
+ value: 8
325
+ per_gpu_train_batch_size:
326
+ desc: null
327
+ value: null
328
+ per_gpu_eval_batch_size:
329
+ desc: null
330
+ value: null
331
+ gradient_accumulation_steps:
332
+ desc: null
333
+ value: 4
334
+ eval_accumulation_steps:
335
+ desc: null
336
+ value: null
337
+ eval_delay:
338
+ desc: null
339
+ value: 0
340
+ learning_rate:
341
+ desc: null
342
+ value: 3.0e-05
343
+ weight_decay:
344
+ desc: null
345
+ value: 0.0001
346
+ adam_beta1:
347
+ desc: null
348
+ value: 0.9
349
+ adam_beta2:
350
+ desc: null
351
+ value: 0.999
352
+ adam_epsilon:
353
+ desc: null
354
+ value: 1.0e-08
355
+ max_grad_norm:
356
+ desc: null
357
+ value: 0.01
358
+ num_train_epochs:
359
+ desc: null
360
+ value: 5
361
+ max_steps:
362
+ desc: null
363
+ value: 7000
364
+ lr_scheduler_type:
365
+ desc: null
366
+ value: cosine
367
+ lr_scheduler_kwargs:
368
+ desc: null
369
+ value: {}
370
+ warmup_ratio:
371
+ desc: null
372
+ value: 0.05
373
+ warmup_steps:
374
+ desc: null
375
+ value: 0
376
+ log_level:
377
+ desc: null
378
+ value: passive
379
+ log_level_replica:
380
+ desc: null
381
+ value: warning
382
+ log_on_each_node:
383
+ desc: null
384
+ value: true
385
+ logging_dir:
386
+ desc: null
387
+ value: /kaggle/working/runs/Apr10_11-44-26_57bca537cf5f
388
+ logging_strategy:
389
+ desc: null
390
+ value: steps
391
+ logging_first_step:
392
+ desc: null
393
+ value: false
394
+ logging_steps:
395
+ desc: null
396
+ value: 30
397
+ logging_nan_inf_filter:
398
+ desc: null
399
+ value: true
400
+ save_strategy:
401
+ desc: null
402
+ value: steps
403
+ save_steps:
404
+ desc: null
405
+ value: 30
406
+ save_total_limit:
407
+ desc: null
408
+ value: 1
409
+ save_safetensors:
410
+ desc: null
411
+ value: true
412
+ save_on_each_node:
413
+ desc: null
414
+ value: false
415
+ save_only_model:
416
+ desc: null
417
+ value: false
418
+ no_cuda:
419
+ desc: null
420
+ value: false
421
+ use_cpu:
422
+ desc: null
423
+ value: false
424
+ use_mps_device:
425
+ desc: null
426
+ value: false
427
+ seed:
428
+ desc: null
429
+ value: 42
430
+ data_seed:
431
+ desc: null
432
+ value: null
433
+ jit_mode_eval:
434
+ desc: null
435
+ value: false
436
+ use_ipex:
437
+ desc: null
438
+ value: false
439
+ bf16:
440
+ desc: null
441
+ value: false
442
+ fp16:
443
+ desc: null
444
+ value: false
445
+ fp16_opt_level:
446
+ desc: null
447
+ value: O1
448
+ half_precision_backend:
449
+ desc: null
450
+ value: auto
451
+ bf16_full_eval:
452
+ desc: null
453
+ value: false
454
+ fp16_full_eval:
455
+ desc: null
456
+ value: false
457
+ tf32:
458
+ desc: null
459
+ value: null
460
+ local_rank:
461
+ desc: null
462
+ value: 0
463
+ ddp_backend:
464
+ desc: null
465
+ value: null
466
+ tpu_num_cores:
467
+ desc: null
468
+ value: null
469
+ tpu_metrics_debug:
470
+ desc: null
471
+ value: false
472
+ debug:
473
+ desc: null
474
+ value: []
475
+ dataloader_drop_last:
476
+ desc: null
477
+ value: false
478
+ eval_steps:
479
+ desc: null
480
+ value: null
481
+ dataloader_num_workers:
482
+ desc: null
483
+ value: 8
484
+ dataloader_prefetch_factor:
485
+ desc: null
486
+ value: null
487
+ past_index:
488
+ desc: null
489
+ value: -1
490
+ run_name:
491
+ desc: null
492
+ value: /kaggle/working/
493
+ disable_tqdm:
494
+ desc: null
495
+ value: false
496
+ remove_unused_columns:
497
+ desc: null
498
+ value: true
499
+ label_names:
500
+ desc: null
501
+ value: null
502
+ load_best_model_at_end:
503
+ desc: null
504
+ value: false
505
+ metric_for_best_model:
506
+ desc: null
507
+ value: null
508
+ greater_is_better:
509
+ desc: null
510
+ value: null
511
+ ignore_data_skip:
512
+ desc: null
513
+ value: false
514
+ fsdp:
515
+ desc: null
516
+ value: []
517
+ fsdp_min_num_params:
518
+ desc: null
519
+ value: 0
520
+ fsdp_config:
521
+ desc: null
522
+ value:
523
+ min_num_params: 0
524
+ xla: false
525
+ xla_fsdp_v2: false
526
+ xla_fsdp_grad_ckpt: false
527
+ fsdp_transformer_layer_cls_to_wrap:
528
+ desc: null
529
+ value: null
530
+ accelerator_config:
531
+ desc: null
532
+ value:
533
+ split_batches: false
534
+ dispatch_batches: null
535
+ even_batches: true
536
+ use_seedable_sampler: true
537
+ deepspeed:
538
+ desc: null
539
+ value: null
540
+ label_smoothing_factor:
541
+ desc: null
542
+ value: 0.0
543
+ optim:
544
+ desc: null
545
+ value: paged_adamw_8bit
546
+ optim_args:
547
+ desc: null
548
+ value: null
549
+ adafactor:
550
+ desc: null
551
+ value: false
552
+ group_by_length:
553
+ desc: null
554
+ value: false
555
+ length_column_name:
556
+ desc: null
557
+ value: length
558
+ report_to:
559
+ desc: null
560
+ value:
561
+ - tensorboard
562
+ - wandb
563
+ ddp_find_unused_parameters:
564
+ desc: null
565
+ value: null
566
+ ddp_bucket_cap_mb:
567
+ desc: null
568
+ value: null
569
+ ddp_broadcast_buffers:
570
+ desc: null
571
+ value: null
572
+ dataloader_pin_memory:
573
+ desc: null
574
+ value: true
575
+ dataloader_persistent_workers:
576
+ desc: null
577
+ value: false
578
+ skip_memory_metrics:
579
+ desc: null
580
+ value: true
581
+ use_legacy_prediction_loop:
582
+ desc: null
583
+ value: false
584
+ push_to_hub:
585
+ desc: null
586
+ value: false
587
+ resume_from_checkpoint:
588
+ desc: null
589
+ value: null
590
+ hub_model_id:
591
+ desc: null
592
+ value: null
593
+ hub_strategy:
594
+ desc: null
595
+ value: every_save
596
+ hub_token:
597
+ desc: null
598
+ value: <HUB_TOKEN>
599
+ hub_private_repo:
600
+ desc: null
601
+ value: false
602
+ hub_always_push:
603
+ desc: null
604
+ value: false
605
+ gradient_checkpointing:
606
+ desc: null
607
+ value: true
608
+ gradient_checkpointing_kwargs:
609
+ desc: null
610
+ value: null
611
+ include_inputs_for_metrics:
612
+ desc: null
613
+ value: false
614
+ fp16_backend:
615
+ desc: null
616
+ value: auto
617
+ push_to_hub_model_id:
618
+ desc: null
619
+ value: null
620
+ push_to_hub_organization:
621
+ desc: null
622
+ value: null
623
+ push_to_hub_token:
624
+ desc: null
625
+ value: <PUSH_TO_HUB_TOKEN>
626
+ mp_parameters:
627
+ desc: null
628
+ value: ''
629
+ auto_find_batch_size:
630
+ desc: null
631
+ value: true
632
+ full_determinism:
633
+ desc: null
634
+ value: false
635
+ torchdynamo:
636
+ desc: null
637
+ value: null
638
+ ray_scope:
639
+ desc: null
640
+ value: last
641
+ ddp_timeout:
642
+ desc: null
643
+ value: 1800
644
+ torch_compile:
645
+ desc: null
646
+ value: false
647
+ torch_compile_backend:
648
+ desc: null
649
+ value: null
650
+ torch_compile_mode:
651
+ desc: null
652
+ value: null
653
+ dispatch_batches:
654
+ desc: null
655
+ value: null
656
+ split_batches:
657
+ desc: null
658
+ value: null
659
+ include_tokens_per_second:
660
+ desc: null
661
+ value: false
662
+ include_num_input_tokens_seen:
663
+ desc: null
664
+ value: false
665
+ neftune_noise_alpha:
666
+ desc: null
667
+ value: null
668
+ optim_target_modules:
669
+ desc: null
670
+ value: null
wandb/run-20240410_114629-hdrbjcbp/files/output.log ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
wandb/run-20240410_114629-hdrbjcbp/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240410_114629-hdrbjcbp/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T11:46:30.239646",
5
+ "startedAt": "2024-04-10T11:46:29.685763",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "57bca537cf5f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5569.163795471191
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240410_114629-hdrbjcbp/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 58}}
wandb/run-20240410_114629-hdrbjcbp/logs/debug-internal.log ADDED
@@ -0,0 +1,223 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 11:46:29,692 INFO StreamThr :295 [internal.py:wandb_internal():86] W&B internal server running at pid: 295, started at: 2024-04-10 11:46:29.691905
2
+ 2024-04-10 11:46:29,694 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-10 11:46:29,990 INFO WriterThread:295 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/run-hdrbjcbp.wandb
4
+ 2024-04-10 11:46:29,991 DEBUG SenderThread:295 [sender.py:send():379] send: header
5
+ 2024-04-10 11:46:29,993 DEBUG SenderThread:295 [sender.py:send():379] send: run
6
+ 2024-04-10 11:46:30,118 INFO SenderThread:295 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files
7
+ 2024-04-10 11:46:30,118 INFO SenderThread:295 [sender.py:_start_run_threads():1124] run started: hdrbjcbp with start time 1712749589.691686
8
+ 2024-04-10 11:46:30,126 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-10 11:46:30,126 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-10 11:46:30,214 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-10 11:46:30,225 DEBUG HandlerThread:295 [system_info.py:__init__():26] System info init
12
+ 2024-04-10 11:46:30,225 DEBUG HandlerThread:295 [system_info.py:__init__():41] System info init done
13
+ 2024-04-10 11:46:30,225 INFO HandlerThread:295 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-10 11:46:30,225 INFO SystemMonitor:295 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-10 11:46:30,225 INFO HandlerThread:295 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-10 11:46:30,226 INFO SystemMonitor:295 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-10 11:46:30,227 INFO SystemMonitor:295 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-10 11:46:30,228 INFO SystemMonitor:295 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-10 11:46:30,229 INFO SystemMonitor:295 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-10 11:46:30,229 INFO SystemMonitor:295 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-10 11:46:30,239 DEBUG HandlerThread:295 [system_info.py:probe():150] Probing system
22
+ 2024-04-10 11:46:30,241 DEBUG HandlerThread:295 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-10 11:46:30,241 DEBUG HandlerThread:295 [system_info.py:probe():198] Probing system done
24
+ 2024-04-10 11:46:30,241 DEBUG HandlerThread:295 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T11:46:30.239646', 'startedAt': '2024-04-10T11:46:29.685763', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '57bca537cf5f', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5569.163795471191}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-10 11:46:30,241 INFO HandlerThread:295 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-10 11:46:30,241 INFO HandlerThread:295 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-10 11:46:30,242 DEBUG HandlerThread:295 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-10 11:46:31,121 INFO Thread-12 :295 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/conda-environment.yaml
29
+ 2024-04-10 11:46:45,257 ERROR HandlerThread:295 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-10 11:46:45,258 DEBUG HandlerThread:295 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-10 11:46:45,258 INFO HandlerThread:295 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-10 11:46:45,264 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-10 11:46:45,264 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-10 11:46:45,264 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-10 11:46:45,265 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-10 11:46:45,265 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-10 11:46:45,265 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-10 11:46:45,266 DEBUG SenderThread:295 [sender.py:send():379] send: files
49
+ 2024-04-10 11:46:45,266 INFO SenderThread:295 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-10 11:46:45,453 INFO wandb-upload_0:295 [upload_job.py:push():131] Uploaded file /tmp/tmpr9bzdyscwandb/y4g1ya4v-wandb-metadata.json
51
+ 2024-04-10 11:46:46,124 INFO Thread-12 :295 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/wandb-metadata.json
52
+ 2024-04-10 11:46:46,184 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-10 11:46:46,184 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-10 11:46:46,187 DEBUG SenderThread:295 [sender.py:send():379] send: telemetry
55
+ 2024-04-10 11:46:46,188 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: stop_status
56
+ 2024-04-10 11:46:46,189 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: stop_status
57
+ 2024-04-10 11:46:46,200 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: internal_messages
58
+ 2024-04-10 11:46:46,280 DEBUG SenderThread:295 [sender.py:send():379] send: config
59
+ 2024-04-10 11:46:46,281 DEBUG SenderThread:295 [sender.py:send():379] send: metric
60
+ 2024-04-10 11:46:46,282 DEBUG SenderThread:295 [sender.py:send():379] send: telemetry
61
+ 2024-04-10 11:46:46,282 DEBUG SenderThread:295 [sender.py:send():379] send: metric
62
+ 2024-04-10 11:46:46,282 WARNING SenderThread:295 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
63
+ 2024-04-10 11:46:46,282 DEBUG SenderThread:295 [sender.py:send():379] send: telemetry
64
+ 2024-04-10 11:46:47,124 INFO Thread-12 :295 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/output.log
65
+ 2024-04-10 11:46:47,124 INFO Thread-12 :295 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/requirements.txt
66
+ 2024-04-10 11:46:49,125 INFO Thread-12 :295 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/output.log
67
+ 2024-04-10 11:46:50,323 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-10 11:46:55,325 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-04-10 11:47:00,330 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-10 11:47:01,129 INFO Thread-12 :295 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/config.yaml
71
+ 2024-04-10 11:47:01,186 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: internal_messages
72
+ 2024-04-10 11:47:01,188 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: stop_status
73
+ 2024-04-10 11:47:01,188 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: stop_status
74
+ 2024-04-10 11:47:06,304 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-04-10 11:47:11,304 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-10 11:47:16,185 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: stop_status
77
+ 2024-04-10 11:47:16,185 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: stop_status
78
+ 2024-04-10 11:47:16,189 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: internal_messages
79
+ 2024-04-10 11:47:17,266 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-04-10 11:47:22,267 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-10 11:47:27,268 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-10 11:47:28,598 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: pause
83
+ 2024-04-10 11:47:28,598 INFO HandlerThread:295 [handler.py:handle_request_pause():708] stopping system metrics thread
84
+ 2024-04-10 11:47:28,599 INFO HandlerThread:295 [system_monitor.py:finish():203] Stopping system monitor
85
+ 2024-04-10 11:47:28,599 DEBUG SystemMonitor:295 [system_monitor.py:_start():172] Starting system metrics aggregation loop
86
+ 2024-04-10 11:47:28,599 DEBUG SystemMonitor:295 [system_monitor.py:_start():179] Finished system metrics aggregation loop
87
+ 2024-04-10 11:47:28,599 INFO HandlerThread:295 [interfaces.py:finish():202] Joined cpu monitor
88
+ 2024-04-10 11:47:28,599 DEBUG SystemMonitor:295 [system_monitor.py:_start():183] Publishing last batch of metrics
89
+ 2024-04-10 11:47:28,600 INFO HandlerThread:295 [interfaces.py:finish():202] Joined disk monitor
90
+ 2024-04-10 11:47:28,610 INFO HandlerThread:295 [interfaces.py:finish():202] Joined gpu monitor
91
+ 2024-04-10 11:47:28,610 INFO HandlerThread:295 [interfaces.py:finish():202] Joined memory monitor
92
+ 2024-04-10 11:47:28,610 INFO HandlerThread:295 [interfaces.py:finish():202] Joined network monitor
93
+ 2024-04-10 11:47:28,611 DEBUG SenderThread:295 [sender.py:send():379] send: stats
94
+ 2024-04-10 11:47:31,185 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: stop_status
95
+ 2024-04-10 11:47:31,185 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: stop_status
96
+ 2024-04-10 11:47:31,188 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: internal_messages
97
+ 2024-04-10 11:47:32,303 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-04-10 11:47:33,655 DEBUG SenderThread:295 [sender.py:send():379] send: exit
99
+ 2024-04-10 11:47:33,655 INFO SenderThread:295 [sender.py:send_exit():586] handling exit code: 0
100
+ 2024-04-10 11:47:33,655 INFO SenderThread:295 [sender.py:send_exit():588] handling runtime: 58
101
+ 2024-04-10 11:47:33,655 INFO SenderThread:295 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
102
+ 2024-04-10 11:47:33,655 INFO SenderThread:295 [sender.py:send_exit():594] send defer
103
+ 2024-04-10 11:47:33,656 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
104
+ 2024-04-10 11:47:33,656 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 0
105
+ 2024-04-10 11:47:33,656 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
106
+ 2024-04-10 11:47:33,656 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 0
107
+ 2024-04-10 11:47:33,656 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 1
108
+ 2024-04-10 11:47:33,656 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
109
+ 2024-04-10 11:47:33,656 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 1
110
+ 2024-04-10 11:47:33,656 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
111
+ 2024-04-10 11:47:33,657 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 1
112
+ 2024-04-10 11:47:33,657 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 2
113
+ 2024-04-10 11:47:33,657 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
114
+ 2024-04-10 11:47:33,657 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 2
115
+ 2024-04-10 11:47:33,657 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
116
+ 2024-04-10 11:47:33,657 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 2
117
+ 2024-04-10 11:47:33,657 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 3
118
+ 2024-04-10 11:47:33,657 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
119
+ 2024-04-10 11:47:33,657 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 3
120
+ 2024-04-10 11:47:33,657 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
121
+ 2024-04-10 11:47:33,658 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 3
122
+ 2024-04-10 11:47:33,658 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 4
123
+ 2024-04-10 11:47:33,658 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
124
+ 2024-04-10 11:47:33,658 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 4
125
+ 2024-04-10 11:47:33,658 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
126
+ 2024-04-10 11:47:33,658 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 4
127
+ 2024-04-10 11:47:33,658 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 5
128
+ 2024-04-10 11:47:33,658 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
129
+ 2024-04-10 11:47:33,658 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 5
130
+ 2024-04-10 11:47:33,658 DEBUG SenderThread:295 [sender.py:send():379] send: summary
131
+ 2024-04-10 11:47:33,659 INFO SenderThread:295 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
132
+ 2024-04-10 11:47:33,659 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
133
+ 2024-04-10 11:47:33,659 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 5
134
+ 2024-04-10 11:47:33,659 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 6
135
+ 2024-04-10 11:47:33,659 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
136
+ 2024-04-10 11:47:33,659 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 6
137
+ 2024-04-10 11:47:33,659 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
138
+ 2024-04-10 11:47:33,660 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 6
139
+ 2024-04-10 11:47:33,660 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 7
140
+ 2024-04-10 11:47:33,660 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: status_report
141
+ 2024-04-10 11:47:33,660 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
142
+ 2024-04-10 11:47:33,660 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 7
143
+ 2024-04-10 11:47:33,660 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
144
+ 2024-04-10 11:47:33,660 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 7
145
+ 2024-04-10 11:47:34,141 INFO Thread-12 :295 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/wandb-summary.json
146
+ 2024-04-10 11:47:34,346 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 8
147
+ 2024-04-10 11:47:34,346 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
148
+ 2024-04-10 11:47:34,346 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 8
149
+ 2024-04-10 11:47:34,347 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
150
+ 2024-04-10 11:47:34,347 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 8
151
+ 2024-04-10 11:47:34,347 INFO SenderThread:295 [job_builder.py:build():318] Attempting to build job artifact
152
+ 2024-04-10 11:47:34,349 INFO SenderThread:295 [job_builder.py:_get_source_type():466] no source found
153
+ 2024-04-10 11:47:34,349 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 9
154
+ 2024-04-10 11:47:34,349 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
155
+ 2024-04-10 11:47:34,349 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 9
156
+ 2024-04-10 11:47:34,349 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
157
+ 2024-04-10 11:47:34,349 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 9
158
+ 2024-04-10 11:47:34,349 INFO SenderThread:295 [dir_watcher.py:finish():358] shutting down directory watcher
159
+ 2024-04-10 11:47:34,655 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: poll_exit
160
+ 2024-04-10 11:47:35,142 INFO Thread-12 :295 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/output.log
161
+ 2024-04-10 11:47:35,143 INFO SenderThread:295 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files
162
+ 2024-04-10 11:47:35,143 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/config.yaml config.yaml
163
+ 2024-04-10 11:47:35,143 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/conda-environment.yaml conda-environment.yaml
164
+ 2024-04-10 11:47:35,146 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/output.log output.log
165
+ 2024-04-10 11:47:35,147 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/wandb-metadata.json wandb-metadata.json
166
+ 2024-04-10 11:47:35,147 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/requirements.txt requirements.txt
167
+ 2024-04-10 11:47:35,150 INFO SenderThread:295 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/wandb-summary.json wandb-summary.json
168
+ 2024-04-10 11:47:35,153 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 10
169
+ 2024-04-10 11:47:35,154 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: poll_exit
170
+ 2024-04-10 11:47:35,154 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
171
+ 2024-04-10 11:47:35,154 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 10
172
+ 2024-04-10 11:47:35,158 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
173
+ 2024-04-10 11:47:35,158 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 10
174
+ 2024-04-10 11:47:35,158 INFO SenderThread:295 [file_pusher.py:finish():172] shutting down file pusher
175
+ 2024-04-10 11:47:35,307 INFO wandb-upload_0:295 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/config.yaml
176
+ 2024-04-10 11:47:35,353 INFO wandb-upload_3:295 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/wandb-summary.json
177
+ 2024-04-10 11:47:35,381 INFO wandb-upload_1:295 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/output.log
178
+ 2024-04-10 11:47:35,381 INFO wandb-upload_2:295 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/files/requirements.txt
179
+ 2024-04-10 11:47:35,582 INFO Thread-11 (_thread_body):295 [sender.py:transition_state():614] send defer: 11
180
+ 2024-04-10 11:47:35,582 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
181
+ 2024-04-10 11:47:35,582 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 11
182
+ 2024-04-10 11:47:35,583 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
183
+ 2024-04-10 11:47:35,583 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 11
184
+ 2024-04-10 11:47:35,583 INFO SenderThread:295 [file_pusher.py:join():178] waiting for file pusher
185
+ 2024-04-10 11:47:35,583 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 12
186
+ 2024-04-10 11:47:35,583 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
187
+ 2024-04-10 11:47:35,583 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 12
188
+ 2024-04-10 11:47:35,584 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
189
+ 2024-04-10 11:47:35,584 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 12
190
+ 2024-04-10 11:47:35,584 INFO SenderThread:295 [file_stream.py:finish():614] file stream finish called
191
+ 2024-04-10 11:47:35,656 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: poll_exit
192
+ 2024-04-10 11:47:35,666 INFO SenderThread:295 [file_stream.py:finish():618] file stream finish is done
193
+ 2024-04-10 11:47:35,666 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 13
194
+ 2024-04-10 11:47:35,666 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: poll_exit
195
+ 2024-04-10 11:47:35,666 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
196
+ 2024-04-10 11:47:35,666 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 13
197
+ 2024-04-10 11:47:35,666 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
198
+ 2024-04-10 11:47:35,666 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 13
199
+ 2024-04-10 11:47:35,667 INFO SenderThread:295 [sender.py:transition_state():614] send defer: 14
200
+ 2024-04-10 11:47:35,667 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: defer
201
+ 2024-04-10 11:47:35,667 INFO HandlerThread:295 [handler.py:handle_request_defer():172] handle defer: 14
202
+ 2024-04-10 11:47:35,667 DEBUG SenderThread:295 [sender.py:send():379] send: final
203
+ 2024-04-10 11:47:35,667 DEBUG SenderThread:295 [sender.py:send():379] send: footer
204
+ 2024-04-10 11:47:35,667 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: defer
205
+ 2024-04-10 11:47:35,667 INFO SenderThread:295 [sender.py:send_request_defer():610] handle sender defer: 14
206
+ 2024-04-10 11:47:35,668 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: poll_exit
207
+ 2024-04-10 11:47:35,668 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: poll_exit
208
+ 2024-04-10 11:47:35,669 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: poll_exit
209
+ 2024-04-10 11:47:35,669 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: poll_exit
210
+ 2024-04-10 11:47:35,669 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: server_info
211
+ 2024-04-10 11:47:35,670 DEBUG SenderThread:295 [sender.py:send_request():406] send_request: server_info
212
+ 2024-04-10 11:47:35,672 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: get_summary
213
+ 2024-04-10 11:47:35,673 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: sampled_history
214
+ 2024-04-10 11:47:35,673 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: internal_messages
215
+ 2024-04-10 11:47:35,690 INFO MainThread:295 [wandb_run.py:_footer_history_summary_info():3920] rendering history
216
+ 2024-04-10 11:47:35,690 INFO MainThread:295 [wandb_run.py:_footer_history_summary_info():3952] rendering summary
217
+ 2024-04-10 11:47:35,690 INFO MainThread:295 [wandb_run.py:_footer_sync_info():3879] logging synced files
218
+ 2024-04-10 11:47:35,690 DEBUG HandlerThread:295 [handler.py:handle_request():146] handle_request: shutdown
219
+ 2024-04-10 11:47:35,690 INFO HandlerThread:295 [handler.py:finish():866] shutting down handler
220
+ 2024-04-10 11:47:36,670 INFO WriterThread:295 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/run-hdrbjcbp.wandb
221
+ 2024-04-10 11:47:36,690 INFO SenderThread:295 [sender.py:finish():1546] shutting down sender
222
+ 2024-04-10 11:47:36,690 INFO SenderThread:295 [file_pusher.py:finish():172] shutting down file pusher
223
+ 2024-04-10 11:47:36,690 INFO SenderThread:295 [file_pusher.py:join():178] waiting for file pusher
wandb/run-20240410_114629-hdrbjcbp/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 11:46:29,687 INFO MainThread:248 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Configure stats pid to 248
3
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/logs/debug.log
10
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_114629-hdrbjcbp/logs/debug-internal.log
11
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7eba35f6dc60>
12
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:init():617] starting backend
16
+ 2024-04-10 11:46:29,688 INFO MainThread:248 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-10 11:46:29,690 INFO MainThread:248 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-10 11:46:29,691 INFO MainThread:248 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-10 11:46:29,704 INFO MainThread:248 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-10 11:46:29,990 INFO MainThread:248 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-10 11:46:29,992 INFO MainThread:248 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-10 11:46:30,125 INFO MainThread:248 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-10 11:46:30,208 INFO MainThread:248 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-10 11:46:30,208 INFO MainThread:248 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-10 11:46:46,184 INFO MainThread:248 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-10 11:46:46,185 INFO MainThread:248 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-10 11:46:46,186 INFO MainThread:248 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-10 11:46:46,186 INFO MainThread:248 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-10 11:46:46,188 INFO MainThread:248 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-10 11:46:46,194 INFO MainThread:248 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 7000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_11-44-26_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 30, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 30, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-10 11:47:28,598 INFO MainThread:248 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-10 11:47:28,598 INFO MainThread:248 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240410_114629-hdrbjcbp/run-hdrbjcbp.wandb ADDED
Binary file (12.4 kB). View file
 
wandb/run-20240410_115017-r9p6sok2/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_115017-r9p6sok2/files/config.yaml ADDED
@@ -0,0 +1,687 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712749817.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ - 62
48
+ 4: 3.10.13
49
+ 5: 0.16.5
50
+ 6: 4.39.3
51
+ 8:
52
+ - 1
53
+ - 2
54
+ - 5
55
+ 9:
56
+ 1: transformers_trainer
57
+ 13: linux-x86_64
58
+ m:
59
+ - 1: train/global_step
60
+ 6:
61
+ - 3
62
+ - 1: train/loss
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/grad_norm
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/learning_rate
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: train/epoch
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ vocab_size:
79
+ desc: null
80
+ value: 32000
81
+ max_position_embeddings:
82
+ desc: null
83
+ value: 2048
84
+ hidden_size:
85
+ desc: null
86
+ value: 2048
87
+ intermediate_size:
88
+ desc: null
89
+ value: 5632
90
+ num_hidden_layers:
91
+ desc: null
92
+ value: 22
93
+ num_attention_heads:
94
+ desc: null
95
+ value: 32
96
+ num_key_value_heads:
97
+ desc: null
98
+ value: 4
99
+ hidden_act:
100
+ desc: null
101
+ value: silu
102
+ initializer_range:
103
+ desc: null
104
+ value: 0.02
105
+ rms_norm_eps:
106
+ desc: null
107
+ value: 1.0e-05
108
+ pretraining_tp:
109
+ desc: null
110
+ value: 1
111
+ use_cache:
112
+ desc: null
113
+ value: false
114
+ rope_theta:
115
+ desc: null
116
+ value: 10000.0
117
+ rope_scaling:
118
+ desc: null
119
+ value: null
120
+ attention_bias:
121
+ desc: null
122
+ value: false
123
+ attention_dropout:
124
+ desc: null
125
+ value: 0.0
126
+ return_dict:
127
+ desc: null
128
+ value: true
129
+ output_hidden_states:
130
+ desc: null
131
+ value: false
132
+ output_attentions:
133
+ desc: null
134
+ value: false
135
+ torchscript:
136
+ desc: null
137
+ value: false
138
+ torch_dtype:
139
+ desc: null
140
+ value: float32
141
+ use_bfloat16:
142
+ desc: null
143
+ value: false
144
+ tf_legacy_loss:
145
+ desc: null
146
+ value: false
147
+ pruned_heads:
148
+ desc: null
149
+ value: {}
150
+ tie_word_embeddings:
151
+ desc: null
152
+ value: false
153
+ chunk_size_feed_forward:
154
+ desc: null
155
+ value: 0
156
+ is_encoder_decoder:
157
+ desc: null
158
+ value: false
159
+ is_decoder:
160
+ desc: null
161
+ value: false
162
+ cross_attention_hidden_size:
163
+ desc: null
164
+ value: null
165
+ add_cross_attention:
166
+ desc: null
167
+ value: false
168
+ tie_encoder_decoder:
169
+ desc: null
170
+ value: false
171
+ max_length:
172
+ desc: null
173
+ value: 20
174
+ min_length:
175
+ desc: null
176
+ value: 0
177
+ do_sample:
178
+ desc: null
179
+ value: false
180
+ early_stopping:
181
+ desc: null
182
+ value: false
183
+ num_beams:
184
+ desc: null
185
+ value: 1
186
+ num_beam_groups:
187
+ desc: null
188
+ value: 1
189
+ diversity_penalty:
190
+ desc: null
191
+ value: 0.0
192
+ temperature:
193
+ desc: null
194
+ value: 1.0
195
+ top_k:
196
+ desc: null
197
+ value: 50
198
+ top_p:
199
+ desc: null
200
+ value: 1.0
201
+ typical_p:
202
+ desc: null
203
+ value: 1.0
204
+ repetition_penalty:
205
+ desc: null
206
+ value: 1.0
207
+ length_penalty:
208
+ desc: null
209
+ value: 1.0
210
+ no_repeat_ngram_size:
211
+ desc: null
212
+ value: 0
213
+ encoder_no_repeat_ngram_size:
214
+ desc: null
215
+ value: 0
216
+ bad_words_ids:
217
+ desc: null
218
+ value: null
219
+ num_return_sequences:
220
+ desc: null
221
+ value: 1
222
+ output_scores:
223
+ desc: null
224
+ value: false
225
+ return_dict_in_generate:
226
+ desc: null
227
+ value: false
228
+ forced_bos_token_id:
229
+ desc: null
230
+ value: null
231
+ forced_eos_token_id:
232
+ desc: null
233
+ value: null
234
+ remove_invalid_values:
235
+ desc: null
236
+ value: false
237
+ exponential_decay_length_penalty:
238
+ desc: null
239
+ value: null
240
+ suppress_tokens:
241
+ desc: null
242
+ value: null
243
+ begin_suppress_tokens:
244
+ desc: null
245
+ value: null
246
+ architectures:
247
+ desc: null
248
+ value:
249
+ - LlamaForCausalLM
250
+ finetuning_task:
251
+ desc: null
252
+ value: null
253
+ id2label:
254
+ desc: null
255
+ value:
256
+ '0': LABEL_0
257
+ '1': LABEL_1
258
+ label2id:
259
+ desc: null
260
+ value:
261
+ LABEL_0: 0
262
+ LABEL_1: 1
263
+ tokenizer_class:
264
+ desc: null
265
+ value: null
266
+ prefix:
267
+ desc: null
268
+ value: null
269
+ bos_token_id:
270
+ desc: null
271
+ value: 1
272
+ pad_token_id:
273
+ desc: null
274
+ value: null
275
+ eos_token_id:
276
+ desc: null
277
+ value: 2
278
+ sep_token_id:
279
+ desc: null
280
+ value: null
281
+ decoder_start_token_id:
282
+ desc: null
283
+ value: null
284
+ task_specific_params:
285
+ desc: null
286
+ value: null
287
+ problem_type:
288
+ desc: null
289
+ value: null
290
+ _name_or_path:
291
+ desc: null
292
+ value: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
293
+ transformers_version:
294
+ desc: null
295
+ value: 4.39.3
296
+ model_type:
297
+ desc: null
298
+ value: llama
299
+ quantization_config:
300
+ desc: null
301
+ value:
302
+ quant_method: QuantizationMethod.BITS_AND_BYTES
303
+ _load_in_8bit: false
304
+ _load_in_4bit: true
305
+ llm_int8_threshold: 6.0
306
+ llm_int8_skip_modules: null
307
+ llm_int8_enable_fp32_cpu_offload: false
308
+ llm_int8_has_fp16_weight: false
309
+ bnb_4bit_quant_type: nf4
310
+ bnb_4bit_use_double_quant: false
311
+ bnb_4bit_compute_dtype: float16
312
+ bnb_4bit_quant_storage: uint8
313
+ load_in_4bit: true
314
+ load_in_8bit: false
315
+ output_dir:
316
+ desc: null
317
+ value: /kaggle/working/
318
+ overwrite_output_dir:
319
+ desc: null
320
+ value: false
321
+ do_train:
322
+ desc: null
323
+ value: false
324
+ do_eval:
325
+ desc: null
326
+ value: false
327
+ do_predict:
328
+ desc: null
329
+ value: false
330
+ evaluation_strategy:
331
+ desc: null
332
+ value: 'no'
333
+ prediction_loss_only:
334
+ desc: null
335
+ value: false
336
+ per_device_train_batch_size:
337
+ desc: null
338
+ value: 8
339
+ per_device_eval_batch_size:
340
+ desc: null
341
+ value: 8
342
+ per_gpu_train_batch_size:
343
+ desc: null
344
+ value: null
345
+ per_gpu_eval_batch_size:
346
+ desc: null
347
+ value: null
348
+ gradient_accumulation_steps:
349
+ desc: null
350
+ value: 4
351
+ eval_accumulation_steps:
352
+ desc: null
353
+ value: null
354
+ eval_delay:
355
+ desc: null
356
+ value: 0
357
+ learning_rate:
358
+ desc: null
359
+ value: 3.0e-05
360
+ weight_decay:
361
+ desc: null
362
+ value: 0.0001
363
+ adam_beta1:
364
+ desc: null
365
+ value: 0.9
366
+ adam_beta2:
367
+ desc: null
368
+ value: 0.999
369
+ adam_epsilon:
370
+ desc: null
371
+ value: 1.0e-08
372
+ max_grad_norm:
373
+ desc: null
374
+ value: 0.01
375
+ num_train_epochs:
376
+ desc: null
377
+ value: 5
378
+ max_steps:
379
+ desc: null
380
+ value: 3000
381
+ lr_scheduler_type:
382
+ desc: null
383
+ value: cosine
384
+ lr_scheduler_kwargs:
385
+ desc: null
386
+ value: {}
387
+ warmup_ratio:
388
+ desc: null
389
+ value: 0.05
390
+ warmup_steps:
391
+ desc: null
392
+ value: 0
393
+ log_level:
394
+ desc: null
395
+ value: passive
396
+ log_level_replica:
397
+ desc: null
398
+ value: warning
399
+ log_on_each_node:
400
+ desc: null
401
+ value: true
402
+ logging_dir:
403
+ desc: null
404
+ value: /kaggle/working/runs/Apr10_11-48-06_57bca537cf5f
405
+ logging_strategy:
406
+ desc: null
407
+ value: steps
408
+ logging_first_step:
409
+ desc: null
410
+ value: false
411
+ logging_steps:
412
+ desc: null
413
+ value: 30
414
+ logging_nan_inf_filter:
415
+ desc: null
416
+ value: true
417
+ save_strategy:
418
+ desc: null
419
+ value: steps
420
+ save_steps:
421
+ desc: null
422
+ value: 30
423
+ save_total_limit:
424
+ desc: null
425
+ value: 1
426
+ save_safetensors:
427
+ desc: null
428
+ value: true
429
+ save_on_each_node:
430
+ desc: null
431
+ value: false
432
+ save_only_model:
433
+ desc: null
434
+ value: false
435
+ no_cuda:
436
+ desc: null
437
+ value: false
438
+ use_cpu:
439
+ desc: null
440
+ value: false
441
+ use_mps_device:
442
+ desc: null
443
+ value: false
444
+ seed:
445
+ desc: null
446
+ value: 42
447
+ data_seed:
448
+ desc: null
449
+ value: null
450
+ jit_mode_eval:
451
+ desc: null
452
+ value: false
453
+ use_ipex:
454
+ desc: null
455
+ value: false
456
+ bf16:
457
+ desc: null
458
+ value: false
459
+ fp16:
460
+ desc: null
461
+ value: false
462
+ fp16_opt_level:
463
+ desc: null
464
+ value: O1
465
+ half_precision_backend:
466
+ desc: null
467
+ value: auto
468
+ bf16_full_eval:
469
+ desc: null
470
+ value: false
471
+ fp16_full_eval:
472
+ desc: null
473
+ value: false
474
+ tf32:
475
+ desc: null
476
+ value: null
477
+ local_rank:
478
+ desc: null
479
+ value: 0
480
+ ddp_backend:
481
+ desc: null
482
+ value: null
483
+ tpu_num_cores:
484
+ desc: null
485
+ value: null
486
+ tpu_metrics_debug:
487
+ desc: null
488
+ value: false
489
+ debug:
490
+ desc: null
491
+ value: []
492
+ dataloader_drop_last:
493
+ desc: null
494
+ value: false
495
+ eval_steps:
496
+ desc: null
497
+ value: null
498
+ dataloader_num_workers:
499
+ desc: null
500
+ value: 8
501
+ dataloader_prefetch_factor:
502
+ desc: null
503
+ value: null
504
+ past_index:
505
+ desc: null
506
+ value: -1
507
+ run_name:
508
+ desc: null
509
+ value: /kaggle/working/
510
+ disable_tqdm:
511
+ desc: null
512
+ value: false
513
+ remove_unused_columns:
514
+ desc: null
515
+ value: true
516
+ label_names:
517
+ desc: null
518
+ value: null
519
+ load_best_model_at_end:
520
+ desc: null
521
+ value: false
522
+ metric_for_best_model:
523
+ desc: null
524
+ value: null
525
+ greater_is_better:
526
+ desc: null
527
+ value: null
528
+ ignore_data_skip:
529
+ desc: null
530
+ value: false
531
+ fsdp:
532
+ desc: null
533
+ value: []
534
+ fsdp_min_num_params:
535
+ desc: null
536
+ value: 0
537
+ fsdp_config:
538
+ desc: null
539
+ value:
540
+ min_num_params: 0
541
+ xla: false
542
+ xla_fsdp_v2: false
543
+ xla_fsdp_grad_ckpt: false
544
+ fsdp_transformer_layer_cls_to_wrap:
545
+ desc: null
546
+ value: null
547
+ accelerator_config:
548
+ desc: null
549
+ value:
550
+ split_batches: false
551
+ dispatch_batches: null
552
+ even_batches: true
553
+ use_seedable_sampler: true
554
+ deepspeed:
555
+ desc: null
556
+ value: null
557
+ label_smoothing_factor:
558
+ desc: null
559
+ value: 0.0
560
+ optim:
561
+ desc: null
562
+ value: paged_adamw_8bit
563
+ optim_args:
564
+ desc: null
565
+ value: null
566
+ adafactor:
567
+ desc: null
568
+ value: false
569
+ group_by_length:
570
+ desc: null
571
+ value: false
572
+ length_column_name:
573
+ desc: null
574
+ value: length
575
+ report_to:
576
+ desc: null
577
+ value:
578
+ - tensorboard
579
+ - wandb
580
+ ddp_find_unused_parameters:
581
+ desc: null
582
+ value: null
583
+ ddp_bucket_cap_mb:
584
+ desc: null
585
+ value: null
586
+ ddp_broadcast_buffers:
587
+ desc: null
588
+ value: null
589
+ dataloader_pin_memory:
590
+ desc: null
591
+ value: true
592
+ dataloader_persistent_workers:
593
+ desc: null
594
+ value: false
595
+ skip_memory_metrics:
596
+ desc: null
597
+ value: true
598
+ use_legacy_prediction_loop:
599
+ desc: null
600
+ value: false
601
+ push_to_hub:
602
+ desc: null
603
+ value: false
604
+ resume_from_checkpoint:
605
+ desc: null
606
+ value: null
607
+ hub_model_id:
608
+ desc: null
609
+ value: null
610
+ hub_strategy:
611
+ desc: null
612
+ value: every_save
613
+ hub_token:
614
+ desc: null
615
+ value: <HUB_TOKEN>
616
+ hub_private_repo:
617
+ desc: null
618
+ value: false
619
+ hub_always_push:
620
+ desc: null
621
+ value: false
622
+ gradient_checkpointing:
623
+ desc: null
624
+ value: true
625
+ gradient_checkpointing_kwargs:
626
+ desc: null
627
+ value: null
628
+ include_inputs_for_metrics:
629
+ desc: null
630
+ value: false
631
+ fp16_backend:
632
+ desc: null
633
+ value: auto
634
+ push_to_hub_model_id:
635
+ desc: null
636
+ value: null
637
+ push_to_hub_organization:
638
+ desc: null
639
+ value: null
640
+ push_to_hub_token:
641
+ desc: null
642
+ value: <PUSH_TO_HUB_TOKEN>
643
+ mp_parameters:
644
+ desc: null
645
+ value: ''
646
+ auto_find_batch_size:
647
+ desc: null
648
+ value: true
649
+ full_determinism:
650
+ desc: null
651
+ value: false
652
+ torchdynamo:
653
+ desc: null
654
+ value: null
655
+ ray_scope:
656
+ desc: null
657
+ value: last
658
+ ddp_timeout:
659
+ desc: null
660
+ value: 1800
661
+ torch_compile:
662
+ desc: null
663
+ value: false
664
+ torch_compile_backend:
665
+ desc: null
666
+ value: null
667
+ torch_compile_mode:
668
+ desc: null
669
+ value: null
670
+ dispatch_batches:
671
+ desc: null
672
+ value: null
673
+ split_batches:
674
+ desc: null
675
+ value: null
676
+ include_tokens_per_second:
677
+ desc: null
678
+ value: false
679
+ include_num_input_tokens_seen:
680
+ desc: null
681
+ value: false
682
+ neftune_noise_alpha:
683
+ desc: null
684
+ value: null
685
+ optim_target_modules:
686
+ desc: null
687
+ value: null
wandb/run-20240410_115017-r9p6sok2/files/output.log ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
38
+ warnings.warn(
39
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
40
+ warnings.warn(
41
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
42
+ warnings.warn(
43
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
44
+ warnings.warn(
45
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
46
+ warnings.warn(
47
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
48
+ warnings.warn(
49
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
50
+ warnings.warn(
51
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
52
+ warnings.warn(
53
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
54
+ warnings.warn(
55
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
56
+ warnings.warn(
57
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
58
+ warnings.warn(
59
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
60
+ warnings.warn(
61
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
62
+ warnings.warn(
63
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
64
+ warnings.warn(
65
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
66
+ warnings.warn(
67
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
68
+ warnings.warn(
69
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
70
+ warnings.warn(
71
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
72
+ warnings.warn(
73
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
74
+ warnings.warn(
75
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
76
+ warnings.warn(
77
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
78
+ warnings.warn(
79
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
80
+ warnings.warn(
81
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
82
+ warnings.warn(
83
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
84
+ warnings.warn(
85
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
86
+ warnings.warn(
87
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
88
+ warnings.warn(
89
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
90
+ warnings.warn(
91
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
92
+ warnings.warn(
93
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
94
+ warnings.warn(
95
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
96
+ warnings.warn(
97
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
98
+ warnings.warn(
99
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
100
+ warnings.warn(
101
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
102
+ warnings.warn(
103
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
104
+ warnings.warn(
105
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
106
+ warnings.warn(
107
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
108
+ warnings.warn(
109
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
110
+ warnings.warn(
111
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
112
+ warnings.warn(
113
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
114
+ warnings.warn(
115
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
116
+ warnings.warn(
117
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
118
+ warnings.warn(
119
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
120
+ warnings.warn(
121
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
122
+ warnings.warn(
123
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
124
+ warnings.warn(
125
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
126
+ warnings.warn(
127
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
128
+ warnings.warn(
129
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
130
+ warnings.warn(
131
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
132
+ warnings.warn(
133
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
134
+ warnings.warn(
135
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
136
+ warnings.warn(
137
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
138
+ warnings.warn(
139
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
140
+ warnings.warn(
141
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
142
+ warnings.warn(
143
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
144
+ warnings.warn(
145
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
146
+ warnings.warn(
147
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
148
+ warnings.warn(
149
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
150
+ warnings.warn(
151
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
152
+ warnings.warn(
153
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
154
+ warnings.warn(
155
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
156
+ warnings.warn(
157
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
158
+ warnings.warn(
159
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
160
+ warnings.warn(
161
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
162
+ warnings.warn(
163
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
164
+ warnings.warn(
165
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
166
+ warnings.warn(
167
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
168
+ warnings.warn(
169
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
170
+ warnings.warn(
171
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
172
+ warnings.warn(
173
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
174
+ warnings.warn(
175
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
176
+ warnings.warn(
177
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
178
+ warnings.warn(
179
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
180
+ warnings.warn(
181
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
182
+ warnings.warn(
183
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
184
+ warnings.warn(
185
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
186
+ warnings.warn(
187
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
188
+ warnings.warn(
189
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
190
+ warnings.warn(
191
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
192
+ warnings.warn(
193
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
194
+ warnings.warn(
195
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
196
+ warnings.warn(
197
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
198
+ warnings.warn(
199
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
200
+ warnings.warn(
201
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
202
+ warnings.warn(
203
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
204
+ warnings.warn(
205
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
206
+ warnings.warn(
207
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
208
+ warnings.warn(
209
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
210
+ warnings.warn(
211
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
212
+ warnings.warn(
213
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
214
+ warnings.warn(
215
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
216
+ warnings.warn(
217
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
218
+ warnings.warn(
219
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
220
+ warnings.warn(
221
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
222
+ warnings.warn(
223
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
224
+ warnings.warn(
225
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
226
+ warnings.warn(
227
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
228
+ warnings.warn(
229
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
230
+ warnings.warn(
231
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
232
+ warnings.warn(
233
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
234
+ warnings.warn(
wandb/run-20240410_115017-r9p6sok2/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240410_115017-r9p6sok2/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T11:50:18.398347",
5
+ "startedAt": "2024-04-10T11:50:17.776100",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "57bca537cf5f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5569.5014572143555
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240410_115017-r9p6sok2/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 2.778, "train/grad_norm": 0.0, "train/learning_rate": 0.0, "train/epoch": 0.07, "train/global_step": 3000, "_timestamp": 1712782518.9167209, "_runtime": 32701.13459777832, "_step": 100, "train_runtime": 32702.7018, "train_samples_per_second": 2.936, "train_steps_per_second": 0.092, "total_flos": 8.629801849965773e+16, "train_loss": 2.7816907374064126, "_wandb": {"runtime": 32705}}
wandb/run-20240410_115017-r9p6sok2/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240410_115017-r9p6sok2/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 11:50:17,777 INFO MainThread:399 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Configure stats pid to 399
3
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_115017-r9p6sok2/logs/debug.log
10
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_115017-r9p6sok2/logs/debug-internal.log
11
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78febf850df0>
12
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-10 11:50:17,778 INFO MainThread:399 [wandb_init.py:init():617] starting backend
16
+ 2024-04-10 11:50:17,779 INFO MainThread:399 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-10 11:50:17,780 INFO MainThread:399 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-10 11:50:17,781 INFO MainThread:399 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-10 11:50:17,793 INFO MainThread:399 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-10 11:50:18,119 INFO MainThread:399 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-10 11:50:18,122 INFO MainThread:399 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-10 11:50:18,282 INFO MainThread:399 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-10 11:50:18,367 INFO MainThread:399 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-10 11:50:18,367 INFO MainThread:399 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-10 11:50:34,333 INFO MainThread:399 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-10 11:50:34,333 INFO MainThread:399 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-10 11:50:34,335 INFO MainThread:399 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-10 11:50:34,335 INFO MainThread:399 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-10 11:50:34,336 INFO MainThread:399 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-10 11:50:34,341 INFO MainThread:399 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 3000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_11-48-06_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 30, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 30, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
+ 2024-04-10 20:55:23,877 INFO MainThread:399 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-04-10 20:55:23,878 INFO MainThread:399 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240410_115017-r9p6sok2/run-r9p6sok2.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4725c6fe803b6839ef60eb5cc2822245767fe4156105830b798eafc63a1c4a9a
3
+ size 1022364
wandb/run-20240410_210030-g0qpa26d/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_210030-g0qpa26d/files/config.yaml ADDED
@@ -0,0 +1,670 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712782830.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ vocab_size:
62
+ desc: null
63
+ value: 32000
64
+ max_position_embeddings:
65
+ desc: null
66
+ value: 2048
67
+ hidden_size:
68
+ desc: null
69
+ value: 2048
70
+ intermediate_size:
71
+ desc: null
72
+ value: 5632
73
+ num_hidden_layers:
74
+ desc: null
75
+ value: 22
76
+ num_attention_heads:
77
+ desc: null
78
+ value: 32
79
+ num_key_value_heads:
80
+ desc: null
81
+ value: 4
82
+ hidden_act:
83
+ desc: null
84
+ value: silu
85
+ initializer_range:
86
+ desc: null
87
+ value: 0.02
88
+ rms_norm_eps:
89
+ desc: null
90
+ value: 1.0e-05
91
+ pretraining_tp:
92
+ desc: null
93
+ value: 1
94
+ use_cache:
95
+ desc: null
96
+ value: false
97
+ rope_theta:
98
+ desc: null
99
+ value: 10000.0
100
+ rope_scaling:
101
+ desc: null
102
+ value: null
103
+ attention_bias:
104
+ desc: null
105
+ value: false
106
+ attention_dropout:
107
+ desc: null
108
+ value: 0.0
109
+ return_dict:
110
+ desc: null
111
+ value: true
112
+ output_hidden_states:
113
+ desc: null
114
+ value: false
115
+ output_attentions:
116
+ desc: null
117
+ value: false
118
+ torchscript:
119
+ desc: null
120
+ value: false
121
+ torch_dtype:
122
+ desc: null
123
+ value: float32
124
+ use_bfloat16:
125
+ desc: null
126
+ value: false
127
+ tf_legacy_loss:
128
+ desc: null
129
+ value: false
130
+ pruned_heads:
131
+ desc: null
132
+ value: {}
133
+ tie_word_embeddings:
134
+ desc: null
135
+ value: false
136
+ chunk_size_feed_forward:
137
+ desc: null
138
+ value: 0
139
+ is_encoder_decoder:
140
+ desc: null
141
+ value: false
142
+ is_decoder:
143
+ desc: null
144
+ value: false
145
+ cross_attention_hidden_size:
146
+ desc: null
147
+ value: null
148
+ add_cross_attention:
149
+ desc: null
150
+ value: false
151
+ tie_encoder_decoder:
152
+ desc: null
153
+ value: false
154
+ max_length:
155
+ desc: null
156
+ value: 20
157
+ min_length:
158
+ desc: null
159
+ value: 0
160
+ do_sample:
161
+ desc: null
162
+ value: false
163
+ early_stopping:
164
+ desc: null
165
+ value: false
166
+ num_beams:
167
+ desc: null
168
+ value: 1
169
+ num_beam_groups:
170
+ desc: null
171
+ value: 1
172
+ diversity_penalty:
173
+ desc: null
174
+ value: 0.0
175
+ temperature:
176
+ desc: null
177
+ value: 1.0
178
+ top_k:
179
+ desc: null
180
+ value: 50
181
+ top_p:
182
+ desc: null
183
+ value: 1.0
184
+ typical_p:
185
+ desc: null
186
+ value: 1.0
187
+ repetition_penalty:
188
+ desc: null
189
+ value: 1.0
190
+ length_penalty:
191
+ desc: null
192
+ value: 1.0
193
+ no_repeat_ngram_size:
194
+ desc: null
195
+ value: 0
196
+ encoder_no_repeat_ngram_size:
197
+ desc: null
198
+ value: 0
199
+ bad_words_ids:
200
+ desc: null
201
+ value: null
202
+ num_return_sequences:
203
+ desc: null
204
+ value: 1
205
+ output_scores:
206
+ desc: null
207
+ value: false
208
+ return_dict_in_generate:
209
+ desc: null
210
+ value: false
211
+ forced_bos_token_id:
212
+ desc: null
213
+ value: null
214
+ forced_eos_token_id:
215
+ desc: null
216
+ value: null
217
+ remove_invalid_values:
218
+ desc: null
219
+ value: false
220
+ exponential_decay_length_penalty:
221
+ desc: null
222
+ value: null
223
+ suppress_tokens:
224
+ desc: null
225
+ value: null
226
+ begin_suppress_tokens:
227
+ desc: null
228
+ value: null
229
+ architectures:
230
+ desc: null
231
+ value:
232
+ - LlamaForCausalLM
233
+ finetuning_task:
234
+ desc: null
235
+ value: null
236
+ id2label:
237
+ desc: null
238
+ value:
239
+ '0': LABEL_0
240
+ '1': LABEL_1
241
+ label2id:
242
+ desc: null
243
+ value:
244
+ LABEL_0: 0
245
+ LABEL_1: 1
246
+ tokenizer_class:
247
+ desc: null
248
+ value: null
249
+ prefix:
250
+ desc: null
251
+ value: null
252
+ bos_token_id:
253
+ desc: null
254
+ value: 1
255
+ pad_token_id:
256
+ desc: null
257
+ value: null
258
+ eos_token_id:
259
+ desc: null
260
+ value: 2
261
+ sep_token_id:
262
+ desc: null
263
+ value: null
264
+ decoder_start_token_id:
265
+ desc: null
266
+ value: null
267
+ task_specific_params:
268
+ desc: null
269
+ value: null
270
+ problem_type:
271
+ desc: null
272
+ value: null
273
+ _name_or_path:
274
+ desc: null
275
+ value: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
276
+ transformers_version:
277
+ desc: null
278
+ value: 4.39.3
279
+ model_type:
280
+ desc: null
281
+ value: llama
282
+ quantization_config:
283
+ desc: null
284
+ value:
285
+ quant_method: QuantizationMethod.BITS_AND_BYTES
286
+ _load_in_8bit: false
287
+ _load_in_4bit: true
288
+ llm_int8_threshold: 6.0
289
+ llm_int8_skip_modules: null
290
+ llm_int8_enable_fp32_cpu_offload: false
291
+ llm_int8_has_fp16_weight: false
292
+ bnb_4bit_quant_type: nf4
293
+ bnb_4bit_use_double_quant: false
294
+ bnb_4bit_compute_dtype: float16
295
+ bnb_4bit_quant_storage: uint8
296
+ load_in_4bit: true
297
+ load_in_8bit: false
298
+ output_dir:
299
+ desc: null
300
+ value: /kaggle/working/
301
+ overwrite_output_dir:
302
+ desc: null
303
+ value: false
304
+ do_train:
305
+ desc: null
306
+ value: false
307
+ do_eval:
308
+ desc: null
309
+ value: false
310
+ do_predict:
311
+ desc: null
312
+ value: false
313
+ evaluation_strategy:
314
+ desc: null
315
+ value: 'no'
316
+ prediction_loss_only:
317
+ desc: null
318
+ value: false
319
+ per_device_train_batch_size:
320
+ desc: null
321
+ value: 8
322
+ per_device_eval_batch_size:
323
+ desc: null
324
+ value: 8
325
+ per_gpu_train_batch_size:
326
+ desc: null
327
+ value: null
328
+ per_gpu_eval_batch_size:
329
+ desc: null
330
+ value: null
331
+ gradient_accumulation_steps:
332
+ desc: null
333
+ value: 4
334
+ eval_accumulation_steps:
335
+ desc: null
336
+ value: null
337
+ eval_delay:
338
+ desc: null
339
+ value: 0
340
+ learning_rate:
341
+ desc: null
342
+ value: 3.0e-05
343
+ weight_decay:
344
+ desc: null
345
+ value: 0.0001
346
+ adam_beta1:
347
+ desc: null
348
+ value: 0.9
349
+ adam_beta2:
350
+ desc: null
351
+ value: 0.999
352
+ adam_epsilon:
353
+ desc: null
354
+ value: 1.0e-08
355
+ max_grad_norm:
356
+ desc: null
357
+ value: 0.01
358
+ num_train_epochs:
359
+ desc: null
360
+ value: 5
361
+ max_steps:
362
+ desc: null
363
+ value: 10
364
+ lr_scheduler_type:
365
+ desc: null
366
+ value: cosine
367
+ lr_scheduler_kwargs:
368
+ desc: null
369
+ value: {}
370
+ warmup_ratio:
371
+ desc: null
372
+ value: 0.05
373
+ warmup_steps:
374
+ desc: null
375
+ value: 0
376
+ log_level:
377
+ desc: null
378
+ value: passive
379
+ log_level_replica:
380
+ desc: null
381
+ value: warning
382
+ log_on_each_node:
383
+ desc: null
384
+ value: true
385
+ logging_dir:
386
+ desc: null
387
+ value: /kaggle/working/runs/Apr10_20-58-11_57bca537cf5f
388
+ logging_strategy:
389
+ desc: null
390
+ value: steps
391
+ logging_first_step:
392
+ desc: null
393
+ value: false
394
+ logging_steps:
395
+ desc: null
396
+ value: 10
397
+ logging_nan_inf_filter:
398
+ desc: null
399
+ value: true
400
+ save_strategy:
401
+ desc: null
402
+ value: steps
403
+ save_steps:
404
+ desc: null
405
+ value: 10
406
+ save_total_limit:
407
+ desc: null
408
+ value: 1
409
+ save_safetensors:
410
+ desc: null
411
+ value: true
412
+ save_on_each_node:
413
+ desc: null
414
+ value: false
415
+ save_only_model:
416
+ desc: null
417
+ value: false
418
+ no_cuda:
419
+ desc: null
420
+ value: false
421
+ use_cpu:
422
+ desc: null
423
+ value: false
424
+ use_mps_device:
425
+ desc: null
426
+ value: false
427
+ seed:
428
+ desc: null
429
+ value: 42
430
+ data_seed:
431
+ desc: null
432
+ value: null
433
+ jit_mode_eval:
434
+ desc: null
435
+ value: false
436
+ use_ipex:
437
+ desc: null
438
+ value: false
439
+ bf16:
440
+ desc: null
441
+ value: false
442
+ fp16:
443
+ desc: null
444
+ value: false
445
+ fp16_opt_level:
446
+ desc: null
447
+ value: O1
448
+ half_precision_backend:
449
+ desc: null
450
+ value: auto
451
+ bf16_full_eval:
452
+ desc: null
453
+ value: false
454
+ fp16_full_eval:
455
+ desc: null
456
+ value: false
457
+ tf32:
458
+ desc: null
459
+ value: null
460
+ local_rank:
461
+ desc: null
462
+ value: 0
463
+ ddp_backend:
464
+ desc: null
465
+ value: null
466
+ tpu_num_cores:
467
+ desc: null
468
+ value: null
469
+ tpu_metrics_debug:
470
+ desc: null
471
+ value: false
472
+ debug:
473
+ desc: null
474
+ value: []
475
+ dataloader_drop_last:
476
+ desc: null
477
+ value: false
478
+ eval_steps:
479
+ desc: null
480
+ value: null
481
+ dataloader_num_workers:
482
+ desc: null
483
+ value: 8
484
+ dataloader_prefetch_factor:
485
+ desc: null
486
+ value: null
487
+ past_index:
488
+ desc: null
489
+ value: -1
490
+ run_name:
491
+ desc: null
492
+ value: /kaggle/working/
493
+ disable_tqdm:
494
+ desc: null
495
+ value: false
496
+ remove_unused_columns:
497
+ desc: null
498
+ value: true
499
+ label_names:
500
+ desc: null
501
+ value: null
502
+ load_best_model_at_end:
503
+ desc: null
504
+ value: false
505
+ metric_for_best_model:
506
+ desc: null
507
+ value: null
508
+ greater_is_better:
509
+ desc: null
510
+ value: null
511
+ ignore_data_skip:
512
+ desc: null
513
+ value: false
514
+ fsdp:
515
+ desc: null
516
+ value: []
517
+ fsdp_min_num_params:
518
+ desc: null
519
+ value: 0
520
+ fsdp_config:
521
+ desc: null
522
+ value:
523
+ min_num_params: 0
524
+ xla: false
525
+ xla_fsdp_v2: false
526
+ xla_fsdp_grad_ckpt: false
527
+ fsdp_transformer_layer_cls_to_wrap:
528
+ desc: null
529
+ value: null
530
+ accelerator_config:
531
+ desc: null
532
+ value:
533
+ split_batches: false
534
+ dispatch_batches: null
535
+ even_batches: true
536
+ use_seedable_sampler: true
537
+ deepspeed:
538
+ desc: null
539
+ value: null
540
+ label_smoothing_factor:
541
+ desc: null
542
+ value: 0.0
543
+ optim:
544
+ desc: null
545
+ value: paged_adamw_8bit
546
+ optim_args:
547
+ desc: null
548
+ value: null
549
+ adafactor:
550
+ desc: null
551
+ value: false
552
+ group_by_length:
553
+ desc: null
554
+ value: false
555
+ length_column_name:
556
+ desc: null
557
+ value: length
558
+ report_to:
559
+ desc: null
560
+ value:
561
+ - tensorboard
562
+ - wandb
563
+ ddp_find_unused_parameters:
564
+ desc: null
565
+ value: null
566
+ ddp_bucket_cap_mb:
567
+ desc: null
568
+ value: null
569
+ ddp_broadcast_buffers:
570
+ desc: null
571
+ value: null
572
+ dataloader_pin_memory:
573
+ desc: null
574
+ value: true
575
+ dataloader_persistent_workers:
576
+ desc: null
577
+ value: false
578
+ skip_memory_metrics:
579
+ desc: null
580
+ value: true
581
+ use_legacy_prediction_loop:
582
+ desc: null
583
+ value: false
584
+ push_to_hub:
585
+ desc: null
586
+ value: false
587
+ resume_from_checkpoint:
588
+ desc: null
589
+ value: null
590
+ hub_model_id:
591
+ desc: null
592
+ value: null
593
+ hub_strategy:
594
+ desc: null
595
+ value: every_save
596
+ hub_token:
597
+ desc: null
598
+ value: <HUB_TOKEN>
599
+ hub_private_repo:
600
+ desc: null
601
+ value: false
602
+ hub_always_push:
603
+ desc: null
604
+ value: false
605
+ gradient_checkpointing:
606
+ desc: null
607
+ value: true
608
+ gradient_checkpointing_kwargs:
609
+ desc: null
610
+ value: null
611
+ include_inputs_for_metrics:
612
+ desc: null
613
+ value: false
614
+ fp16_backend:
615
+ desc: null
616
+ value: auto
617
+ push_to_hub_model_id:
618
+ desc: null
619
+ value: null
620
+ push_to_hub_organization:
621
+ desc: null
622
+ value: null
623
+ push_to_hub_token:
624
+ desc: null
625
+ value: <PUSH_TO_HUB_TOKEN>
626
+ mp_parameters:
627
+ desc: null
628
+ value: ''
629
+ auto_find_batch_size:
630
+ desc: null
631
+ value: true
632
+ full_determinism:
633
+ desc: null
634
+ value: false
635
+ torchdynamo:
636
+ desc: null
637
+ value: null
638
+ ray_scope:
639
+ desc: null
640
+ value: last
641
+ ddp_timeout:
642
+ desc: null
643
+ value: 1800
644
+ torch_compile:
645
+ desc: null
646
+ value: false
647
+ torch_compile_backend:
648
+ desc: null
649
+ value: null
650
+ torch_compile_mode:
651
+ desc: null
652
+ value: null
653
+ dispatch_batches:
654
+ desc: null
655
+ value: null
656
+ split_batches:
657
+ desc: null
658
+ value: null
659
+ include_tokens_per_second:
660
+ desc: null
661
+ value: false
662
+ include_num_input_tokens_seen:
663
+ desc: null
664
+ value: false
665
+ neftune_noise_alpha:
666
+ desc: null
667
+ value: null
668
+ optim_target_modules:
669
+ desc: null
670
+ value: null
wandb/run-20240410_210030-g0qpa26d/files/output.log ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
wandb/run-20240410_210030-g0qpa26d/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.1
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240410_210030-g0qpa26d/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T21:00:31.409648",
5
+ "startedAt": "2024-04-10T21:00:30.774771",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "57bca537cf5f",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.142,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.142,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.142,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.142,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.142,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5569.839115142822
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 2.7928, "train/grad_norm": 0.0, "train/learning_rate": 0.0, "train/epoch": 0.0, "train/global_step": 10, "_timestamp": 1712782952.4825826, "_runtime": 121.7013726234436, "_step": 1, "train_runtime": 123.5321, "train_samples_per_second": 2.59, "train_steps_per_second": 0.081, "total_flos": 267882145972224.0, "train_loss": 2.792811393737793}
wandb/run-20240410_210030-g0qpa26d/logs/debug-internal.log ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 21:00:30,782 INFO StreamThr :603 [internal.py:wandb_internal():86] W&B internal server running at pid: 603, started at: 2024-04-10 21:00:30.781407
2
+ 2024-04-10 21:00:30,783 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-10 21:00:31,093 INFO WriterThread:603 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/run-g0qpa26d.wandb
4
+ 2024-04-10 21:00:31,094 DEBUG SenderThread:603 [sender.py:send():379] send: header
5
+ 2024-04-10 21:00:31,097 DEBUG SenderThread:603 [sender.py:send():379] send: run
6
+ 2024-04-10 21:00:31,284 INFO SenderThread:603 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files
7
+ 2024-04-10 21:00:31,284 INFO SenderThread:603 [sender.py:_start_run_threads():1124] run started: g0qpa26d with start time 1712782830.78121
8
+ 2024-04-10 21:00:31,291 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-10 21:00:31,292 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-10 21:00:31,382 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-10 21:00:31,393 DEBUG HandlerThread:603 [system_info.py:__init__():26] System info init
12
+ 2024-04-10 21:00:31,394 DEBUG HandlerThread:603 [system_info.py:__init__():41] System info init done
13
+ 2024-04-10 21:00:31,394 INFO HandlerThread:603 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-10 21:00:31,394 INFO SystemMonitor:603 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-10 21:00:31,394 INFO HandlerThread:603 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-10 21:00:31,395 INFO SystemMonitor:603 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-10 21:00:31,396 INFO SystemMonitor:603 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-10 21:00:31,396 INFO SystemMonitor:603 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-10 21:00:31,398 INFO SystemMonitor:603 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-10 21:00:31,399 INFO SystemMonitor:603 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-10 21:00:31,409 DEBUG HandlerThread:603 [system_info.py:probe():150] Probing system
22
+ 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [system_info.py:probe():198] Probing system done
24
+ 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T21:00:31.409648', 'startedAt': '2024-04-10T21:00:30.774771', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '57bca537cf5f', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5569.839115142822}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-10 21:00:31,411 INFO HandlerThread:603 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-10 21:00:31,412 INFO HandlerThread:603 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-10 21:00:31,412 DEBUG HandlerThread:603 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-10 21:00:32,286 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/conda-environment.yaml
29
+ 2024-04-10 21:00:46,426 ERROR HandlerThread:603 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-10 21:00:46,427 DEBUG HandlerThread:603 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-10 21:00:46,427 INFO HandlerThread:603 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-10 21:00:46,432 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-10 21:00:46,434 DEBUG SenderThread:603 [sender.py:send():379] send: files
49
+ 2024-04-10 21:00:46,434 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-10 21:00:46,686 INFO wandb-upload_0:603 [upload_job.py:push():131] Uploaded file /tmp/tmptu_ruxaxwandb/y8c9lp5m-wandb-metadata.json
51
+ 2024-04-10 21:00:47,289 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-metadata.json
52
+ 2024-04-10 21:00:47,385 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-10 21:00:47,385 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-10 21:00:47,389 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
55
+ 2024-04-10 21:00:47,401 DEBUG SenderThread:603 [sender.py:send():379] send: config
56
+ 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: metric
57
+ 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
58
+ 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: metric
59
+ 2024-04-10 21:00:47,403 WARNING SenderThread:603 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-10 21:00:47,403 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
61
+ 2024-04-10 21:00:47,403 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
62
+ 2024-04-10 21:00:47,404 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
63
+ 2024-04-10 21:00:47,406 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-04-10 21:00:48,290 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/requirements.txt
65
+ 2024-04-10 21:00:48,290 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/output.log
66
+ 2024-04-10 21:00:50,291 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/output.log
67
+ 2024-04-10 21:00:51,626 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-10 21:00:56,627 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-04-10 21:01:01,633 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-04-10 21:01:02,296 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/config.yaml
71
+ 2024-04-10 21:01:02,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
72
+ 2024-04-10 21:01:02,388 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
73
+ 2024-04-10 21:01:02,391 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
74
+ 2024-04-10 21:01:07,482 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-04-10 21:01:12,483 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-04-10 21:01:17,386 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
77
+ 2024-04-10 21:01:17,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
78
+ 2024-04-10 21:01:17,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
79
+ 2024-04-10 21:01:17,493 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
80
+ 2024-04-10 21:01:22,494 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-10 21:01:27,495 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-04-10 21:01:31,399 DEBUG SystemMonitor:603 [system_monitor.py:_start():172] Starting system metrics aggregation loop
83
+ 2024-04-10 21:01:31,400 DEBUG SenderThread:603 [sender.py:send():379] send: stats
84
+ 2024-04-10 21:01:32,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
85
+ 2024-04-10 21:01:32,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
86
+ 2024-04-10 21:01:32,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
87
+ 2024-04-10 21:01:32,732 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-04-10 21:01:37,733 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-04-10 21:01:42,734 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
90
+ 2024-04-10 21:01:47,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
91
+ 2024-04-10 21:01:47,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
92
+ 2024-04-10 21:01:47,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
93
+ 2024-04-10 21:01:48,467 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-04-10 21:01:53,468 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
95
+ 2024-04-10 21:01:58,469 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
96
+ 2024-04-10 21:02:01,401 DEBUG SenderThread:603 [sender.py:send():379] send: stats
97
+ 2024-04-10 21:02:02,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
98
+ 2024-04-10 21:02:02,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
99
+ 2024-04-10 21:02:02,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
100
+ 2024-04-10 21:02:04,452 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
101
+ 2024-04-10 21:02:09,454 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
102
+ 2024-04-10 21:02:14,455 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
103
+ 2024-04-10 21:02:17,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
104
+ 2024-04-10 21:02:17,388 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
105
+ 2024-04-10 21:02:17,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
106
+ 2024-04-10 21:02:19,519 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-04-10 21:02:24,520 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-04-10 21:02:29,521 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
109
+ 2024-04-10 21:02:31,403 DEBUG SenderThread:603 [sender.py:send():379] send: stats
110
+ 2024-04-10 21:02:31,907 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: partial_history
111
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
112
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
113
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
114
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
115
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: history
116
+ 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
117
+ 2024-04-10 21:02:31,910 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
118
+ 2024-04-10 21:02:32,337 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json
119
+ 2024-04-10 21:02:32,392 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
120
+ 2024-04-10 21:02:32,392 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
121
+ 2024-04-10 21:02:32,392 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
122
+ 2024-04-10 21:02:32,482 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
123
+ 2024-04-10 21:02:32,524 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: partial_history
124
+ 2024-04-10 21:02:32,525 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
125
+ 2024-04-10 21:02:32,525 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
126
+ 2024-04-10 21:02:32,526 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
127
+ 2024-04-10 21:02:32,526 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
128
+ 2024-04-10 21:02:32,526 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
129
+ 2024-04-10 21:02:32,527 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
130
+ 2024-04-10 21:02:32,527 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
131
+ 2024-04-10 21:02:32,527 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
132
+ 2024-04-10 21:02:32,527 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
133
+ 2024-04-10 21:02:32,528 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
134
+ 2024-04-10 21:02:32,528 DEBUG SenderThread:603 [sender.py:send():379] send: history
135
+ 2024-04-10 21:02:32,528 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
136
+ 2024-04-10 21:02:32,528 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
137
+ 2024-04-10 21:02:33,337 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json
wandb/run-20240410_210030-g0qpa26d/logs/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Configure stats pid to 556
3
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug.log
10
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug-internal.log
11
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ce03cb61630>
12
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():617] starting backend
16
+ 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-10 21:00:30,779 INFO MainThread:556 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-10 21:00:30,780 INFO MainThread:556 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-10 21:00:30,794 INFO MainThread:556 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-10 21:00:31,093 INFO MainThread:556 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-10 21:00:31,096 INFO MainThread:556 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-10 21:00:31,291 INFO MainThread:556 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-10 21:00:47,389 INFO MainThread:556 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-10 21:00:47,395 INFO MainThread:556 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_20-58-11_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_210030-g0qpa26d/run-g0qpa26d.wandb ADDED
Binary file (10.6 kB). View file