slurm submission log: 2024-05-19 09:14:41.566584 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-2782455 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 1024 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631089 ############################### slurm submission log: 2024-05-19 09:16:20.758126 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-914626 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 1024 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631150 ############################### slurm submission log: 2024-05-19 09:25:08.150536 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-758301 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 768 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631221 ############################### slurm submission log: 2024-05-19 09:27:21.164411 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-4168844 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 640 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631284 ############################### slurm submission log: 2024-05-19 09:28:17.764747 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-3776623 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 512 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631348 ############################### slurm submission log: 2024-05-19 09:29:20.389708 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-4689387 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 256 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631409 ############################### slurm submission log: 2024-05-19 09:30:33.987595 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-1409388 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 340 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631469 ############################### slurm submission log: 2024-05-19 09:31:39.572545 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-1319359 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 320 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631529 ############################### slurm submission log: 2024-05-19 09:34:33.193118 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-742694 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 256 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631593 ############################### slurm submission log: 2024-05-19 09:45:20.459710 created following sbatch script: ############################### #!/bin/bash #SBATCH --account=nlp #SBATCH --cpus-per-task=16 #SBATCH --gres=gpu:1 #SBATCH --job-name=tthrush-job-3186286 #SBATCH --mem=100G #SBATCH --nodelist=sphinx2 #SBATCH --open-mode=append #SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/train_job_output.txt #SBATCH --partition=sphinx #SBATCH --time=14-0 # activate your desired anaconda environment . /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection # cd to working directory cd . # launch commands srun --unbuffered run_as_child_processes 'torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 320 --seed 1 --num_train_epochs 0.2' ############################### submission to slurm complete! ############################### slurm submission output Submitted batch job 7631663 ############################### /var/lib/slurm/slurmd/job7631663/slurm_script: line 15: /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh: No such file or directory CommandNotFoundError: Your shell has not been properly configured to use 'conda activate'. To initialize your shell, run $ conda init Currently supported shells are: - bash - fish - tcsh - xonsh - zsh - powershell See 'conda init --help' for more information and options. IMPORTANT: You may need to close and restart your shell after running 'conda init'. ############################### start time: 2024-05-19 10:14:28.271980 machine: sphinx2 conda env: pretraining-coreset-selection ############################### running following processes torchrun --master_port 29517 --nproc_per_node=1 train_llm.py --dataset_id /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa --output_dir /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1 --output_hub_id pythia-14m_piqa --model_id EleutherAI/pythia-14m --learning_rate 6e-3 --warmup_ratio=0.1 --gradient_accumulation_steps 1 --per_device_train_batch_size 320 --seed 1 --num_train_epochs 0.2 ############################### command outputs: 05/19/2024 10:14:33 - INFO - __main__ - Script parameters ScriptArguments(seed=1, dataset_id='/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/train_data_more_data/piqa', output_dir='/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1', output_hub_id='pythia-14m_piqa', hf_hub_token=True, model_id='EleutherAI/pythia-14m', per_device_train_batch_size=320, num_train_epochs=0.2, learning_rate=0.006, gradient_accumulation_steps=1, from_scratch=True, warmup_ratio=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, weight_decay=0.01, lr_scheduler_type='cosine', local_rank=0, resume_from_checkpoint=False, deepspeed=None, peft=False) 0%| | 0/6837 [00:00