pythia-14m_piqa / eval_job_output.txt
Tristan's picture
Training in progress, epoch 0
0aea66e verified
raw
history blame
No virus
14 kB
slurm submission log: 2024-05-19 09:14:41.870170
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631089
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-4203243
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631090
###############################
slurm submission log: 2024-05-19 09:16:21.073119
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631150
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-3010961
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631151
###############################
slurm submission log: 2024-05-19 09:25:08.396129
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631221
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-4399372
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631222
###############################
slurm submission log: 2024-05-19 09:27:21.404018
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631284
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-1884627
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631285
###############################
slurm submission log: 2024-05-19 09:28:18.004148
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631348
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-1636654
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631349
###############################
slurm submission log: 2024-05-19 09:29:20.681949
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631409
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-2759412
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631410
###############################
slurm submission log: 2024-05-19 09:30:34.292755
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631469
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-4781467
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631470
###############################
slurm submission log: 2024-05-19 09:31:39.869529
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631529
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-2913641
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631530
###############################
slurm submission log: 2024-05-19 09:34:33.468022
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631593
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-2085637
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx2
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631594
###############################
slurm submission log: 2024-05-19 09:45:20.755632
created following sbatch script:
###############################
#!/bin/bash
#SBATCH --account=nlp
#SBATCH --cpus-per-task=16
#SBATCH --dependency=afterok:7631663
#SBATCH --gres=gpu:1
#SBATCH --job-name=tthrush-job-1519057
#SBATCH --mem=60G
#SBATCH --nodelist=sphinx1
#SBATCH --open-mode=append
#SBATCH --output=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/eval_job_output.txt
#SBATCH --partition=sphinx
#SBATCH --time=14-0
# activate your desired anaconda environment
. /nlp/scr/tthrush/miniconda3/envs/pretraining-coreset-selection/etc/profile.d/conda.sh ; conda activate pretraining-coreset-selection
# cd to working directory
cd .
# launch commands
srun --unbuffered run_as_child_processes 'lm_eval --model hf --model_args pretrained=/juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1,revision=main,dtype=float16,trust_remote_code=True --tasks xnli_en,xnli_fr,sciq,piqa,lambada,arc_easy --device cuda --output_path /juice5/scr5/tthrush/pretraining-coreset-selection/llm_pretraining/14m_llm_seeds_more_data/pythia-14m_piqa_1/perf'
###############################
submission to slurm complete!
###############################
slurm submission output
Submitted batch job 7631664
###############################