Spaces:
				
			
			
	
			
			
		Build error
		
	
	
	
			
			
	
	
	
	
		
		
		Build error
		
	llama3 p2 epoch 1 results
Browse files
    	
        llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml
    CHANGED
    
    | 
         @@ -25,7 +25,7 @@ output_dir: saves/llama3-8b/lora/sft_bf16_p1_full 
     | 
|
| 25 | 
         
             
            logging_steps: 10
         
     | 
| 26 | 
         
             
            save_steps: 175
         
     | 
| 27 | 
         
             
            plot_loss: true
         
     | 
| 28 | 
         
            -
             
     | 
| 29 | 
         | 
| 30 | 
         
             
            ### train
         
     | 
| 31 | 
         
             
            per_device_train_batch_size: 16
         
     | 
| 
         | 
|
| 25 | 
         
             
            logging_steps: 10
         
     | 
| 26 | 
         
             
            save_steps: 175
         
     | 
| 27 | 
         
             
            plot_loss: true
         
     | 
| 28 | 
         
            +
            overwrite_output_dir: true
         
     | 
| 29 | 
         | 
| 30 | 
         
             
            ### train
         
     | 
| 31 | 
         
             
            per_device_train_batch_size: 16
         
     | 
    	
        results/llama3-8b_lora_sft_bf16-p2.csv
    CHANGED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        scripts/eval-mgtv-llama3_8b.sh
    CHANGED
    
    | 
         @@ -17,7 +17,6 @@ grep MemTotal /proc/meminfo 
     | 
|
| 17 | 
         
             
            #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
         
     | 
| 18 | 
         | 
| 19 | 
         
             
            export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
         
     | 
| 20 | 
         
            -
            export START_EPOCH=1
         
     | 
| 21 | 
         
             
            export RESIZE_TOKEN_EMBEDDINGS=true
         
     | 
| 22 | 
         | 
| 23 | 
         
             
            #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
         
     | 
| 
         @@ -27,11 +26,13 @@ export MODEL_PREFIX=llama3-8b_lora_sft_bf16 
     | 
|
| 27 | 
         
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
         
     | 
| 28 | 
         
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
         
     | 
| 29 | 
         
             
            export USING_P1_PROMPT_TEMPLATE=true
         
     | 
| 
         | 
|
| 30 | 
         
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         
     | 
| 31 | 
         
            -
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         
     | 
| 32 | 
         | 
| 33 | 
         
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
         
     | 
| 34 | 
         
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
         
     | 
| 35 | 
         
             
            export USING_P1_PROMPT_TEMPLATE=false
         
     | 
| 
         | 
|
| 36 | 
         
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         
     | 
| 37 | 
         
             
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         
     | 
| 
         | 
|
| 17 | 
         
             
            #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
         
     | 
| 18 | 
         | 
| 19 | 
         
             
            export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
         
     | 
| 
         | 
|
| 20 | 
         
             
            export RESIZE_TOKEN_EMBEDDINGS=true
         
     | 
| 21 | 
         | 
| 22 | 
         
             
            #export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
         
     | 
| 
         | 
|
| 26 | 
         
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
         
     | 
| 27 | 
         
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
         
     | 
| 28 | 
         
             
            export USING_P1_PROMPT_TEMPLATE=true
         
     | 
| 29 | 
         
            +
            export START_EPOCH=3
         
     | 
| 30 | 
         
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         
     | 
| 31 | 
         
            +
            #python llm_toolkit/eval_logical_reasoning_all_epochs.py
         
     | 
| 32 | 
         | 
| 33 | 
         
             
            export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
         
     | 
| 34 | 
         
             
            export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
         
     | 
| 35 | 
         
             
            export USING_P1_PROMPT_TEMPLATE=false
         
     | 
| 36 | 
         
            +
            export START_EPOCH=1
         
     | 
| 37 | 
         
             
            echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
         
     | 
| 38 | 
         
             
            python llm_toolkit/eval_logical_reasoning_all_epochs.py
         
     |