dh-mc commited on
Commit
473e849
·
1 Parent(s): 62df289

ready for eval

Browse files
data/Qwen2-72B-Instruct_metrics.csv CHANGED
@@ -1,2 +1,8 @@
1
  epoch,model,accuracy,precision,recall,f1
2
- 0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.7486666666666667,0.8038991443169622,0.7486666666666667,0.7615870542896964
 
 
 
 
 
 
 
1
  epoch,model,accuracy,precision,recall,f1
2
+ 0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.7473333333333333,0.804122252986722,0.7473333333333333,0.7607828719113865
3
+ 0.2,Qwen/Qwen2-72B-Instruct/checkpoint-35_torch.bfloat16_4bit_lf,0.7583333333333333,0.8199928526815756,0.7583333333333333,0.782751089787442
4
+ 0.4,Qwen/Qwen2-72B-Instruct/checkpoint-70_torch.bfloat16_4bit_lf,0.7366666666666667,0.8224865755517643,0.7366666666666667,0.7700627366337021
5
+ 0.6,Qwen/Qwen2-72B-Instruct/checkpoint-105_torch.bfloat16_4bit_lf,0.757,0.8253824826209251,0.757,0.784000409833628
6
+ 0.8,Qwen/Qwen2-72B-Instruct/checkpoint-140_torch.bfloat16_4bit_lf,0.7893333333333333,0.8229104753645825,0.7893333333333333,0.8033124955993173
7
+ 1.0,Qwen/Qwen2-72B-Instruct/checkpoint-175_torch.bfloat16_4bit_lf,0.7376666666666667,0.8243654864769323,0.7376666666666667,0.7699617360961548
8
+ 1.2,Qwen/Qwen2-72B-Instruct/checkpoint-210_torch.bfloat16_4bit_lf,0.763,0.8318882808702871,0.763,0.7901075708186186
data/Qwen2-72B-Instruct_results.csv CHANGED
The diff for this file is too large to render. See raw diff
 
notebooks/00_Data Analysis.ipynb CHANGED
The diff for this file is too large to render. See raw diff
 
notebooks/02b_Qwen2-72B-Instruct_analysis.ipynb CHANGED
The diff for this file is too large to render. See raw diff
 
scripts/eval-mgtv-qwen2_72b.sh CHANGED
@@ -15,16 +15,47 @@ grep MemTotal /proc/meminfo
15
 
16
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
17
  export RESIZE_TOKEN_EMBEDDINGS=true
18
- export START_EPOCH=0
19
  export USING_LLAMA_FACTORY=true
 
20
  export LOAD_IN_4BIT=true
21
- export MODEL_NAME=Qwen/Qwen2-72B-Instruct
22
 
 
23
  export MODEL_PREFIX=Qwen2-72B-Instruct_p2
 
24
 
25
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
26
  export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
27
- export USING_P1_PROMPT_TEMPLATE=false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
30
  python llm_toolkit/eval_logical_reasoning_all_epochs.py
 
15
 
16
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
17
  export RESIZE_TOKEN_EMBEDDINGS=true
 
18
  export USING_LLAMA_FACTORY=true
19
+ export USING_P1_PROMPT_TEMPLATE=false
20
  export LOAD_IN_4BIT=true
 
21
 
22
+ export START_EPOCH=7
23
  export MODEL_PREFIX=Qwen2-72B-Instruct_p2
24
+ export MODEL_NAME=Qwen/$Qwen2-72B-Instruct
25
 
26
  export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
27
  export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
28
+
29
+ echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
30
+ python llm_toolkit/eval_logical_reasoning_all_epochs.py
31
+
32
+ export START_EPOCH=0
33
+ export MODEL_PREFIX=Llama3.1-70B-Chinese-Chat
34
+ export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
35
+
36
+ export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
37
+ export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
38
+
39
+ echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
40
+ python llm_toolkit/eval_logical_reasoning_all_epochs.py
41
+
42
+
43
+ export LOAD_IN_4BIT=false
44
+
45
+ export MODEL_PREFIX=Llama3.1-8B-Chinese-Chat
46
+ export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
47
+
48
+ export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
49
+ export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
50
+
51
+ echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
52
+ python llm_toolkit/eval_logical_reasoning_all_epochs.py
53
+
54
+ export MODEL_PREFIX=Mistral-7B-v0.3-Chinese-Chat
55
+ export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
56
+
57
+ export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
58
+ export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
59
 
60
  echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
61
  python llm_toolkit/eval_logical_reasoning_all_epochs.py