Spaces:
Build error
Build error
ready for eval
Browse files
data/Qwen2-72B-Instruct_metrics.csv
CHANGED
@@ -1,2 +1,8 @@
|
|
1 |
epoch,model,accuracy,precision,recall,f1
|
2 |
-
0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
epoch,model,accuracy,precision,recall,f1
|
2 |
+
0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.7473333333333333,0.804122252986722,0.7473333333333333,0.7607828719113865
|
3 |
+
0.2,Qwen/Qwen2-72B-Instruct/checkpoint-35_torch.bfloat16_4bit_lf,0.7583333333333333,0.8199928526815756,0.7583333333333333,0.782751089787442
|
4 |
+
0.4,Qwen/Qwen2-72B-Instruct/checkpoint-70_torch.bfloat16_4bit_lf,0.7366666666666667,0.8224865755517643,0.7366666666666667,0.7700627366337021
|
5 |
+
0.6,Qwen/Qwen2-72B-Instruct/checkpoint-105_torch.bfloat16_4bit_lf,0.757,0.8253824826209251,0.757,0.784000409833628
|
6 |
+
0.8,Qwen/Qwen2-72B-Instruct/checkpoint-140_torch.bfloat16_4bit_lf,0.7893333333333333,0.8229104753645825,0.7893333333333333,0.8033124955993173
|
7 |
+
1.0,Qwen/Qwen2-72B-Instruct/checkpoint-175_torch.bfloat16_4bit_lf,0.7376666666666667,0.8243654864769323,0.7376666666666667,0.7699617360961548
|
8 |
+
1.2,Qwen/Qwen2-72B-Instruct/checkpoint-210_torch.bfloat16_4bit_lf,0.763,0.8318882808702871,0.763,0.7901075708186186
|
data/Qwen2-72B-Instruct_results.csv
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
notebooks/00_Data Analysis.ipynb
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
notebooks/02b_Qwen2-72B-Instruct_analysis.ipynb
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
scripts/eval-mgtv-qwen2_72b.sh
CHANGED
@@ -15,16 +15,47 @@ grep MemTotal /proc/meminfo
|
|
15 |
|
16 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
17 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
18 |
-
export START_EPOCH=0
|
19 |
export USING_LLAMA_FACTORY=true
|
|
|
20 |
export LOAD_IN_4BIT=true
|
21 |
-
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
22 |
|
|
|
23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
|
|
24 |
|
25 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
26 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
|
29 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
30 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
|
|
15 |
|
16 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
17 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
|
|
18 |
export USING_LLAMA_FACTORY=true
|
19 |
+
export USING_P1_PROMPT_TEMPLATE=false
|
20 |
export LOAD_IN_4BIT=true
|
|
|
21 |
|
22 |
+
export START_EPOCH=7
|
23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
24 |
+
export MODEL_NAME=Qwen/$Qwen2-72B-Instruct
|
25 |
|
26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
27 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
28 |
+
|
29 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
30 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
31 |
+
|
32 |
+
export START_EPOCH=0
|
33 |
+
export MODEL_PREFIX=Llama3.1-70B-Chinese-Chat
|
34 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
35 |
+
|
36 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
37 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
38 |
+
|
39 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
40 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
41 |
+
|
42 |
+
|
43 |
+
export LOAD_IN_4BIT=false
|
44 |
+
|
45 |
+
export MODEL_PREFIX=Llama3.1-8B-Chinese-Chat
|
46 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
47 |
+
|
48 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
49 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
50 |
+
|
51 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
52 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
53 |
+
|
54 |
+
export MODEL_PREFIX=Mistral-7B-v0.3-Chinese-Chat
|
55 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
56 |
+
|
57 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
58 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
59 |
|
60 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
61 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|