inflaton commited on
Commit
be560ea
·
1 Parent(s): 71e4b14

internlm2 best results

Browse files
llm_toolkit/eval_logical_reasoning.py CHANGED
@@ -91,10 +91,10 @@ print(f"(3) GPU = {gpu_stats.name}. Max memory = {max_memory} GB.")
91
  print(f"{start_gpu_memory} GB of memory reserved.")
92
 
93
  if adapter_name_or_path is not None:
94
- model_name += "_" + adapter_name_or_path.split("/")[-1]
95
 
96
  save_results(
97
- model_name,
98
  results_path,
99
  datasets["test"],
100
  predictions,
 
91
  print(f"{start_gpu_memory} GB of memory reserved.")
92
 
93
  if adapter_name_or_path is not None:
94
+ model_name += "/" + adapter_name_or_path.split("/")[-1]
95
 
96
  save_results(
97
+ f"{model_name}_{dtype}{'_4bit' if load_in_4bit else ''}",
98
  results_path,
99
  datasets["test"],
100
  predictions,
results/mgtv-results_internlm_best.csv ADDED
The diff for this file is too large to render. See raw diff
 
scripts/eval-mgtv-best.sh CHANGED
@@ -17,14 +17,30 @@ grep MemTotal /proc/meminfo
17
  #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes] && cd $BASEDIR
18
  #pip install transformers==4.41.2
19
 
20
- export USING_LLAMA_FACTORY=true
 
21
 
22
  export MODEL_NAME=internlm/internlm2_5-7b-chat-1m
23
- export ADAPTER_NAME_OR_PATH=llama-factory/saves/internlm2_5_7b/lora/sft_bf16_p2_full/checkpoint-88
24
- #export ADAPTER_NAME_OR_PATH=inflaton-ai/InternLM_2_5-7b_LoRA-Adapter
25
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
26
  export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_internlm_best.csv
 
27
  export USE_FLOAT32_FOR_INFERENCE=true
 
 
 
 
 
 
 
 
 
 
 
 
28
 
 
 
 
29
  echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
30
  python llm_toolkit/eval_logical_reasoning.py
 
17
  #cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes] && cd $BASEDIR
18
  #pip install transformers==4.41.2
19
 
20
+ export USING_LLAMA_FACTORY=false
21
+ export LOAD_IN_4BIT=false
22
 
23
  export MODEL_NAME=internlm/internlm2_5-7b-chat-1m
24
+ export ADAPTER_NAME_OR_PATH=inflaton-ai/InternLM_2_5-7b_LoRA-Adapter
 
25
  export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
26
  export LOGICAL_REASONING_RESULTS_PATH=results/mgtv-results_internlm_best.csv
27
+
28
  export USE_FLOAT32_FOR_INFERENCE=true
29
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
30
+ python llm_toolkit/eval_logical_reasoning.py
31
+
32
+ export USE_FLOAT32_FOR_INFERENCE=false
33
+ export USE_BF16_FOR_INFERENCE=true
34
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
35
+ python llm_toolkit/eval_logical_reasoning.py
36
+
37
+ export USE_FLOAT32_FOR_INFERENCE=false
38
+ export USE_BF16_FOR_INFERENCE=false
39
+ echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
40
+ python llm_toolkit/eval_logical_reasoning.py
41
 
42
+ export LOAD_IN_4BIT=true
43
+ export USE_FLOAT32_FOR_INFERENCE=false
44
+ export USE_BF16_FOR_INFERENCE=true
45
  echo "Eval $MODEL_NAME with $ADAPTER_NAME_OR_PATH"
46
  python llm_toolkit/eval_logical_reasoning.py
scripts/eval-mgtv.sh CHANGED
@@ -1 +1 @@
1
- eval-mgtv-internlm.sh
 
1
+ eval-mgtv-best.sh