qnguyen3 commited on
Commit
05c8ca0
1 Parent(s): b040da0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -25,9 +25,8 @@ Hello world!<|im_end|>
25
 
26
  ## Evaluation
27
 
28
- We evaluated our model via peer comparison on multiple publicly available dataset using
29
- <a href="https://github.com/hieunguyen1053/lm-evaluation-harness"> @hieunguyen1053 fork of lm-evaluation-harness </a>
30
- , and combine the results with that provided by the authors of VinaLLaMA. The results are bellow:
31
  | Model | Model size | arc_vi (acc) | hellaswag_vi (acc) | mmlu_vi (acc) | truthfulqa_vi (acc) | Average |
32
  | ------------------------ | ---------- | ------------ | ------------------ | ------------- | ------------------- | ------- |
33
  | URA-LLaMA-13B | | 0,3752 | 0,4830 | 0,3973 | 0,4574 | 0,4282 |
 
25
 
26
  ## Evaluation
27
 
28
+ This table is copied from VBD-Llama2 with updated results from VinaLLaMA-12.5B-chat-DUS
29
+
 
30
  | Model | Model size | arc_vi (acc) | hellaswag_vi (acc) | mmlu_vi (acc) | truthfulqa_vi (acc) | Average |
31
  | ------------------------ | ---------- | ------------ | ------------------ | ------------- | ------------------- | ------- |
32
  | URA-LLaMA-13B | | 0,3752 | 0,4830 | 0,3973 | 0,4574 | 0,4282 |