File size: 3,050 Bytes
ac444da
 
 
 
 
 
 
 
5feb78b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
05c8ca0
 
5feb78b
 
154578c
 
 
 
 
 
 
 
 
ba40d8c
5feb78b
 
ac444da
 
 
5feb78b
ac444da
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
---
license: apache-2.0
tags:
- merge
- mergekit
- vilm/vinallama-7b-chat
---

# VinaLLaMA - State-of-the-art Vietnamese LLMs

![image](https://i.ibb.co/W0dq12n/vinallama.png)

Read our [Paper](https://huggingface.co/papers/2312.11011)

Prompt Format (ChatML):

```
<|im_start|>system
Bạn là một trợ lí AI hữu ích. Hãy trả lời người dùng một cách chính xác.
<|im_end|>
<|im_start|>user
Hello world!<|im_end|>
<|im_start|>assistant
```

## Evaluation

This table is copied from VBD-Llama2 with updated results from VinaLLaMA-12.5B-chat-DUS

| Model                    | Model size | arc_vi (acc) | hellaswag_vi (acc) | mmlu_vi (acc) | truthfulqa_vi (acc) | Average |
| ------------------------ | ---------- | ------------ | ------------------ | ------------- | ------------------- | ------- |
| URA-LLaMA-13B            | 13B        | 0,3752       | 0,4830             | 0,3973        | 0,4574              | 0,4282  |
| BLOOMZ-7B                | 7B         | 0,3205       | 0,4930             | <b>0,3975</b> | 0,4523              | 0,4158  |
| PhoGPT-7B5-Instruct      | 7B         | 0,2470       | 0,2578             | 0,2413        | 0,4759              | 0,3055  |
| SeaLLM-7B-chat           | 7B         | 0,3607       | 0,5112             | 0,3339        | 0,4948              | 0,4252  |
| Vietcuna-7b-v3           | 7B         | 0,3419       | 0,4939             | 0,3354        | 0,4807              | 0,4130  |
| VinaLLaMA-2.7B-chat      | 7B         | 0,3273       | 0,4814             | 0,3051        | 0,4972              | 0,4028  |
| VinaLLaMA-7B-chat        | 7B         | 0,4239       | 0,5407             | 0,3932        | 0,5251              | 0,4707  |
| VBD-LLaMA2-7B-50b        | 7B         | 0,3222       | 0,5195             | 0,2964        | 0,4614              | 0,3999  |
| VBD-LLaMA2-7B-50b-Chat   | 7B         | 0,3585       | 0,5207             | 0,3444        | 0,5179              | 0,4354  |
| <b>VinaLLaMA-12.5B-chat-DUS</b> | 12.5B      | <b>0,4325</b>| <b>0,5816</b>      | 0,3875        | <b>0,5850</b>       | <b>0,4967</b>  |

# Merging Methods

This model is a merge of the following models made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
 * [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)


## 🧩 Configuration

```yaml
slices:
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [0, 16]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [8, 16]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [8, 16]      
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [16, 24]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [16, 24]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [24, 28]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [24, 28]
  - sources:
    - model: vilm/vinallama-7b-chat
      layer_range: [28, 32]
merge_method: passthrough
dtype: bfloat16
```