GGUF
mergekit
Merge
Not-For-All-Audiences
nsfw
File size: 2,351 Bytes
c817d7d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
---
base_model:
- NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3
- rombodawg/Open_Gpt4_8x7B_v0.2
- mistralai/Mixtral-8x7B-Instruct-v0.1
tags:
- mergekit
- merge
- not-for-all-audiences
- nsfw
license: cc-by-nc-4.0
---

<!-- description start -->
## Description

This repo contains fp16 files of NoromaidxOpenGPT4-2.

The model was created by merging Noromaid-8x7b-Instruct with Open_Gpt4_8x7B_v0.2 the exact same way [Rombodawg](https://huggingface.co/rombodawg) done his merge.

The only difference between [NoromaidxOpenGPT4-1](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-1-GGUF-iMatrix/) and [NoromaidxOpenGPT4-2](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2-GGUF-iMatrix/) is that the first iteration use Mixtral-8x7B as a base for the merge (f16), where the second use Open_Gpt4_8x7B_v0.2 as a base (bf16).

After further testing and usage, the two model was released, because they each have their own qualities.

You can download the imatrix file to do many other quant [HERE](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2/blob/main/imatrix-2.dat).
<!-- description end -->
<!-- prompt-template start -->
### Prompt template:

## Alpaca

```
### Instruction:
{system prompt}

### Input:
{prompt}

### Response:
{output}
```

## Mistral

```
[INST] {prompt} [/INST] 
```

## Merge Details
### Merge Method

This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [rombodawg/Open_Gpt4_8x7B_v0.2](https://huggingface.co/rombodawg/Open_Gpt4_8x7B_v0.2) as a base.

### Models Merged

The following models were included in the merge:
* [NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3](https://huggingface.co/NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3)
* [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: mistralai/Mixtral-8x7B-Instruct-v0.1
    parameters:
      density: .5
      weight: 1
  - model: NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3
    parameters:
      density: .5
      weight: .7
merge_method: ties
base_model: rombodawg/Open_Gpt4_8x7B_v0.2
parameters:
  normalize: true
  int8_mask: true
dtype: bfloat16
```

### Support

If you want to support us, you can [here](https://ko-fi.com/undiai).