File size: 2,041 Bytes
00585cb
01acdfd
 
 
 
 
 
 
 
 
 
 
00585cb
01acdfd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
00585cb
01acdfd
 
 
 
7467365
01acdfd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
---
language:
- en
base_model:
- meta-llama/Meta-Llama-3-8B-Instruct
library_name: transformers
tags:
- meta
- llama-3
- pytorch
- mergekit
- merge
license: llama3
license_link: LICENSE
pipeline_tag: text-generation
widget:
  - example_title: Hello
    messages:
    - role: user
      content: Hey my name is Corwin! How are you?
  - example_title: Hellriding out of Amber
    messages:
    - role: system
      content: You are a helpful and honest assistant. Please, respond concisely and truthfully.
    - role: user
      content: Can you recommend a good destination for a hellride out of Amber?
inference:
  parameters:
    max_new_tokens: 300
    stop:
    - <|end_of_text|>
    - <|eot_id|>
---
# llama-3-experiment-v1-9B-GGUF

This is an experimental merge, replicating additional layers to the model without post-merge healing. There is damage to the model, but it appears to be tolerable as is. The resulting impact on narrative text completion may be of interest.

Light testing performed with instruct prompting and the following sampler settings:
- temp=1 and minP=0.02
- temp=1 and smoothing factor=0.33

Full weights: [grimjim/llama-3-experiment-v1-9B](https://huggingface.co/grimjim/llama-3-experiment-v1-9B)

GGUF quants: [grimjim/llama-3-experiment-v1-9B-GGUF](https://huggingface.co/grimjim/llama-3-experiment-v1-9B-GGUF)

This is a merge of pre-trained language model meta-llama/Meta-Llama-3-8B-Instruct created using [mergekit](https://github.com/cg123/mergekit).

Built with Meta Llama 3.

## Merge Details
### Merge Method

This model was merged using the passthrough merge method.

### Models Merged

The following models were included in the merge:
* meta-llama/Meta-Llama-3-8B-Instruct

### Configuration

The following YAML configuration was used to produce this model:

```yaml
slices:
  - sources:
    - model: meta-llama/Meta-Llama-3-8B-Instruct
      layer_range: [0, 12]
  - sources:
    - model: meta-llama/Meta-Llama-3-8B-Instruct
      layer_range: [8, 32]
merge_method: passthrough
dtype: bfloat16

```