DPOpenHermes-11B / README.md
winglian's picture
Create README.md
6b78354
|
raw
history blame contribute delete
No virus
533 Bytes
---
license: apache-2.0
datasets:
- teknium/openhermes
- argilla/ultrafeedback-binarized-preferences
- Intel/orca_dpo_pairs
language:
- en
library_name: transformers
---
# DPOpenHermes 11B
This is a mergekit merge of DPOpenHermes-7B from seperate versions of it.
```
slices:
- sources:
- model: openaccess-ai-collective/DPOpenHermes-7B
revision: dpo-v0
layer_range: [0, 24]
- sources:
- model: openaccess-ai-collective/DPOpenHermes-7B
layer_range: [8, 32]
merge_method: passthrough
dtype: bfloat16
```