Azazelle commited on
Commit
4cedade
1 Parent(s): 197c87d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
 
2
  base_model:
 
3
  - akjindal53244/Mistral-7B-v0.1-Open-Platypus
4
  - WizardLM/WizardMath-7B-V1.1
5
- - Open-Orca/Mistral-7B-OpenOrca
6
  - mistralai/Mistral-7B-v0.1
7
  library_name: transformers
8
  tags:
9
  - mergekit
10
  - merge
11
-
12
  ---
13
- # Moko-DARE
14
 
15
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
@@ -22,9 +23,9 @@ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](
22
  ### Models Merged
23
 
24
  The following models were included in the merge:
 
25
  * [akjindal53244/Mistral-7B-v0.1-Open-Platypus](https://huggingface.co/akjindal53244/Mistral-7B-v0.1-Open-Platypus)
26
  * [WizardLM/WizardMath-7B-V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1)
27
- * [Open-Orca/Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)
28
 
29
  ### Configuration
30
 
 
1
  ---
2
+ pipeline_tag: text-generation
3
  base_model:
4
+ - Open-Orca/Mistral-7B-OpenOrca
5
  - akjindal53244/Mistral-7B-v0.1-Open-Platypus
6
  - WizardLM/WizardMath-7B-V1.1
 
7
  - mistralai/Mistral-7B-v0.1
8
  library_name: transformers
9
  tags:
10
  - mergekit
11
  - merge
12
+ license: cc-by-4.0
13
  ---
14
+ # Mocha-Dare-7b-ex
15
 
16
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
 
 
23
  ### Models Merged
24
 
25
  The following models were included in the merge:
26
+ * [Open-Orca/Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)
27
  * [akjindal53244/Mistral-7B-v0.1-Open-Platypus](https://huggingface.co/akjindal53244/Mistral-7B-v0.1-Open-Platypus)
28
  * [WizardLM/WizardMath-7B-V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1)
 
29
 
30
  ### Configuration
31