Update README.md
Browse files
README.md
CHANGED
@@ -113,18 +113,6 @@ model-index:
|
|
113 |
|
114 |
# dfurman/Qwen2-72B-Orpo-v0.1
|
115 |
|
116 |
-
## Introduction
|
117 |
-
|
118 |
-
> Qwen2 is the new series of Qwen large language models. For Qwen2, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters, including a Mixture-of-Experts model. This repo contains the instruction-tuned 72B Qwen2 model.
|
119 |
-
>
|
120 |
-
> Compared with the state-of-the-art opensource language models, including the previous released Qwen1.5, Qwen2 has generally surpassed most opensource models and demonstrated competitiveness against proprietary models across a series of benchmarks targeting for language understanding, language generation, multilingual capability, coding, mathematics, reasoning, etc.
|
121 |
-
>
|
122 |
-
> Qwen2-72B-Instruct supports a context length of up to 131,072 tokens, enabling the processing of extensive inputs. Please refer to [this section](#processing-long-texts) for detailed instructions on how to deploy Qwen2 for handling long texts.
|
123 |
-
>
|
124 |
-
> For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2/), [GitHub](https://github.com/QwenLM/Qwen2), and [Documentation](https://qwen.readthedocs.io/en/latest/).
|
125 |
-
>
|
126 |
-
> - [Qwen2](https://huggingface.co/Qwen/Qwen2-72B-Instruct)
|
127 |
-
|
128 |
## This finetune
|
129 |
|
130 |
Qwen2-72B-Orpo-v0.1 is a QLoRA finetune of `Qwen/Qwen2-72B-Instruct` on 1.5k rows of `mlabonne/orpo-dpo-mix-40k`.
|
|
|
113 |
|
114 |
# dfurman/Qwen2-72B-Orpo-v0.1
|
115 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
116 |
## This finetune
|
117 |
|
118 |
Qwen2-72B-Orpo-v0.1 is a QLoRA finetune of `Qwen/Qwen2-72B-Instruct` on 1.5k rows of `mlabonne/orpo-dpo-mix-40k`.
|