File size: 9,273 Bytes
8700a59 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
youri-7b-instruction - GGUF
- Model creator: https://huggingface.co/rinna/
- Original model: https://huggingface.co/rinna/youri-7b-instruction/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [youri-7b-instruction.Q2_K.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q2_K.gguf) | Q2_K | 2.36GB |
| [youri-7b-instruction.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.IQ3_XS.gguf) | IQ3_XS | 2.6GB |
| [youri-7b-instruction.IQ3_S.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.IQ3_S.gguf) | IQ3_S | 2.75GB |
| [youri-7b-instruction.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q3_K_S.gguf) | Q3_K_S | 2.75GB |
| [youri-7b-instruction.IQ3_M.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.IQ3_M.gguf) | IQ3_M | 2.9GB |
| [youri-7b-instruction.Q3_K.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q3_K.gguf) | Q3_K | 3.07GB |
| [youri-7b-instruction.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q3_K_M.gguf) | Q3_K_M | 3.07GB |
| [youri-7b-instruction.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q3_K_L.gguf) | Q3_K_L | 3.35GB |
| [youri-7b-instruction.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.IQ4_XS.gguf) | IQ4_XS | 3.4GB |
| [youri-7b-instruction.Q4_0.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q4_0.gguf) | Q4_0 | 3.56GB |
| [youri-7b-instruction.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.IQ4_NL.gguf) | IQ4_NL | 3.58GB |
| [youri-7b-instruction.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q4_K_S.gguf) | Q4_K_S | 3.59GB |
| [youri-7b-instruction.Q4_K.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q4_K.gguf) | Q4_K | 3.8GB |
| [youri-7b-instruction.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q4_K_M.gguf) | Q4_K_M | 3.8GB |
| [youri-7b-instruction.Q4_1.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q4_1.gguf) | Q4_1 | 3.95GB |
| [youri-7b-instruction.Q5_0.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q5_0.gguf) | Q5_0 | 4.33GB |
| [youri-7b-instruction.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q5_K_S.gguf) | Q5_K_S | 4.33GB |
| [youri-7b-instruction.Q5_K.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q5_K.gguf) | Q5_K | 4.45GB |
| [youri-7b-instruction.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q5_K_M.gguf) | Q5_K_M | 4.45GB |
| [youri-7b-instruction.Q5_1.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q5_1.gguf) | Q5_1 | 4.72GB |
| [youri-7b-instruction.Q6_K.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q6_K.gguf) | Q6_K | 5.15GB |
| [youri-7b-instruction.Q8_0.gguf](https://huggingface.co/RichardErkhov/rinna_-_youri-7b-instruction-gguf/blob/main/youri-7b-instruction.Q8_0.gguf) | Q8_0 | 6.67GB |
Original model description:
---
thumbnail: https://github.com/rinnakk/japanese-pretrained-models/blob/master/rinna.png
license: llama2
language:
- ja
- en
inference: false
datasets:
- databricks/databricks-dolly-15k
- kunishou/databricks-dolly-15k-ja
- izumi-lab/llm-japanese-dataset
---
# `rinna/youri-7b-instruction`
![rinna-icon](./rinna.png)
# Overview
The model is the instruction-tuned version of [`rinna/youri-7b`](https://huggingface.co/rinna/youri-7b). It adopts the Alpaca input format.
* **Model architecture**
A 32-layer, 4096-hidden-size transformer-based language model. Refer to the [llama2 paper](https://arxiv.org/abs/2307.09288) for architecture details.
* **Fine-tuning**
The fine-tuning data is the subset of the following datasets.
* [Databricks Dolly data](https://huggingface.co/datasets/databricks/databricks-dolly-15k)
* [Japanese Databricks Dolly data](https://huggingface.co/datasets/kunishou/databricks-dolly-15k-ja)
* [FLAN Instruction Tuning data](https://github.com/google-research/FLAN) and its Japanese translation
* [Izumi lab LLM Japanese dataset](https://github.com/masanorihirano/llm-japanese-dataset/tree/main)
* The following sections are used
* alt
* aozora-txt
* CourseraParallel
* ParaNatCom
* Tab-delimited_Bilingual_Sentence_Pairs
* tanaka-corpus
* wikinews
* wordnet
* yasashi-japanese
* The [remaining sections](https://github.com/masanorihirano/llm-japanese-dataset/tree/main/datasets-cc-by-sa) contain commonly used evaluation corpora so they are skipped to prevent data leak.
* **Contributors**
- [Tianyu Zhao](https://huggingface.co/tianyuz)
- [Kei Sawada](https://huggingface.co/keisawada)
---
# Benchmarking
Please refer to [rinna's LM benchmark page](https://rinnakk.github.io/research/benchmarks/lm/index.html).
---
# How to use the model
~~~~python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("rinna/youri-7b-instruction")
model = AutoModelForCausalLM.from_pretrained("rinna/youri-7b-instruction")
if torch.cuda.is_available():
model = model.to("cuda")
instruction = "次の日本語を英語に翻訳してください。"
input = "大規模言語モデル(だいきぼげんごモデル、英: large language model、LLM)は、多数のパラメータ(数千万から数十億)を持つ人工ニューラルネットワークで構成されるコンピュータ言語モデルで、膨大なラベルなしテキストを使用して自己教師あり学習または半教師あり学習によって訓練が行われる。"
prompt = f"""
以下は、タスクを説明する指示と、文脈のある入力の組み合わせです。要求を適切に満たす応答を書きなさい。
### 指示:
{instruction}
### 入力:
{input}
### 応答:
"""
token_ids = tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt")
with torch.no_grad():
output_ids = model.generate(
token_ids.to(model.device),
max_new_tokens=200,
do_sample=True,
temperature=0.5,
pad_token_id=tokenizer.pad_token_id,
bos_token_id=tokenizer.bos_token_id,
eos_token_id=tokenizer.eos_token_id
)
output = tokenizer.decode(output_ids.tolist()[0])
print(output)
"""
以下は、タスクを説明する指示と、文脈のある入力の組み合わせです。要求を適切に満たす応答を書きなさい。
### 指示:
次の日本語を英語に翻訳してください。
### 入力:
大規模言語モデル(だいきぼげんごモデル、英: large language model、LLM)は、多数のパラメータ(数千万から数十億)を持つ人工ニューラルネットワークで構成されるコンピュータ言語モデルで、膨大なラベルなしテキストを使 用して自己教師あり学習または半教師あり学習によって訓練が行われる。
### 応答:
Large language models (LLMs) are computer language models that are composed of artificial neural networks with millions to billions of parameters that are trained via self-supervised or semi-supervised learning using vast unlabeled text.</s>
"""
~~~~
---
# Tokenization
The model uses the original llama-2 tokenizer.
---
# How to cite
~~~
@misc{rinna-youri-7b-instruction,
title = {rinna/youri-7b-instruction},
author={Zhao, Tianyu and Sawada, Kei}
url = {https://huggingface.co/rinna/youri-7b-instruction},
}
@inproceedings{sawada2024release,
title = {Release of Pre-Trained Models for the {J}apanese Language},
author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
month = {5},
year = {2024},
url = {https://arxiv.org/abs/2404.01657},
}
~~~
---
# License
[The llama2 license](https://ai.meta.com/llama/license/)
|