Update README.md
Browse files
README.md
CHANGED
@@ -27,12 +27,14 @@ This repository provides large language models developed by [LLM-jp](https://llm
|
|
27 |
|
28 |
| Model Variant |
|
29 |
| :--- |
|
30 |
-
|**Instruction models**|
|
|
|
31 |
| [llm-jp-13b-instruct-full-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1) |
|
|
|
|
|
32 |
| [llm-jp-13b-instruct-full-jaster-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-jaster-v1.0) |
|
33 |
| [llm-jp-13b-instruct-full-jaster-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-jaster-dolly-oasst-v1.0) |
|
34 |
| [llm-jp-13b-instruct-full-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-dolly-oasst-v1.0) |
|
35 |
-
| [llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1) |
|
36 |
| [llm-jp-13b-instruct-lora-jaster-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-jaster-v1.0) |
|
37 |
| [llm-jp-13b-instruct-lora-jaster-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-jaster-dolly-oasst-v1.0) |
|
38 |
| [llm-jp-13b-instruct-lora-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-dolly-oasst-v1.0) |
|
@@ -58,18 +60,18 @@ Checkpoints format: Hugging Face Transformers (Megatron-DeepSpeed format models
|
|
58 |
```python
|
59 |
import torch
|
60 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
61 |
-
tokenizer = AutoTokenizer.from_pretrained("llm-jp/llm-jp-13b-instruct-
|
62 |
-
model = AutoModelForCausalLM.from_pretrained("llm-jp/llm-jp-13b-instruct-
|
63 |
-
text = "θͺηΆθ¨θͺε¦ηγ¨γ―δ½γ"
|
64 |
-
text = text + "### εηοΌ"
|
65 |
tokenized_input = tokenizer.encode(text, add_special_tokens=False, return_tensors="pt").to(model.device)
|
66 |
with torch.no_grad():
|
67 |
output = model.generate(
|
68 |
tokenized_input,
|
69 |
-
max_new_tokens=
|
70 |
do_sample=True,
|
71 |
top_p=0.95,
|
72 |
temperature=0.7,
|
|
|
73 |
)[0]
|
74 |
print(tokenizer.decode(output))
|
75 |
```
|
@@ -130,11 +132,12 @@ The models have been fine-tuned on the following datasets.
|
|
130 |
| Language | Dataset | description |
|
131 |
|:---|:---:|:---:|
|
132 |
|Japanese|[jaster](https://github.com/llm-jp/llm-jp-eval)| An automatically transformed data from the existing Japanese NLP datasets |
|
133 |
-
|English|[databricks-dolly-15k
|
134 |
|Japanese|[databricks-dolly-15k-ja](https://huggingface.co/datasets/llm-jp/databricks-dolly-15k-ja)| A translated one by DeepL in LLM-jp |
|
135 |
|English|[oasst1-21k-en](https://huggingface.co/datasets/llm-jp/oasst1-21k-en)| English subset of [oasst1 dataset](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
136 |
|Japanese|[oasst1-21k-ja](https://huggingface.co/datasets/llm-jp/oasst1-21k-ja)| A translated one by DeepL in LLM-jp |
|
137 |
|Japanese|[ichikara_003_001](https://liat-aip.sakura.ne.jp/wp/llm%E3%81%AE%E3%81%9F%E3%82%81%E3%81%AE%E6%97%A5%E6%9C%AC%E8%AA%9E%E3%82%A4%E3%83%B3%E3%82%B9%E3%83%88%E3%83%A9%E3%82%AF%E3%82%B7%E3%83%A7%E3%83%B3%E3%83%87%E3%83%BC%E3%82%BF%E4%BD%9C%E6%88%90/)| ichikara-instruction dataset (ver.003-001)
|
|
|
138 |
|
139 |
|
140 |
## Evaluation
|
|
|
27 |
|
28 |
| Model Variant |
|
29 |
| :--- |
|
30 |
+
|**Instruction models ver1.1**|
|
31 |
+
| [llm-jp-13b-dpo-lora-hh_rlhf_ja-v1.1](https://huggingface.co/llm-jp/llm-jp-13b-dpo-lora-hh_rlhf_ja-v1.1)|
|
32 |
| [llm-jp-13b-instruct-full-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1) |
|
33 |
+
| [llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1) |
|
34 |
+
|**Instruction models ver1.0**|
|
35 |
| [llm-jp-13b-instruct-full-jaster-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-jaster-v1.0) |
|
36 |
| [llm-jp-13b-instruct-full-jaster-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-jaster-dolly-oasst-v1.0) |
|
37 |
| [llm-jp-13b-instruct-full-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-full-dolly-oasst-v1.0) |
|
|
|
38 |
| [llm-jp-13b-instruct-lora-jaster-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-jaster-v1.0) |
|
39 |
| [llm-jp-13b-instruct-lora-jaster-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-jaster-dolly-oasst-v1.0) |
|
40 |
| [llm-jp-13b-instruct-lora-dolly-oasst-v1.0](https://huggingface.co/llm-jp/llm-jp-13b-instruct-lora-dolly-oasst-v1.0) |
|
|
|
60 |
```python
|
61 |
import torch
|
62 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
63 |
+
tokenizer = AutoTokenizer.from_pretrained("llm-jp/llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1")
|
64 |
+
model = AutoModelForCausalLM.from_pretrained("llm-jp/llm-jp-13b-instruct-lora-dolly_en-dolly_ja-ichikara_003_001-oasst_en-oasst_ja-v1.1", device_map="auto", torch_dtype=torch.float16)
|
65 |
+
text = "δ»₯δΈγ―γγΏγΉγ―γθͺ¬ζγγζη€Ίγ§γγθ¦ζ±γι©εγ«ζΊγγεΏηγζΈγγͺγγγ\n\n### ζη€Ί:\n{instruction}\n\n### εΏη:\n".format(instruction="θͺηΆθ¨θͺε¦ηγ¨γ―δ½γ")
|
|
|
66 |
tokenized_input = tokenizer.encode(text, add_special_tokens=False, return_tensors="pt").to(model.device)
|
67 |
with torch.no_grad():
|
68 |
output = model.generate(
|
69 |
tokenized_input,
|
70 |
+
max_new_tokens=512,
|
71 |
do_sample=True,
|
72 |
top_p=0.95,
|
73 |
temperature=0.7,
|
74 |
+
repetition_penalty=1.1,
|
75 |
)[0]
|
76 |
print(tokenizer.decode(output))
|
77 |
```
|
|
|
132 |
| Language | Dataset | description |
|
133 |
|:---|:---:|:---:|
|
134 |
|Japanese|[jaster](https://github.com/llm-jp/llm-jp-eval)| An automatically transformed data from the existing Japanese NLP datasets |
|
135 |
+
|English|[databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k)| - |
|
136 |
|Japanese|[databricks-dolly-15k-ja](https://huggingface.co/datasets/llm-jp/databricks-dolly-15k-ja)| A translated one by DeepL in LLM-jp |
|
137 |
|English|[oasst1-21k-en](https://huggingface.co/datasets/llm-jp/oasst1-21k-en)| English subset of [oasst1 dataset](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
138 |
|Japanese|[oasst1-21k-ja](https://huggingface.co/datasets/llm-jp/oasst1-21k-ja)| A translated one by DeepL in LLM-jp |
|
139 |
|Japanese|[ichikara_003_001](https://liat-aip.sakura.ne.jp/wp/llm%E3%81%AE%E3%81%9F%E3%82%81%E3%81%AE%E6%97%A5%E6%9C%AC%E8%AA%9E%E3%82%A4%E3%83%B3%E3%82%B9%E3%83%88%E3%83%A9%E3%82%AF%E3%82%B7%E3%83%A7%E3%83%B3%E3%83%87%E3%83%BC%E3%82%BF%E4%BD%9C%E6%88%90/)| ichikara-instruction dataset (ver.003-001)
|
140 |
+
|Japanese|[hh-rlhf-12k-ja](https://huggingface.co/datasets/llm-jp/hh-rlhf-12k-ja)| A translated one by DeepL in LLM-jp |
|
141 |
|
142 |
|
143 |
## Evaluation
|