blockblockblock's picture
Upload folder using huggingface_hub
818542f verified
---
license: apache-2.0
language:
- ja
- en
tags:
- japanese
- causal-lm
inference: false
---
# CyberAgentLM3-22B-Chat (CALM3-22B-Chat)
## Model Description
CyberAgentLM3 is a decoder-only language model pre-trained on 2.0 trillion tokens from scratch.
CyberAgentLM3-Chat is a fine-tuned model specialized for dialogue use cases.
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
model = AutoModelForCausalLM.from_pretrained("cyberagent/calm3-22b-chat", device_map="auto", torch_dtype="auto")
tokenizer = AutoTokenizer.from_pretrained("cyberagent/calm3-22b-chat")
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
messages = [
{"role": "system", "content": "あなたは親切なAIアシスタントです。"},
{"role": "user", "content": "AIによって私たちの暮らしはどのように変わりますか?"}
]
input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
output_ids = model.generate(input_ids,
max_new_tokens=1024,
temperature=0.5,
streamer=streamer)
```
## Prompt Format
CALM3-Chat uses ChatML as the prompt format.
```
<|im_start|>system
あなたは親切なAIアシスタントです。<|im_end|>
<|im_start|>user
AIによって私たちの暮らしはどのように変わりますか?<|im_end|>
<|im_start|>assistant
```
## Model Details
* **Model size**: 22B
* **Context length**: 16384
* **Model type**: Transformer-based Language Model
* **Language(s)**: Japanese, English
* **Developed by**: [CyberAgent, Inc.](https://www.cyberagent.co.jp/)
* **License**: Apache-2.0
## Author
[Ryosuke Ishigami](https://huggingface.co/rishigami)
## How to cite
```tex
@misc{cyberagent-calm3-22b-chat,
title={cyberagent/calm3-22b-chat},
url={https://huggingface.co/cyberagent/calm3-22b-chat},
author={Ryosuke Ishigami},
year={2024},
}
```