Antler 7B
![drawing](/Elizezen/Antler-7B-GGUF/resolve/main/OIG3.UAjshTXCEJU.jpg)
Model Description
GGUF version of Antler-7B.
This is a 7B-parameter decoder-only Japanese language model fine-tuned on novel datasets, built on top of the base model Japanese Stable LM Base Gamma 7B. Japanese Stable LM Instruct Gamma 7B
Usage
Ensure you are using Transformers 4.34.0 or newer.
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Elizezen/Antler-7B")
model = AutoModelForCausalLM.from_pretrained(
"Elizezen/Antler-7B",
torch_dtype="auto",
)
model.eval()
if torch.cuda.is_available():
model = model.to("cuda")
input_ids = tokenizer.encode(
"吾輩は猫である。名前はまだない",,
add_special_tokens=True,
return_tensors="pt"
)
tokens = model.generate(
input_ids.to(device=model.device),
max_new_tokens=512,
temperature=1,
top_p=0.95,
do_sample=True,
)
out = tokenizer.decode(tokens[0][input_ids.shape[1]:], skip_special_tokens=True).strip()
print(out)
Datasets
- less than 1GB of web novels(non-PG)
- 70GB of web novels(PG)
Intended Use
The model is mainly intended to be used for generating novels. It may not be so capable with instruction-based responses.
- Downloads last month
- 19
Unable to determine this model's library. Check the
docs
.