YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Quantization made by Richard Erkhov.
mptk-1b - bnb 4bits
- Model creator: https://huggingface.co/team-lucid/
- Original model: https://huggingface.co/team-lucid/mptk-1b/
Original model description:
license: apache-2.0 language: - ko
MPTK-1B
MPTK-1B๋ ํ๊ตญ์ด/์์ด์ฝ๋ ๋ฐ์ดํฐ์ ์์ ํ์ต๋ 1.3B ํ๋ผ๋ฏธํฐ์ decoder-only transformer ์ธ์ด๋ชจ๋ธ์ ๋๋ค.
์ด ๋ชจ๋ธ์ ๊ตฌ๊ธ์ TPU Research Cloud(TRC)๋ฅผ ํตํด ์ง์๋ฐ์ Cloud TPU๋ก ํ์ต๋์์ต๋๋ค.
Model Details
Model Description
๋ค๋ฅธ decoder-only transformer์์ ์ผ๋ถ ์์ ๋ ์ํคํ ์ฒ์ธ MPT๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ํฉ๋๋ค.
- ALiBi (Attention with Linear Biases)๋ฅผ ์ฌ์ฉํฉ๋๋ค
- bias๋ฅผ ์ฌ์ฉํ์ง ์์ต๋๋ค.
Hyperparameter | Value |
---|---|
n_parameters | 1.3B |
n_layers | 24 |
n_heads | 16 |
d_model | 2048 |
vocab size | 50432 |
sequence length | 2048 |
Uses
How to Get Started with the Model
fp16์ผ๋ก ์คํ ์ NaN์ด ๋ฐ์ํ ์ ์์ต๋๋ค. ๋ฐ๋ผ์ fp32 ํน์ bf16๋ก ์คํํ๊ธฐ๋ฅผ ๊ถ์ฅํฉ๋๋ค.
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
tokenizer = AutoTokenizer.from_pretrained("team-lucid/mptk-1b")
model = AutoModelForCausalLM.from_pretrained("team-lucid/mptk-1b")
pipe = pipeline('text-generation', model=model, tokenizer=tokenizer, device='cuda:0')
with torch.autocast('cuda', dtype=torch.bfloat16):
print(
pipe(
'๋ํ๋ฏผ๊ตญ์ ์๋๋',
max_new_tokens=100,
do_sample=True,
)
)
Training Details
Training Data
OSCAR, mC4, wikipedia, namuwiki ๋ฑ ํ๊ตญ์ด ๋ฐ์ดํฐ์ RefinedWeb, The Stack ์์ ์ผ๋ถ๋ฅผ ์ถ๊ฐํด ํ์ตํ์์ต๋๋ค.
Training Hyperparameters
Hyperparameter | Value |
---|---|
Precision | bfloat16 |
Optimizer | Lion |
Learning rate | 2e-4 |
Batch size | 1024 |
- Downloads last month
- 6
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.