humanda5
gemini llm model
c246b94
raw
history blame contribute delete
745 Bytes
# portfolio/npc_social_network/models/openkollm_setup.py
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
from dotenv import load_dotenv
import os
def load_openkollm():
load_dotenv() # .env ํŒŒ์ผ์—์„œ ํ™˜๊ฒฝ ๋ณ€์ˆ˜ ๋กœ๋“œ
model_id = "beomi/KoAlpaca-Polyglot-12.8B" # GPU๊ฐ€ ์—†์œผ๋ฉด ๋ป—์Œ
access_token = os.getenv("HF_token")
tokenizer = AutoTokenizer.from_pretrained(model_id, token=access_token)
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto", # GPU ์ž๋™ ํ• ๋‹น
# load_in_8bit=True, # GPU ์‚ฌ์šฉํ•ด์„œ 8bit๋กœ ๋กœ๋“œ
torch_dtype="auto"
)
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
return pipe