import os | |
#os.system('git clone https://github.com/tloen/alpaca-lora.git') | |
import streamlit as st | |
from streamlit_chat import message as st_message | |
import bitsandbytes as bnb | |
from transformers import GenerationConfig | |
from transformers import AutoTokenizer, AutoConfig, LlamaForCausalLM, LlamaTokenizer | |
tokenizer = LlamaTokenizer.from_pretrained("wxjiao/alpaca-7b") | |
model2 = LlamaForCausalLM.from_pretrained( | |
"wxjiao/alpaca-7b", | |
load_in_8bit_fp32_cpu_offload=True, | |
from_pretrained="auto", | |
) | |