export DATA=h2oai/openassistant_oasst1_h2ogpt export BASE_MODEL=tiiuae/falcon-7b # confirmed working with 0.9.2 # export BASE_MODEL=openlm-research/open_llama_3b # fails with OOM on 48GB card?? # export BASE_MODEL=Salesforce/xgen-7b-8k-base # fails since tokenizer not yet supported (have to hack to force LLaMa tokenizer) export CUDA_VISIBLE_DEVICES=0 export MODEL=model-test export MODEL_NAME=`echo $MODEL | sed 's@/@_@g'` export HF_PORT=1000 #export TGI_VERSION=latest # works #export TGI_VERSION=0.9.1 # fails export TGI_VERSION=0.9.3 # works # Train LoRA rm -rf $MODEL.lora python finetune.py --data_path=$DATA --base_model=$BASE_MODEL --num_epochs=0.01 --output_dir=$MODEL.lora # Merge LoRA, export model to $MODEL dir (via env var) rm -rf $MODEL python src/export_hf_checkpoint.py # Load model with TGI docker run --gpus all --shm-size 1g -e CUDA_VISIBLE_DEVICES=$CUDA_VISIBLE_DEVICES -p $HF_PORT:80 -v $HOME/.cache/huggingface/hub/:/data -v $PWD/$MODEL:/$MODEL ghcr.io/huggingface/text-generation-inference:$TGI_VERSION --model-id /$MODEL --max-input-length 2048 --max-total-tokens 4096 --max-stop-sequences 6 --sharded false --disable-custom-kernels --trust-remote-code