

eval "$(conda shell.bash hook)"
conda create --prefix=/data/nfs-ten1/nfs/meichaoyang001/envs/long_bench python=3.10 -y
conda activate /data/nfs-ten1/nfs/meichaoyang001/envs/long_bench


#pip install torch==2.1.2 torchvision==0.16.2 torchaudio==2.1.2 --index-url https://download.pytorch.org/whl/cu118
cd /data/nfs-ten1/nfs/meichaoyang001/pypi/
pip install torch-2.1.2+cu118-cp310-cp310-linux_x86_64.whl torchaudio-2.1.2+cu118-cp310-cp310-linux_x86_64.whl torchvision-0.16.2+cu118-cp310-cp310-linux_x86_64.whl
pip install flash_attn-2.3.6+cu118torch2.1cxx11abiFALSE-cp310-cp310-linux_x86_64.whl # /data/nfs-ten1/nfs/meichaoyang001/project/flash-attention_0119/
pip install cupy-cuda11x
#pip install vllm-0.3.1+cu118-cp310-cp310-manylinux1_x86_64.whl
#pip uninstall torch -y && pip install torch-2.1.2+cu118-cp310-cp310-linux_x86_64.whl
pip uninstall xformers -y && pip install xformers-0.0.23.post1+cu118-cp310-cp310-manylinux2014_x86_64.whl
cd -

pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple

# cd ../ && git clone https://gitee.com/meichaoyang/flash-attention.git
# cd flash-attention && pip install flash-attn --no-build-isolation && pip install csrc/layer_norm && pip install csrc/rotary

# git clone https://modelscope.cn/qwen/Qwen-14B-Chat.git Qwen-14B-Chat_1221
# git clone https://modelscope.cn/qwen/Qwen-7B-Chat.git Qwen-7B-Chat_1221
