Model does not reply ( Is typing.. ) / MetaIX/GPT4-X-Alpaca-30B-4bit
#14
by
ilnurshams
- opened
The same problem here, "'LlamaForCausalLM' object has no attribute 'generate_with_streaming'" in the console. Linux, RTX 4090, 24 VRAM
Okay, I fixed it. I don't know what was the reason.
- Fresh install of oobabooga one-click installers
- Start start_windows file. Don't dowload any model!
- I manually downloaded the necessary model files from https://huggingface.co/MetaIX/GPT4-X-Alpaca-30B-4bit/tree/main
( see picture in attachment ) and put in oobabooga models folder. - I installed pytorch and cuda via Conda ( download and install Conda first, then run Anaconda Prompt ( miniconda3 ) as admin ). There I run these code ( in order to install pytorch and cuda ). You also need to have python 3.10 as I had it already ( download it )!
conda create --name gptq python=3.10 -y
conda activate gptq
conda install pytorch torchvision torchaudio pytorch-cuda=11.7 -c pytorch -c nvidia
( I used this tutorial, but it is for linux. https://github.com/qwopqwop200/GPTQ-for-LLaMa
I used only code from tutorial to install pytorch and cuda )
- Then I run update_windows file in oobabooga main folder.
Done!
ilnurshams
changed discussion status to
closed