error when trying inference locally
#8
by
hussainwali1
- opened
RuntimeError: Please install flash-attn==1.0.3.post0 and triton==2.0.0.dev20221202
If you would like to use the triton implementation, you will need to do as the message says and install flash attention and triton. Otherwise you can stick with the torch implementation without issue.
daking
changed discussion status to
closed