-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to fix install cuda12.1, python=3.9, flash-atten=2.3.2 #598
Comments
Can you check if you can download that wheel manually (e.g. with wget)? |
Do you have write permission to /home/dodx/tmp? |
yes , I has just fix , I can run but I have a question Warning: import flash_attn rms_norm fail, please install FlashAttention layer_norm to get higher efficiency https://github.com/Dao-AILab/flash-attention/tree/main/csrc/layer_norm so I use rtx 3090 don't use layer_norm right |
You can try the layer_norm, I think it should work but I haven't tested extensively on 3080. |
thanks u so much for reply me :)) |
@batman-do see this
|
I got error this /data/dodx/anaconda3/envs/flash_attention/lib/python3.10/site-packages/torch/include/ATen/cuda/CUDAContext.h(76): error: identifier "cusparseHandle_t" is undefined 2 errors detected in the compilation of "/data/dodx/GenerateAI/test_LLM_local/flash-attention/csrc/flash_attn/src/flash_bwd_hdim128_bf16_sm80.cu". /data/dodx/anaconda3/envs/flash_attention/lib/python3.10/site-packages/torch/include/ATen/cuda/CUDAContext.h(76): error: identifier "cusparseHandle_t" is undefined 2 errors detected in the compilation of "/data/dodx/GenerateAI/test_LLM_local/flash-attention/csrc/flash_attn/src/flash_bwd_hdim128_fp16_sm80.cu". /data/dodx/anaconda3/envs/flash_attention/lib/python3.10/site-packages/torch/include/ATen/cuda/CUDAContext.h(76): error: identifier "cusparseHandle_t" is undefined 2 errors detected in the compilation of "/data/dodx/GenerateAI/test_LLM_local/flash-attention/csrc/flash_attn/src/flash_bwd_hdim160_bf16_sm80.cu". The above exception was the direct cause of the following exception: Traceback (most recent call last): |
i use MAX_JOBS=4 pip install flash-attn --no-build-isolation alterative build from source |
Hi @tridao , why I install layer_norm don't it respond, stopped like this |
It probably takes a very long time if you don't have ninja or lots of CPU cores to compile. You don't have to use that extension. |
Thanks @tridao , I will try maybe later |
@batman-do Hi, I got the exact same bug when trying |
@batman-do It seems that CUDA library cusparseHandle_t not locate properly. I am using venv virtual environment not anaconda. An do make sure And also, layerNorm is deprecate in flash attention2, no need to install. |
use shutil.move(wheel_filename, wheel_path) instead os.rename(src, dst) in setup.py |
Thanks! It works for me. |
It keeps affecting users. Shouldn't have been closed. I guess it happens when /tmp/ or other pip cache directories are on different filesystems or tmpfs. |
can u suggest to me solution for fix error that?, thank you
The text was updated successfully, but these errors were encountered: