Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

部署服务 #3648

Closed
1 task done
cxjtju opened this issue May 9, 2024 · 1 comment
Closed
1 task done

部署服务 #3648

cxjtju opened this issue May 9, 2024 · 1 comment
Labels
solved This problem has been already solved.

Comments

@cxjtju
Copy link

cxjtju commented May 9, 2024

Reminder

  • I have read the README and searched the existing issues.

Reproduction

CUDA_VISIBLE_DEVICES=0,1 API_PORT=8000 llamafactory-cli api examples/inference/llama3_vllm.yaml

Expected behavior

CUDA_VISIBLE_DEVICES=0,1 API_PORT=8000 llamafactory-cli api examples/inference/llama3_vllm.yaml部署服务时报错ImportError: /usr/local/lib/python3.9/dist-packages/flash_attn_2_cuda.cpython-39-x86_64-linux-gnu.so: undefined symbol: _ZNK3c106SymIntltEl,请问设置什么参数可以不使用flash_attn呢?

System Info

No response

Others

No response

@hiyouga
Copy link
Owner

hiyouga commented May 9, 2024

pip uninstall flash_attn

@hiyouga hiyouga added the solved This problem has been already solved. label May 9, 2024
@hiyouga hiyouga closed this as completed May 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
solved This problem has been already solved.
Projects
None yet
Development

No branches or pull requests

2 participants