Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ggml vs Qualcomm SNPE inference engine on qualcomm soc #809

Open
Francis235 opened this issue Apr 30, 2024 · 0 comments
Open

ggml vs Qualcomm SNPE inference engine on qualcomm soc #809

Francis235 opened this issue Apr 30, 2024 · 0 comments

Comments

@Francis235
Copy link

Hello, I plan to deploy the model using ggml on Qualcomm's chip. I'm curious about the comparison between using ggml for inference on an SoC chip (such as a Qualcomm SoC, involving components like CPU, GPU, NPU, etc.) versus leveraging the inference engine provided by the chip itself (such as qualcomm SNPE). Since ggml inference primarily takes place on the CPU, whereas the chip's inference engine can offload computations to the GPU or NPU, does using ggml lead to a significant increase in CPU memory usage and %CPU, potentially impacting other tasks? Has anyone conducted a similar comparative test?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant