Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Rockchip RK3588 / NPU / Mali-G610 GPU #2205

Open
adamcohenhillel opened this issue May 1, 2024 · 3 comments
Open

Support Rockchip RK3588 / NPU / Mali-G610 GPU #2205

adamcohenhillel opened this issue May 1, 2024 · 3 comments
Labels
enhancement New feature or request

Comments

@adamcohenhillel
Copy link

adamcohenhillel commented May 1, 2024

Is your feature request related to a problem? Please describe.
I am trying to run LocalAI more efficiently on a Orange Pi 5B, which have an NPUm and Arm Mali-G610 GPU.

More specifically:
CPU | • 8-core 64-bit processor• Big.Little Architecture: 4 * Cortex-A76 and 4 * Cortex-A55, Big core cluster is 2.4GHz, and Little core cluster is 1.8GHz frequency.
GPU | • Arm Mali-G610 MP4• Compatible with OpenGL ES1.1/2.0/3.2, OpenCL 2.2 and Vulkan 1.2• 3D graphics engine and 2D graphics engine
NPU | Built-in AI accelerator NPU with up to 6 TOPS, supports INT4/INT8/INT16 mixed operation
(reference )
There are different frameworks that supports taking advantage of the NPU / GPU on that board to accelerate models inference, and I was wondering if it is possible to integrate that into LocalAI as well?

There is also an interesting discussion on this topic on the Ollama repo: ggerganov/llama.cpp#722

Rockchip are pretty cheap, and can be a great for edge AI stuff. Would be great to have support for that !!

@adamcohenhillel adamcohenhillel added the enhancement New feature or request label May 1, 2024
@localai-bot
Copy link
Contributor

Hello,

Thank you for opening this issue, and thank you for providing the detailed specifications of the Orange Pi 5B that you are trying to optimize. It's great to hear that Rockchip devices like this offer promising capabilities for edge AI applications.

I will mention this feature request to the maintained of the LocalAI project, @mudler, and the community will evaluate the possibilities of incorporating support for Rockchip RK3588, NPU, and the Mali-G610 GPU into the project.

Please feel free to add any further information or clarification if needed. We will keep you updated on any progress regarding this request.

This conversation is being managed by a bot, an experiment of @mudler.

@fakezeta
Copy link
Collaborator

fakezeta commented May 2, 2024

Hi @adamcohenhillel,
thank you for your interest in LocalAI.

The discussion is on llama.cpp repo, not Ollama and LocalAI already uses llama.cpp as one of the backends so as soon as it will be supported from the upstream repo it will be supported also here on LocalAI.

Let's monitor the updates from the discussion you linked!

@nekkitl
Copy link

nekkitl commented May 13, 2024

+1

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

4 participants