Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support LMDeploy backend #966

Open
zhyncs opened this issue Apr 25, 2024 · 1 comment
Open

feat: support LMDeploy backend #966

zhyncs opened this issue Apr 25, 2024 · 1 comment

Comments

@zhyncs
Copy link

zhyncs commented Apr 25, 2024

Feature request

@aarnphm @ssheng @parano Hi OpenLLM team, thank you for your exceptional work. Currently, OpenLLM supports two backends, vLLM and PyTorch, with good usability but there is still room for improvement in terms of performance. LMDeploy has achieved a good balance between performance and usability, with recent Llama3 8B showing a 1.8x performance improvement over vLLM on LMDeploy. Performance is crucial, especially when the demand for large-scale deployment arises after meeting user requirements. Currently, Meituan is widely using internally. I strongly recommend OpenLLM to consider integrating LMDeploy and making it the default backend. You can refer to the documentation at https://lmdeploy.readthedocs.io/en/latest/ during the research and integration process. Thanks.

Motivation

No response

Other

No response

@zhyncs
Copy link
Author

zhyncs commented Apr 25, 2024

cc @lvhan028 @AllentDan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant