Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[LLm serving] Fix timeout setting bug #2398

Closed
wants to merge 75 commits into from

Conversation

rainyfly
Copy link
Collaborator

@rainyfly rainyfly commented Mar 6, 2024

PR types(PR类型)

Other

Description

  1. Bug fix

root and others added 30 commits September 25, 2023 10:37
* Support chatglm-6b

* Update README.md

* support dynamic batching

* support dynamic batching

* fix dybatch

* Disable dynamic batching for chatglm

---------

Co-authored-by: root <[email protected]>
* add prefix cache for chatglm

* support chatglm
* Support multicards

* fix ptuning diff

* Update engine.py
* support bloom prefix

* support_bloom_prefix

* support bloom prefix
* support bloom prefix

* support_bloom_prefix

* support bloom prefix

* Update code for bloom prefix

* update code
* support bloom prefix

* support_bloom_prefix

* support bloom prefix

* Update code for bloom prefix

* update code

* support bloom prefix
* [LLM] Support dynamic batching for chatglm

* fix bug in triton model

* fix bug

* fix bug
karagg and others added 27 commits November 14, 2023 11:06
* test

* test FastDeploy

* test

* delete run.sh

* delete run.sh

* update run.sh

* update run.sh ci.py

* update ci.py

* update ci.py

---------

Co-authored-by: root <[email protected]>
* add inference load balancer for fastdeploy llm

* add inference load balance controller for llm

* add ic for llm

* add ic for llm

* add fastdeploy ic for llm

* add fastdeploy ic to llm

* Fix asyncio.CancelError exception

* Improve robust for llm service

* Improve robust for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service
* add inference load balancer for fastdeploy llm

* add inference load balance controller for llm

* add ic for llm

* add ic for llm

* add fastdeploy ic for llm

* add fastdeploy ic to llm

* Fix asyncio.CancelError exception

* Improve robust for llm service

* Improve robust for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* add detailed log

* add detailed log
* add inference load balancer for fastdeploy llm

* add inference load balance controller for llm

* add ic for llm

* add ic for llm

* add fastdeploy ic for llm

* add fastdeploy ic to llm

* Fix asyncio.CancelError exception

* Improve robust for llm service

* Improve robust for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* Add detailed log for llm service

* add detailed log

* add detailed log

* add detailed log
* Add warning for server hangs

* Add http nonstream server support

* bump fastdeploy_llm to v1.0.0

* add time log for each request

* baidu-fastdeploy-fastdeploy-3 fix time format
* Add fastapi support

* Add fastapi support

* fix bug
* Add fastapi support

* Add fastapi support

* fix bug

* Update Code for fastapi
Copy link

paddle-bot bot commented Mar 6, 2024

Thanks for your contribution!

@CLAassistant
Copy link

CLAassistant commented Mar 6, 2024

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.
4 out of 5 committers have signed the CLA.

✅ rainyfly
✅ jiangjiajun
✅ Zeref996
✅ karagg
❌ root


root seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.
You have signed the CLA already but the status is still pending? Let us recheck it.

@rainyfly rainyfly closed this Mar 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

5 participants