New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mode tts, gpt4 or vision frozes from time to time without timeout #2258
Comments
Thank you for reporting this issue, DavidGOrtega. We will investigate and try to reproduce the problem in a controlled environment. In the meantime, if you have any other details or find anything that could help, please let us know. We'll get back to you as soon as we have more information or a resolution to the issue. This is an ongoing experiment by @mudler, and we're here to help improve LocalAI. |
This happens when the prompt exceeds the context size and there is no more space for the response - looks like something we could handle on our side and fail cleanly instead. What's your context size window? Can you share your model config/ setup? |
@mudler Im no even using that model as I use mine, and nothing is apparently requesting it. The only thing I did with that model was install it and then delete it after try it. Is that model |
An easy way to hang the system is to make several requests to tts endpoint in a row (in my case no more than three) to generate the speech of a larger text. it hangs and never timeouts. Tested with piper and bark |
I confirm this issue. Use LocalAI (v2.14.0) with Orca2. Here are logs:
|
LocalAI version:
2.14.0
Environment, CPU architecture, OS, and Version:
Linux Ubuntu SMP PREEMPT_DYNAMIC x86_64 x86_64 x86_64 GNU/Linux
90GB RAM 22 vcores
nvidia L4 24GB
Describe the bug
Requests frozen from time to time
My logs are continuously producing:
The text was updated successfully, but these errors were encountered: