LibreChat just outputs nonsense text using LiteLLM and Ollama #1960
Replies: 4 comments 6 replies
-
You can try adjusting the temperature according to what's recommended for the model you are using. That might be affecting the response. addParams:
- temperature: 0.5 https://docs.librechat.ai/install/configuration/custom_config.html#addparams |
Beta Was this translation helpful? Give feedback.
-
I have Ollama BEHIND LiteLLM. So I connect from LibreChat to LiteLLM, which then connects to Ollama. When Using other UIs, everything works fine. The latest Answer I got was <|im_start|>assttassistantistantie he heißtassistantuser?t The "im_start" seems to be some sort of "start of message" signal of the API. But it seems to not be interpreted right. |
Beta Was this translation helpful? Give feedback.
-
I found the problem (for me). While reading the LibreChat/logs/debug*.log I discovered that the default end point was gptPlugins instead of ollama/litellm. |
Beta Was this translation helpful? Give feedback.
-
sorry, direct Ollama is working, but I still face issues with litellm :( I opened another issue |
Beta Was this translation helpful? Give feedback.
-
What happened?
When even asking a simple question, the answer is just nonsense:
That is my librechat.yaml:
Steps to Reproduce
What browsers are you seeing the problem on?
Firefox, Mobile (Android)
Relevant log output
No response
Screenshots
No response
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions