New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Issue] embeddings model needs to be provided when using Azure API #296
Comments
Zep embeds all chat conversations so that they may be recalled via semantic search. Are you using Azure's OpenAI embedding service or Zep's local embedding service? More about Zep's embedding configuration here: https://docs.getzep.com/deployment/embeddings/ |
I want to use long-term memory with summarization. Because my message window is about 5 messages since this a real-time rag I want to summarize the last few messages to handle the flow. So I don't need to use embeddings. However I want to use Azure Open AI models for summarization. It won't let me disable the embedding when using Azure Open AI endpoint and LLM, the following message is coming : |
I'm getting following message when deployed zep image using docker-compose and failed the zep container creation.
embeddings model needs to be provided when using Azure API
Do we have to use an embedding model when using Azure? Can I just use Azure Open AI for the summarizing? I have tried using Azure Embedding Model as well. When using the embedding model my bot responses get very slow.
The text was updated successfully, but these errors were encountered: