Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: GPTCache implementation for GPT Assistants #573

Open
yashaswi-nayak opened this issue Nov 22, 2023 · 0 comments
Open

[Feature]: GPTCache implementation for GPT Assistants #573

yashaswi-nayak opened this issue Nov 22, 2023 · 0 comments

Comments

@yashaswi-nayak
Copy link

Is your feature request related to a problem? Please describe.

We are creating multiple GPT Assistants to support our customer service operations. Currently, there is only support for caching results via ChatCompletion. We would like to see a GPT Assistant implementation of GPTCache.

GPT Assistants will mainly answer multiple queries that are very common like

  • Greeting a user/customer
  • Generic FAQs - questions that have been answered previously (stored in cache)
  • Out-of-Scope Answers that are stored in the cache ( using RAG, we limit the answers to the ones stored in the cache)

Over the course of time, we build a decent cache storage of all answers for user queries. These answers can be pulled from the cache directly instead of requesting Open AI every time. This would save us a lot of tokens and latency time.

Describe the solution you'd like.

I would like to see a solution where we have

  • GPT Assistant adapter class in Open AI adapter
  • Functionality to select assistant via assistant_id
  • Functionality to create and update cache for a specific assistant

Describe an alternate solution.

No response

Anything else? (Additional Context)

No response

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant