New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What's most optimal context size for the average workflow to work well ? #388
Comments
Umm what i feel is it at first didn't work because of the wrong response from the LLM on the prompts which are there. Because it shows "Invalid response from the model, trying again". This comes when the output from the LLM is a bit wrong and doesn't get validated due to its wrong format (as per the prompt). |
I think it's because gemma wasn't good for task, however the result using mistral was a little promising given that it's a 7b model, I tried codellama 34B but it was too much for my computer, I created a pull request for the LM Studio option, maybe someone can try it out and play with the context size : #389 |
I tried to add LM_STUDIO internal server as a model to be in the options, and I only tried it with 2000 context using the google gemma 7b model, I didn't get any results, even upping the number of tokens to 8192 which is the maximum supported by the model, it took so much time and didn't reach the goal, I asked it to create the game of life using python and matplotlib :
once it surpasses the number of tokens it keeps looping to the void, I want to know if there's a way that we can make the most out of the 8192 context size, if not what is the optimal the one should have ?
The text was updated successfully, but these errors were encountered: