Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What's most optimal context size for the average workflow to work well ? #388

Open
ayoubachak opened this issue Apr 7, 2024 · 4 comments

Comments

@ayoubachak
Copy link

I tried to add LM_STUDIO internal server as a model to be in the options, and I only tried it with 2000 context using the google gemma 7b model, I didn't get any results, even upping the number of tokens to 8192 which is the maximum supported by the model, it took so much time and didn't reach the goal, I asked it to create the game of life using python and matplotlib :
image

once it surpasses the number of tokens it keeps looping to the void, I want to know if there's a way that we can make the most out of the 8192 context size, if not what is the optimal the one should have ?

image

@ayoubachak
Copy link
Author

It suddenly changed its mind and started looking arround for SQL content which is completly non-related with the topic, It must be because of gemma's non ability to perform coding tasks due to its small size and it not being optimized for coding :
image

@ayoubachak
Copy link
Author

UPDATE : I've managed to get one of the models to work, I've used mistral instruc 7B, and asked the same question, but it gave me the results in rust instead of python, using python libraries that I don't think are yet available in rust like matplotlib and numpy.
image

The context size I've used was 32768 tokens, which I don't think was necessary to set as the whole process took 5522 tokens as shown in the image above .
image

@obliviousz
Copy link
Contributor

Umm what i feel is it at first didn't work because of the wrong response from the LLM on the prompts which are there. Because it shows "Invalid response from the model, trying again". This comes when the output from the LLM is a bit wrong and doesn't get validated due to its wrong format (as per the prompt).

@ayoubachak
Copy link
Author

Umm what i feel is it at first didn't work because of the wrong response from the LLM on the prompts which are there. Because it shows "Invalid response from the model, trying again". This comes when the output from the LLM is a bit wrong and doesn't get validated due to its wrong format (as per the prompt).

I think it's because gemma wasn't good for task, however the result using mistral was a little promising given that it's a 7b model, I tried codellama 34B but it was too much for my computer, I created a pull request for the LM Studio option, maybe someone can try it out and play with the context size : #389

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants