New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for more sizes of LLaMA #169
Comments
That would be really helpful! |
I think this is the standard repo which people use - https://github.com/qwopqwop200/GPTQ-for-LLaMa |
Hi @cnbeining @shreyansh26, |
Hi again, @cnbeining @shreyansh26 |
We are also working on addition kbit quantisation to generic model, so it should be released soon. |
Hey folks,
Trying to get 13/30B model with 4 bit fine tuning - any chance you folk could release the script used to convert the 7B version of model to 4 bit?
Thanks,
The text was updated successfully, but these errors were encountered: