-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Stuck at Quantization? or just taking a long time to run? #4
Comments
@P15V an hour and a half is too long, it is probably froze up. Try rebooting it and then mounting more SWAP memory, disabling ZRAM, and if needed disable the desktop GUI like here: https://github.com/dusty-nv/jetson-containers/blob/master/docs/setup.md#mounting-swap Also, try testing |
Hello all,
I hope whoever reads this is doing well!! :)
So I'm trying to get this going on my Jetson Nano 8GB. I'm getting stuck (maybe?) at Quantization. I run this command, and I get the terminal output that it's quantizing the model and that this will take a while. And it seems to lock up/get stuck there? I've had it going for the past 1-1.5 hours with no further outputs or such, and the entire Jetson Nano is locked up. I can't interact with it, can't SSH into it.
Do you know if this is normal? or is something going wrong? Am I doing something wrong? I'm going to let it run for a few hours to see if it accomplishes anything.
Thanks for everyone's time!! :) . My run command inside the container :
python3 -m nano_llm.chat --api=mlc
--model Efficient-Large-Model/VILA-2.7b
--max-context-len 128
--max-new-tokens 32
The text was updated successfully, but these errors were encountered: