-
Notifications
You must be signed in to change notification settings - Fork 327
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
4张3080ti跑chatglm2-6b-lora报oom #151
Labels
Comments
通常来说,使用lora训练chatglm2,确实只需要14g左右(在batchsize=1,文本长度也不太长的情况下(比如长度为512))。 但是也有例外,会导致你的显存不够,建议你检查一下下面内容:
基本上就这些,希望可以帮到你 |
好的,我试试。谢谢 |
已解决,transformers更新到最新版 |
请问chatglm2-6b-lora训练可以设置多轮的epoch吗?我没找到在哪设置的 |
请问一下你用的transformers的版本是多少? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
你好,我用4张12g 3080ti想跑chatglm2微调,但是报显存不足; 使用你列的只需14G显存的参数后也不行;使用int8加载模型后在训练时报了oom。 上面都是在模型并行参数开启下跑的。
在关闭模型并行参数时,也会在训练时报oom, 这时候只会用到一张卡。
请问这种情况正常吗,我该如何解决呢?
The text was updated successfully, but these errors were encountered: