Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Official support for finetune-based methods, e.g., vit adapter, and multi-GPU training with DataParallel. #48

Open
TimandXiyu opened this issue Sep 18, 2023 · 1 comment

Comments

@TimandXiyu
Copy link

The implementation of the gradient update in faa_model.py seems to be very constrainted at best. It does not factor in the case where I want to obtain the policy for a finetuning model, it only naively set all layer's parameter to either require or not require gradient. Besides, this is also problematic if the user want to use DP to enable multi-GPU training.

This repo really needs some updates to address the mentioned issues as well as other issues that has been mentioned in other issues.

@saigontrade88
Copy link

@TimandXiyu : can you provide more information related to your issues? It is best if you can provide detailed information related to their implementation. Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants