-
Notifications
You must be signed in to change notification settings - Fork 87
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is It Resizing or Just Fusion at Corresponding Positions? #38
Comments
Yes, I have the same question. The latent feature of the sub-region is directly cropped and not resized. RPG-DiffusionMaster/cross_attention.py Lines 127 to 128 in d2a26e9
Then, the cropped features are fused with the corresponding positions of the base latent features. RPG-DiffusionMaster/cross_attention.py Lines 129 to 133 in d2a26e9
It seems not resized as the paper say. And I'd like to know why this is done, is it because resize doesn't make sense? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I'd like to ask, during the stage of regional latent space fusion in different areas, is this method really resizing to the corresponding positions? Looking at the code, it seems that only the latent spaces of the corresponding positions in each regional image are fused, which is quite confusing?
The text was updated successfully, but these errors were encountered: