Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About Guided Attention loss #45

Open
ywh-my opened this issue Nov 3, 2021 · 2 comments
Open

About Guided Attention loss #45

ywh-my opened this issue Nov 3, 2021 · 2 comments

Comments

@ywh-my
Copy link

ywh-my commented Nov 3, 2021

Hello . I try to use your model to finish melspec to melspec conversion task.

I try to add a guided attention like this :

dal_loss = (hp.lamda_attn_dal * dal_g * attn_probs).abs().mean()
dal_g = guided_attention(Nt, Ns) ## This function is included in your code ,but not used .

And I found that the loss can not decay while seq loss decline fastly . And my train can not converge.

I wish you help ,please!Thank you.

@vskadandale
Copy link

Hi, did you get this working finally? While training on LJSpeech dataset, I notice that the diagonal alignment doesn't appear in decoder attention and encoder-decoder attention, but only in the encoder attention around 160K iterations. Please let know if you had similar issues. Many thanks!

@ywh-my
Copy link
Author

ywh-my commented Mar 31, 2023

sorry , I abandoned this project long ago. >_<

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants