Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improvment to 97% accuracy #10

Open
Ananas120 opened this issue Aug 13, 2020 · 4 comments
Open

Improvment to 97% accuracy #10

Ananas120 opened this issue Aug 13, 2020 · 4 comments

Comments

@Ananas120
Copy link

Hello, i don’t know if this repo is active but if i can help, i used this repo for my project and i find a method to improve the loss / accuracy just by L2-normalizing the output of the encoder
My score with it is actually 0.97% accuracy and 0.02 val-BCEloss training for 25 epochs on a mixt of LibriSpeech and CommonVoice (fr) datasets (360 speakers in train set and 150 in validation set with 200 pairs for each speaker (100 same and 100 not same) (batch_size of size 32 (16 same and 16 not) with embedding dim 64)

@Vatsalparsaniya
Copy link

@Ananas120 Hi, are you using python 2.7 or 3.*?

@Ananas120
Copy link
Author

I re-implemented your model architecture in python 3 in my personal project

@Vatsalparsaniya
Copy link

Vatsalparsaniya commented May 6, 2021

Thank you for your response,
I have just reimplemented train_siamese.py code in python 3 and currently training on google colab.
image

I'm getting accuracy value == 0.5000 since first training step. Is it normal or I'm missing something.
if it is possible can you share training log?

Thank you.

@Ananas120
Copy link
Author

Yes it is normal ! I found that siamese networks need some epoch (sometimes more than 5 or 10 !) to increase their accuracy and this can be explained as follow : they try to separate « same » and « not same » from a scalar value (distance) so they have to find the right threshold before having a better score (before that they always predict either 0 or 1)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants