Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training is too slow on GPU (batch size = 1) #21

Open
nrc53 opened this issue Jan 28, 2019 · 0 comments
Open

Training is too slow on GPU (batch size = 1) #21

nrc53 opened this issue Jan 28, 2019 · 0 comments

Comments

@nrc53
Copy link

nrc53 commented Jan 28, 2019

Hi,

I am training the network on NVIDIA Tesla GPU with a batch size of 1. I am training for 500K iterations, it's taking huge time, around some 600 hours. I listed below the parameters I used.
batch_size = 1
sequence_max_length = 100
words_count = 256
word_size = 32
read_heads = 4
learning_rate = 1e-4
momentum = 0.9

Any suggestions on how to improve the training time?

Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant