Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suggestions on reshaping the input data for action recognition #141

Open
sayakpaul opened this issue Dec 27, 2019 · 0 comments
Open

Suggestions on reshaping the input data for action recognition #141

sayakpaul opened this issue Dec 27, 2019 · 0 comments

Comments

@sayakpaul
Copy link

First of all, thank you very much for your hard work. I am trying to build myself an activity recognition model on the subset of UCF101 dataset (I am using the top 20 activity labels).

So far, I have used a pre-trained VGG16 network to extract the features out of the individual frames extracted from the videos. The final shape I got from the VGG16 network is (20501, 7, 7, 512) (for the train set). I now want to pass these extracted features to an LSTM-based network and I am a bit confused as to how I should reshape it?

How many time steps should I pass in and also how many features in one time-step?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant