Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Timecourse regression model #33

Closed
wants to merge 0 commits into from
Closed

Conversation

namsaraeva
Copy link
Collaborator

A draft of the simple VGG-based neural network with a single output instead of the last classification layer.

@namsaraeva namsaraeva self-assigned this Apr 17, 2024
@namsaraeva namsaraeva changed the base branch from main to development April 17, 2024 14:30
@namsaraeva
Copy link
Collaborator Author

Note to self: look at AdamW as optimizer

Copy link
Collaborator

@sophiamaedler sophiamaedler left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good so far. Probably it would be a good idea to also implement AdamW as a potential optimiser.

@namsaraeva
Copy link
Collaborator Author

AdamW's weight decay default value is 1e-2, but I am not sure if I should add it as self.hparams["weight_decay"] = 10 ** -2 too

@namsaraeva
Copy link
Collaborator Author

namsaraeva commented May 13, 2024

It should be possible to pass on a column name instead of column index for target_col (the col with float timepoint targets) in HDF5SingleCellDatasetRegression

@namsaraeva
Copy link
Collaborator Author

namsaraeva commented May 14, 2024

@sophiamaedler should we stick to the VGG architecture with 11 or 16 layers? for MLP: config A or B?

UPD: stick to the default architecture for now

@namsaraeva
Copy link
Collaborator Author

Consider adding HuberLoss instead of MSE, MAE

@namsaraeva
Copy link
Collaborator Author

Consider adding cross validation or grid search to test different settings for delta in Huber loss to get an optimal value.

"Huber loss will clip gradients to delta for residual (abs) values larger than delta. You want that when some part of your data points poorly fit the model and you would like to limit their influence." - we probably want to limit the influence of phenotypes that are similar to previously seen phenotypes? @sophiamaedler

@namsaraeva
Copy link
Collaborator Author

Current limitation: scatterplot lims when training. We have to somehow pass the min and max values of dataset to the Callback module.

Also implement a flexible Callback module that can be re-used.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants