-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Timecourse regression model #33
Conversation
Note to self: look at AdamW as optimizer |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good so far. Probably it would be a good idea to also implement AdamW as a potential optimiser.
AdamW's weight decay default value is 1e-2, but I am not sure if I should add it as |
It should be possible to pass on a column name instead of column index for target_col (the col with float timepoint targets) in HDF5SingleCellDatasetRegression |
@sophiamaedler should we stick to the VGG architecture with 11 or 16 layers? for MLP: config A or B? UPD: stick to the default architecture for now |
Consider adding HuberLoss instead of MSE, MAE |
Consider adding cross validation or grid search to test different settings for delta in Huber loss to get an optimal value. "Huber loss will clip gradients to delta for residual (abs) values larger than delta. You want that when some part of your data points poorly fit the model and you would like to limit their influence." - we probably want to limit the influence of phenotypes that are similar to previously seen phenotypes? @sophiamaedler |
Current limitation: scatterplot lims when training. We have to somehow pass the min and max values of dataset to the Callback module. Also implement a flexible Callback module that can be re-used. |
de8eeac
to
e001080
Compare
00c5590
to
e001080
Compare
A draft of the simple VGG-based neural network with a single output instead of the last classification layer.