-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Refactor PyTorch for explicit Lighnting/Wandb hyperparameters #30
Conversation
Include Cam's latest features on generic dev
Trivial updates
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM - fwiw i've been using this version of the code in the batch processing system
@@ -26,7 +26,7 @@ | |||
'torch == 1.10.1', | |||
'torchvision == 0.11.2', | |||
'torchaudio == 0.10.1', | |||
'pytorch-lightning', | |||
'pytorch-lightning==1.6.5', # 1.7 requires protobuf version incompatible with tensorflow/tensorboard. Otherwise works. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
not blocking - in theory we could isolate the pytorch and tensorflow installs to their own python virtual envs, thus avoiding this conflict and allowing pytorch lightning to resolve as high as possible
Unless of course the tensorboard dependency i used in pytorch....then please ignore me.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we could isolate the pytorch and tensorflow installs to their own python virtual envs
Good idea but I don't know how to do that.
In practice there is very little advantage to being on the absolute latest package and I suspect the tensorboard team will update their deps shortly.
datamodule=datamodule, | ||
ckpt_path='best' # can optionally point to a specific checkpoint here e.g. "/share/nas2/walml/repos/gz-decals-classifiers/results/early_stopping_1xgpu_greyscale/checkpoints/epoch=26-step=16847.ckpt" | ||
) | ||
# trainer.test( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
any reason to remove the test step of after fitting? Does this speed up the system but impact quality?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Testing routinely is bad practice as you may accidentally tune your hparams to overfit the model. I was being a bit lazy when adding here earlier. I have left it commented as example, with a warning note.
Major changes
Minor (but potentially breaking) changes
cc @patrikasvanagas, @camallen