Skip to content

Commit

Permalink
bump batch size and dial down logging for plain unet
Browse files Browse the repository at this point in the history
  • Loading branch information
henryaddison committed Aug 5, 2024
1 parent bbc20c6 commit b52c8b7
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 10 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -27,10 +27,11 @@ def get_default_configs():

# training
config.training = training = ml_collections.ConfigDict()
training.batch_size = 16#128
training.n_epochs = 20
training.batch_size = 16
training.snapshot_freq = 25
training.log_freq = 50
training.eval_freq = 1000
training.log_freq = 500
training.eval_freq = 5000
## store additional checkpoints for preemption in cloud computing environments
training.snapshot_freq_for_preemption = 1000
## produce samples at each snapshot.
Expand All @@ -41,9 +42,6 @@ def get_default_configs():
training.random_crop_size = 0
training.continuous = True
training.reduce_mean = True
training.n_epochs = 20
training.snapshot_freq = 5
training.eval_freq = 5000
training.sde = ""

# sampling
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,10 +29,8 @@ def get_config():
# training
training = config.training
training.n_epochs = 100
training.batch_size = 64
training.snapshot_freq = 25
training.log_freq = 50
training.eval_freq = 1000
training.snapshot_freq = 20
training.batch_size = 256

# data
data = config.data
Expand Down

0 comments on commit b52c8b7

Please sign in to comment.