fix training resume from snapshot
This commit is contained in:
parent
5d2a69f56d
commit
8cd91e1c20
|
@ -32,7 +32,7 @@ $ENT \
|
|||
sde.prior_model 'models.latent_points_ada_localprior.PVCNN2Prior' \
|
||||
sde.train_vae $train_vae \
|
||||
sde.embedding_scale 1.0 \
|
||||
viz.save_freq 1000 \
|
||||
viz.save_freq 1 \
|
||||
viz.viz_freq -200 viz.log_freq -1 viz.val_freq -10000 \
|
||||
data.batch_size $BS \
|
||||
trainer.type 'trainers.train_2prior' \
|
||||
|
|
|
@ -76,7 +76,9 @@ def main(args, config):
|
|||
else:
|
||||
raise NotImplementedError
|
||||
elif args.pretrained is not None:
|
||||
trainer.load_vae(args.pretrained)
|
||||
logger.info('Resuming training from {}; if you dont want resume training, edit the cmt to change the exp name',
|
||||
args.pretrained)
|
||||
trainer.resume(args.pretrained)
|
||||
|
||||
if not args.eval_generation:
|
||||
trainer.train_epochs()
|
||||
|
|
Loading…
Reference in a new issue