From 6cc8ebd2003a4462e15b957d9438066c62266b4a Mon Sep 17 00:00:00 2001 From: Ze Liu Date: Wed, 14 Apr 2021 20:50:10 +0800 Subject: [PATCH] minor bug fixes (#25) --- main.py | 2 ++ utils.py | 2 ++ 2 files changed, 4 insertions(+) diff --git a/main.py b/main.py index eba7be3..18e9ad6 100644 --- a/main.py +++ b/main.py @@ -110,7 +110,9 @@ def main(config): if resume_file: if config.MODEL.RESUME: logger.warning(f"auto-resume changing resume file from {config.MODEL.RESUME} to {resume_file}") + config.defrost() config.MODEL.RESUME = resume_file + config.freeze() logger.info(f'auto resuming from {resume_file}') else: logger.info(f'no checkpoint found in {config.OUTPUT}, ignoring auto resume') diff --git a/utils.py b/utils.py index 0b2b07e..fd2df67 100644 --- a/utils.py +++ b/utils.py @@ -29,7 +29,9 @@ def load_checkpoint(config, model, optimizer, lr_scheduler, logger): if not config.EVAL_MODE and 'optimizer' in checkpoint and 'lr_scheduler' in checkpoint and 'epoch' in checkpoint: optimizer.load_state_dict(checkpoint['optimizer']) lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) + config.defrost() config.TRAIN.START_EPOCH = checkpoint['epoch'] + 1 + config.freeze() if 'amp' in checkpoint and config.AMP_OPT_LEVEL != "O0" and checkpoint['config'].AMP_OPT_LEVEL != "O0": amp.load_state_dict(checkpoint['amp']) logger.info(f"=> loaded successfully '{config.MODEL.RESUME}' (epoch {checkpoint['epoch']})")