From e52bdabbde3c6895aceb76c1bced295c2646121f Mon Sep 17 00:00:00 2001 From: Teven Date: Tue, 21 Feb 2023 17:56:28 +0100 Subject: [PATCH] Removed debug dump of universal checkpoints --- megatron/training.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/megatron/training.py b/megatron/training.py index bd00bc77e..815426116 100644 --- a/megatron/training.py +++ b/megatron/training.py @@ -460,9 +460,6 @@ def setup_model_and_optimizer(model_provider_func): else: args.iteration = 0 - - from .utils import dump_weights - dump_weights(f'{args.universal_checkpoint=}', args.iteration, model, optimizer) # tp_rank = mpu.get_tensor_model_parallel_rank() # pp_rank = mpu.get_pipeline_model_parallel_rank()