From 39f5f72ce143c764916d751f523e8307e765594d Mon Sep 17 00:00:00 2001 From: awaelchli Date: Sun, 28 Aug 2022 02:06:12 +0200 Subject: [PATCH] reset --- src/pytorch_lightning/strategies/ddp_spawn.py | 1 - src/pytorch_lightning/strategies/tpu_spawn.py | 1 - 2 files changed, 2 deletions(-) diff --git a/src/pytorch_lightning/strategies/ddp_spawn.py b/src/pytorch_lightning/strategies/ddp_spawn.py index 71f2e55157c57..0792c283f3f3b 100644 --- a/src/pytorch_lightning/strategies/ddp_spawn.py +++ b/src/pytorch_lightning/strategies/ddp_spawn.py @@ -171,7 +171,6 @@ def _setup_model(self, model: Module) -> DistributedDataParallel: def setup_distributed(self) -> None: log.detail(f"{self.__class__.__name__}: setting up distributed...") - reset_seed() self.set_world_ranks() rank_zero_only.rank = self.global_rank self._process_group_backend = self._get_process_group_backend() diff --git a/src/pytorch_lightning/strategies/tpu_spawn.py b/src/pytorch_lightning/strategies/tpu_spawn.py index 1e522bc7cf8df..36435ff19cb52 100644 --- a/src/pytorch_lightning/strategies/tpu_spawn.py +++ b/src/pytorch_lightning/strategies/tpu_spawn.py @@ -214,7 +214,6 @@ def reduce( def setup_distributed(self) -> None: self._launched = True - reset_seed() self.set_world_ranks() rank_zero_only.rank = self.global_rank