Skip to content

Commit

Permalink
[pre-commit.ci] auto fixes from pre-commit.com hooks
Browse files Browse the repository at this point in the history
for more information, see https://pre-commit.ci
  • Loading branch information
pre-commit-ci[bot] committed Oct 1, 2022
1 parent 80d24fe commit 05ac035
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion src/lightning_lite/plugins/precision/fsdp.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,9 @@
class FSDPPrecision(NativeMixedPrecision):
"""AMP for Fully Sharded Data Parallel training."""

def __init__(self, precision: Literal[16, "bf16"], device: str, scaler: Optional["ShardedGradScaler"] = None) -> None:
def __init__(
self, precision: Literal[16, "bf16"], device: str, scaler: Optional["ShardedGradScaler"] = None
) -> None:
if not _TORCH_GREATER_EQUAL_1_12:
raise RuntimeError("`FSDPPrecision` is supported from PyTorch v1.12.0 onwards.")

Expand Down

0 comments on commit 05ac035

Please sign in to comment.