Merge Spawn Strategy in Lite #14707
Labels
fabric
lightning.fabric.Fabric
priority: 1
Medium priority task
refactor
strategy: ddp
DistributedDataParallel
Milestone
🚀 Feature
The DDPStrategy and DDPSpawnStrategy are two very similar strategies. Fundamentally, they only differ in the way processes are managed. Their logic can be merged and de-duplicated, since over time we have factored out all things related to process launching and environments to separate classes.
Motivation
Pitch
Merge the two implementations (resulting in the class named DDPStrategy). Register the same class with different init args for launcher settings.
Logical consequence: DDPShardedStrategy and DDPSpawnShardedStrategy will also merge.
Note:
Alternatives
Leave as is.
Additional context
Lite has now their own implementations
If you enjoy Lightning, check out our other projects! ⚡
Metrics: Machine learning metrics for distributed, scalable PyTorch applications.
Lite: enables pure PyTorch users to scale their existing code on any kind of device while retaining full control over their own loops and optimization logic.
Flash: The fastest way to get a Lightning baseline! A collection of tasks for fast prototyping, baselining, fine-tuning, and solving problems with deep learning.
Bolts: Pretrained SOTA Deep Learning models, callbacks, and more for research and production with PyTorch Lightning and PyTorch.
Lightning Transformers: Flexible interface for high-performance research using SOTA Transformers leveraging PyTorch Lightning, Transformers, and Hydra.
cc @justusschock @awaelchli @rohitgr7 @tchaton @carmocca @kaushikb11 @akihironitta
The text was updated successfully, but these errors were encountered: