You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is this a new feature, an improvement, or a change to existing functionality?
New Feature
How would you describe the priority of this feature request
Medium
Please provide a clear description of problem you would like to solve.
Currently, the various weather prediction models within Modulus utilize distinct training recipes, each with unique requirements for configurations, multi-step rollouts, checkpoint saving and loading, runtime optimizations, static datasets, and data loading. This can be cumbersome for new users, who must familiarize themselves with the separate requirements of each model. As such, there is a clear need for a unified training recipe that can facilitate quick experimentation across multiple models for users.
As a first tep, we can develop a basic unified recipe for AFNO and SFNO without model/tensor parallelization.
Describe any alternatives you have considered
N/A
Additional context
N/A
The text was updated successfully, but these errors were encountered:
NickGeneva
changed the title
[FEA 🚀]: Unified training recipe for weather prediction models
🚀[FEA]: Unified training recipe for weather prediction models
Jul 26, 2023
mnabian
changed the title
🚀[FEA]: Unified training recipe for weather prediction models
🚀[FEA]: Basic unified training recipe for AFNO & SFNO (non-parallel)
Aug 10, 2023
Is this a new feature, an improvement, or a change to existing functionality?
New Feature
How would you describe the priority of this feature request
Medium
Please provide a clear description of problem you would like to solve.
Currently, the various weather prediction models within Modulus utilize distinct training recipes, each with unique requirements for configurations, multi-step rollouts, checkpoint saving and loading, runtime optimizations, static datasets, and data loading. This can be cumbersome for new users, who must familiarize themselves with the separate requirements of each model. As such, there is a clear need for a unified training recipe that can facilitate quick experimentation across multiple models for users.
As a first tep, we can develop a basic unified recipe for AFNO and SFNO without model/tensor parallelization.
Describe any alternatives you have considered
N/A
Additional context
N/A
The text was updated successfully, but these errors were encountered: