Add restart reproducibility test into repro CI #30
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Enable restart reproducibility checksum checks - as detailed in #11
I've added the marker
checksum
to the restart checksum tests, added a model-specific checksum comparison method (as format of checksums may different depending on the model).For access-om2 model, it is basically doing the same check as before, i.e any field and checksum value in the
2-day
model run is in either of the first or second1 day
model runs.I have manually tested it with a
1deg_jra55_ryf
configuration where I’ve just updated the executables paths to point to the corresponding pre-release locations with the fix - config.yaml used. Currently it's failing thetest_bit_repro_historical
test - so the checksums (historical-3hr-checksum.json) are different to checksums saved on the release configuration branch, is that to be expected?Also have noticed the tests are taking ages to run - as it's mostly waiting in queues - wondered whether its worth dropping the qsub walltime and memory for short tests. But that might not be the best idea as different configurations may need varying resources..
With adding the
checksum
marker, this pytest will be run as part of the reproducibility checks. It outputs some checksum files if it fails, I haven't added them to thechecksum
output sub-directory as that isrsync
ed to github. Should it bersync
ed to github? I don't think it should be added to thetesting/checksum
on each configuration branch. Another todo, is maybe update the reproducibility fail messages in the CI as the test can fail in either the historical checksums or the restart test.