You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Ran again, and it recomputed stuff and wrote files (good).
Changed the parameter back to its previous value.
Ran again, and *it didn't rerun" because it said it had cached.
I think (5) is a bug -- somehow the file being there its mtime or hash or whatever didn't cause it to run again. This seems weird/bad to me. But I don't think it's too surprising since I think we only check the mtime of the input files and not the output files. Somehow we need to check to make sure that the output files haven't been overwritten (or modified) when determining if we need to rerun or not.
The text was updated successfully, but these errors were encountered:
It seems like we should be able to overcome it by:
Storing and the mtime/hash of the out_files when actually running the step and returning {key: (filename, hash) ... } as the out_files dict of the step.
When we do the joblib caching check, if they say the step is done/complete, we manually additionally check the hash/mtime of all the filenames and force a re-run if they've changed from what's expected.
I think it should work and not add much overhead. (Well, the minimum required amount of overhead I guess.)
I was just running some tests and:
derivatives
(byrm -Rf
ing them)I think (5) is a bug -- somehow the file being there its
mtime
or hash or whatever didn't cause it to run again. This seems weird/bad to me. But I don't think it's too surprising since I think we only check themtime
of the input files and not the output files. Somehow we need to check to make sure that the output files haven't been overwritten (or modified) when determining if we need to rerun or not.The text was updated successfully, but these errors were encountered: