-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
When should we update dvc.lock and dvc push in GitHub Actions? #393
Comments
See previous discussion comments: |
Just copy another comment to here: I am more curious about the recommended way for this demo
In the experiment pull request, it also not update the Would be great to have some recommendations! 😃 To be more specific, it would be great to cover these scenarios and still not mess up the DVC:
|
Another option that I was taught to do was use DVC's run-cache. In the CML runner:
On local machine
This is less automated than the new |
I've been wondering about all these questions myself and haven't found a satisfying answer yet. @mattlbeck Wouldn't |
Relevant links for any future readers: Why? I'm assuming the use case to be for rerunning or inspecting dvc exp/repro conducted from CI/CD (user can preform commit / or / or discard changes instead of CI) |
Discussed in iterative/dvc#6542
Originally posted by Hongbo-Miao September 6, 2021
Currently, my GitHub Actions workflow looks like this: when I open a pull request (change some model codes / params), CML creates a AWS EC2 instance, and DVC pull the data.
Here is my current GitHub Actions workflow:
Click to expand!
My dvc.yaml looks like this:
After training, if I think the change is good because the performance is better based on the reports,
model.pt
needs to be uploaded to AWS S3 in my case.My question is, after
dvc repro
, am I supposed to adddvc push
and then commit? Something likeThis above method will apply when the pull request is open.
However, I kind of feeling the best moment adding would be when I decide merging because I think this is a good pull request that actually improves the machine learning performance. But at this moment, the EC2 instance has been destroyed.
Any suggestion? Thanks!
The text was updated successfully, but these errors were encountered: