-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
dvc metrics diff #2995
Comments
A workaround would be pretty trivial though. |
@efiop yeap! Plain But agree, we just need a nice looking shortcut for that. |
Are you guys contemplating in the specs every stage of the pipeline? {
"train": {
"train_time": "3d 8h 23m 15s",
"memory_consume": "8Gb"
},
"eval": {
"inference_time": 0.001,
"memory_consume": "124Mb",
"top1-error": 0.0385,
"top5-error": 0.039221
}
} |
For starters, I would not support them, but I think at some point we could add functionality allowing the user to define how to calculate the difference between particular metrics, by executing some custom-defined method. That could resemble how we are dealing with Question from me: |
👍 I have the same thoughts - we need a "custom" metrics file to support multiple formats (our basic json + csv are not enough). Probably, different types of metrics can be represented as separate summoning-objects (not necessary metrics inside summoning object). Metrics are a bit more complicated than just numbers. There are numbers, numbers with сonfidence level, confusion matrixes and so one. It is not easy to find a single way of dealing with them. Even for numbers metrics, it is important to know if you are minimizing or maximizing it (is +013 good or bad?) to visualize properly and find the "best one".
Only two. The same idea as |
One of the best features would be observability. One of the best features of dvc is having all the metrics all together. I would vote for having that somehow. |
@DavidGOrtega @dmpetrov Well, that seems like a conflict "behave as git" vs "this might be useful in ML use-case". I guess we can always think about implementing another new metrics command ( |
@DavidGOrtega I'd appreciate if you could provide a use case (with a command example) when it will be helpful (and not easily replicable with multiple |
@dmpetrov is your question is regarding my question
Im just only asking without any intention of going towards it, was only to have a better picture. If your question is why "all together" observability would be one of the best features is to have a general overview of the experiment. Conceptually an experiment can have many permutations of data, parameters and even implementation but at the end what should matter is how it performs according to the metrics that you want to measure. So, yes, you could be able to do multiple diffs but for an experiment with many trials that would be difficult to handle. |
Hi. Is this closed by #3051? (docs on their way too: iterative/dvc.org/pull/933) p.s. I know I'm late to this party but it seems more like a |
Please note that we already provided a short script for this same workaround some time ago in #770 (comment) ! @efiop @dmpetrov
Agree. For certain numeric scales and ranges a simple B-A calculation may yield no meaning. For these cases maybe add |
@jorgeorpinel I was expecting some follow up requirements on this one, but looks like everything mentioned in this ticket is already implemented. So let's close it for now. Thanks for the heads up!
But this diff is a metrics diff, direct comparison with |
@jorgeorpinel Re the ‘git diff’ - it is a problem of Git which just cannot quantify the difference. If they can they would show numbers instead of line-by-line diff 😀 But really, if number diff does not work for dvc metrics it means the metrics were not defined properly. We need to introduce a more strict requirements for metric files. |
Today, we can track metrics but metrics become much more valuable when you can see differences/improvements over time (commits/branches). A new
dvc metrics diff
command is needed.Open question: What should we do about not float/integer metrics? Let's don't support (ignore) them. Any other ideas?
Note, we should deal with float formatting carefully - we don't want to see diff values like
0.0001624000000000000001
.Also, an easy-to-parse output option is required:
The text was updated successfully, but these errors were encountered: