Replies: 1 comment 1 reply
-
|
@dimitry12 Have you tried using DVC Metrics and Comparisons. Try generate comparison metrics as part of your pipeline |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Here is an illustration of the scenario:
In this scenario I care about agreement/disagreement across variations. For example, I only want to spend time inspecting the results if they are sufficiently distinct from baseline.
What's the best approach for doing this with dvc? The very basic constraint I am bumping into is the fact that dvc-experiment variations' outs never even exist at the same time in the workspace.
I can only think of taking "outs" of baseline experiment and duplicating them as silver-labels. Then these silver-labels will exist across all variations and I can metric-compare variations-outs against silver-labels. But this gets out of hand fast because as soon as I find sufficiently distinct new variation, I will need a second set of silver-labels.
Beta Was this translation helpful? Give feedback.
All reactions