You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We need an evaluation metric that gives us some qualitative sense of how well a model is performing.
From a performance what we care about is whether a search query is correctly mapped to the code that goes with that query. So for the test/evaluation we can compute the number of correctly matched and incorrectly matched examples.
Given a training example (Qi, Ci) where Qi is the query and Ci is the code that matches it, the example is correctly classified if
distance(Qi, Ci) <= (Qi, Cj) for j not equal to i for some set of code examples
So with this metric you're taking the mean distance to all the non-matching examples (that you sample) and asking wither the distance to the matching example is much less? That's a good measure. You can also relate that to the distance you plan to use when looking up queries. Also maybe typo, I think you meant distance(Qi, Ci).
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
We need an evaluation metric that gives us some qualitative sense of how well a model is performing.
From a performance what we care about is whether a search query is correctly mapped to the code that goes with that query. So for the test/evaluation we can compute the number of correctly matched and incorrectly matched examples.
Given a training example (Qi, Ci) where Qi is the query and Ci is the code that matches it, the example is correctly classified if
distance(Qi, Ci) <= (Qi, Cj) for j not equal to i for some set of code examples
Related to #239 Train a high quality model
The text was updated successfully, but these errors were encountered: