-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to evaluate retriever? #336
Comments
If you have your custom dataset in SQuAD format, you can evaluate the retriever as shown in Tutorial 5.
See the tutorial for details. (Just be aware that the metrics of reader.eval() are currently broken and show very low numbers. We work on a fix in #331. The retriever metrics that you are interested in are correct) Did that answer your question @samreenkazi? |
@tholor Question is I have( Q,P,A )triple as the training dataset, the reader is trained on it and it was created from Wikipedia dump three years ago, In what form should we give query to retriever and how do we evaluate the result since we don't have ground truth passage in that case |
Not sure if I fully understand. When you have triples of Query, Passage, Answer you do have some ground truth data. You can basically take one triple, give the Query to the Retriever, and then check whether the Passage is within the top-k results of the retriever. Or am I missing anything here? If that's what you are after, see my post above for hints how to do that in haystack. |
Closing due to inactivity. Feel free to re-open. |
(originally posted by @samreenkazi in #334)
┆Issue is synchronized with this Jira Task by Unito
The text was updated successfully, but these errors were encountered: