How to Evaluate Retriever on own test_data? #4034
Replies: 5 comments 7 replies
-
Hi @VikasRathod314! If you already have an annotated dataset, evaluating the retriever should be no problem: have a look at our Evaluation Documentation page and at our Evaluation tutorial. |
Beta Was this translation helpful? Give feedback.
-
@bogdankostic how i can improve the accuracy of retriever. I have trained and tested the retriever but getting less accuracy increased the dataset but also not getting improvement. Please suggest. |
Beta Was this translation helpful? Give feedback.
-
Good Morning @mayankjobanputra, Test data is about (650 rows) Generated from PseudoLabelGenerator. Also, I have checked those data and removed unwanted Questions and their respective pos_doc and neg_doc. Please do suggest to me make a better retriever as. |
Beta Was this translation helpful? Give feedback.
-
@bogdankostic this is strange. On multiple different places I find "Accuracy@k" as the main metric for DPR. Example: https://arxiv.org/pdf/2004.04906.pdf Page 5 |
Beta Was this translation helpful? Give feedback.
-
Hey @bogdankostic is it possible to evaluate the throughput and the indexing time using eval()? I know I get the logger information but is there any way I could get measures that point towards the time it takes to index, throughput, etc? Thank you :) |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I trained the retriever on our dataset using the PseudoLabelGenerator model and passed it to the model and trained
Also, found the eval_bier method to evaluate the models but they used different data formats to eval the model. but it isn't easy to evaluate our trained model. Is there any other way to eval retrievers on document search?
Beta Was this translation helpful? Give feedback.
All reactions