DPR finetuning
#5645
Replies: 1 comment 1 reply
-
Hi @rnyak! Hard-negatives are added to the batch and are additionally used as in-batch negatives. A batch for DPR training would therefore look like this: If we wouldn't add the hard-negative passages, the training task would be a lot easier for the model as the in-batch negatives would only consist of random passages that probably are not related at all to the query. By adding hard-negative passages, we make sure to add passages that are related to the query (by for example being about the same topic), but are not relevant for answering the query. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I was going over this tutorial, but one thing is not clear to me. The DPR format is asking for hard-negs and the default it is
30
. but then there is a statement in the tutorial saying:so my question, if there is in-batch negative sampling, why
hard-negs
are required in the dpr dataset format? Does that mean, if one provides hard-negs, are those gonna be used during fine-tuning, instead of generating hard-negs via in-batch sampling?Thanks.
Beta Was this translation helpful? Give feedback.
All reactions