I have encountered a few cases where the "answers" field appears to include erroneous whitepace, or is missing a hyphen, and as a result some questions are impossible to answer.
For example:
RACE.json.gz, for question f69d72de082a4fe6bcebab8301ca52d1 the answers are:
"The positive effects of early- life exercise."
However the passage text only contains the phrase:
".... the positive effects of early-life exercise lasted for only one week"
Or for MrqaBioASQ, question 78f9bca0ee664b74b0be699e63138b9b, the answers are:
["Interferon signature", "IFN signature"]
but the only related passage phrase is:
"...for the IFN-signature as a..."
As a result it looks like it would be impossible to get an EM score one 1 on these questions if using a purely extractive approach. You can still retrieve a valid answer using the character spans, but evaluation script uses the "answer" field so it will fail models on those questions.
So far all the errors of this sort that I have seen are related to hyphens.
I have encountered a few cases where the "answers" field appears to include erroneous whitepace, or is missing a hyphen, and as a result some questions are impossible to answer.
For example:
RACE.json.gz, for question
f69d72de082a4fe6bcebab8301ca52d1the answers are:"The positive effects of early- life exercise."
However the passage text only contains the phrase:
".... the positive effects of early-life exercise lasted for only one week"
Or for MrqaBioASQ, question
78f9bca0ee664b74b0be699e63138b9b, the answers are:["Interferon signature", "IFN signature"]
but the only related passage phrase is:
"...for the IFN-signature as a..."
As a result it looks like it would be impossible to get an EM score one 1 on these questions if using a purely extractive approach. You can still retrieve a valid answer using the character spans, but evaluation script uses the "answer" field so it will fail models on those questions.
So far all the errors of this sort that I have seen are related to hyphens.