Skip to content

Commit 40d7663

Browse files
Merge pull request #304 from Parry-Parry/master
feat: add DL 2023 qrels for MSMARCO passage and document v2
2 parents aebd707 + 75c558f commit 40d7663

File tree

8 files changed

+99
-3
lines changed

8 files changed

+99
-3
lines changed

ir_datasets/datasets/msmarco_document_v2.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -222,8 +222,15 @@ def _init():
222222
subsets['trec-dl-2023'] = Dataset(
223223
collection,
224224
TsvQueries(dlc['trec-dl-2023/queries'], namespace='msmarco', lang='en'),
225+
TrecQrels(dlc['trec-dl-2023/qrels'], TREC_DL_QRELS_DEFS),
225226
TrecScoredDocs(GzipExtract(dlc['trec-dl-2023/scoreddocs'])),
226227
)
228+
dl23_judged = Lazy(lambda: {q.query_id for q in subsets['trec-dl-2023'].qrels_iter()})
229+
subsets['trec-dl-2023/judged'] = Dataset(
230+
FilteredQueries(subsets['trec-dl-2023'].queries_handler(), dl23_judged),
231+
FilteredScoredDocs(subsets['trec-dl-2023'].scoreddocs_handler(), dl23_judged),
232+
subsets['trec-dl-2023'],
233+
)
227234

228235
subsets['anchor-text'] = Dataset(
229236
MsMarcoV2AnchorTextDocs(

ir_datasets/datasets/msmarco_passage_v2.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -316,8 +316,15 @@ def _init():
316316
subsets['trec-dl-2023'] = Dataset(
317317
collection,
318318
TsvQueries(dlc['trec-dl-2023/queries'], namespace='msmarco', lang='en'),
319+
TrecQrels(dlc['trec-dl-2023/qrels'], TREC_DL_QRELS_DEFS),
319320
TrecScoredDocs(GzipExtract(dlc['trec-dl-2023/scoreddocs'])),
320321
)
322+
dl23_judged = Lazy(lambda: {q.query_id for q in subsets['trec-dl-2023'].qrels_iter()})
323+
subsets['trec-dl-2023/judged'] = Dataset(
324+
FilteredQueries(subsets['trec-dl-2023'].queries_handler(), dl23_judged),
325+
FilteredScoredDocs(subsets['trec-dl-2023'].scoreddocs_handler(), dl23_judged),
326+
subsets['trec-dl-2023'],
327+
)
321328

322329
ir_datasets.registry.register(NAME, Dataset(collection, documentation("_")))
323330
for s in sorted(subsets):

ir_datasets/docs/msmarco-document-v2.yaml

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -130,6 +130,18 @@ trec-dl-2023:
130130
<p>
131131
Official topics for the TREC Deep Learning (DL) 2023 shared task.
132132
</p>
133+
<p>
134+
Note that these qrels are <i>inferred</i> from the passage ranking task; a document''s relevance
135+
label is the maximum of the labels of its passages.
136+
</p>
137+
'
138+
139+
trec-dl-2023/judged:
140+
desc: '
141+
<p>
142+
<a class="ds-ref">msmarco-document-v2/trec-dl-2023</a>, but filtered down to only the queries
143+
with qrels.
144+
</p>
133145
'
134146

135147
anchor-text:

ir_datasets/docs/msmarco-passage-v2.yaml

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -104,4 +104,18 @@ trec-dl-2023:
104104
<p>
105105
Official topics for the TREC Deep Learning (DL) 2023 shared task.
106106
</p>
107+
<p>
108+
Note that the officially-released qrels <i>include</i> relevance labels propagated to
109+
duplicate passages, while results presented in the notebook papers remove duplicate documents.
110+
This means that the results are not directly comparable, and extra care should be taken when
111+
making comparisons among systems to ensure that they were evaluated in the same settings.
112+
</p>
113+
'
114+
115+
trec-dl-2023/judged:
116+
desc: '
117+
<p>
118+
<a class="ds-ref">msmarco-passage-v2/trec-dl-2023</a>, but filtered down to only the queries
119+
with qrels.
120+
</p>
107121
'

ir_datasets/etc/downloads.json

Lines changed: 15 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4711,9 +4711,16 @@
47114711
"size_hint": 902168,
47124712
"expected_md5": "0f5d548e53afb9e319c837ad67f9046a",
47134713
"cache_path": "trec-dl-2023/top100.txt.gz"
4714+
},
4715+
"trec-dl-2023/qrels": {
4716+
"url": "https://trec.nist.gov/data/deep/2023.qrels.docs.withDupes.txt",
4717+
"irds_mirror": true,
4718+
"size_hint": 675015,
4719+
"expected_md5": "1e9c540b3cb03bcc975a583586c04090",
4720+
"cache_path": "trec-dl-2023/qrels.withDupes.txt"
47144721
}
47154722
},
4716-
4723+
47174724
"msmarco-passage": {
47184725
"collectionandqueries": {
47194726
"url": "https://msmarco.z22.web.core.windows.net/msmarcoranking/collectionandqueries.tar.gz",
@@ -4939,6 +4946,13 @@
49394946
"expected_md5": "c339ed75e1556cacb387899f34cadad1",
49404947
"cache_path": "trec-dl-2023/top100.txt.gz"
49414948
},
4949+
"trec-dl-2023/qrels": {
4950+
"url": "https://trec.nist.gov/data/deep/2023.qrels.pass.withDupes.txt",
4951+
"irds_mirror": true,
4952+
"size_hint": 912450,
4953+
"expected_md5": "3a742d51ae65da2ece9c09b304b9e358",
4954+
"cache_path": "trec-dl-2023/qrels.withDupes.txt"
4955+
},
49424956
"dedup_positions": {
49434957
"url": "https://huggingface.co/datasets/macavaney/msmarco-passage-v2-duplicate-ids/resolve/main/dedup_positions.tar.gz",
49444958
"size_hint": 388936118,

ir_datasets/etc/metadata.json

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -509,7 +509,8 @@
509509
"msmarco-document-v2/trec-dl-2021/judged": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 57}, "qrels": {"_ref": "msmarco-document-v2/trec-dl-2021"}, "scoreddocs": {"count": 5700}},
510510
"msmarco-document-v2/trec-dl-2022": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 500}, "scoreddocs": {"count": 50000}, "qrels": {"count": 369638, "fields": {"relevance": {"counts_by_value": {"0": 274971, "1": 49016, "2": 44255, "3": 1396}}}}},
511511
"msmarco-document-v2/trec-dl-2022/judged": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 76}, "qrels": {"_ref": "msmarco-document-v2/trec-dl-2022"}, "scoreddocs": {"count": 7600}},
512-
"msmarco-document-v2/trec-dl-2023": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 700}, "scoreddocs": {"count": 70000}},
512+
"msmarco-document-v2/trec-dl-2023": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 700}, "scoreddocs": {"count": 70000}, "qrels": {"count": 18034, "fields": {"relevance": {"counts_by_value": {"0": 11876, "1": 3015, "2": 1668, "3": 1475}}}}},
513+
"msmarco-document-v2/trec-dl-2023/judged": {"docs": {"_ref": "msmarco-document-v2"}, "queries": {"count": 82}, "qrels": {"_ref": "msmarco-document-v2/trec-dl-2023"}, "scoreddocs": {"count": 8200}},
513514
"msmarco-document/anchor-text": {"docs": {"count": 1703834, "fields": {"doc_id": {"max_len": 8, "common_prefix": "D"}}}},
514515
"msmarco-document/dev": {"docs": {"_ref": "msmarco-document"}, "queries": {"count": 5193}, "qrels": {"count": 5193, "fields": {"relevance": {"counts_by_value": {"1": 5193}}}}, "scoreddocs": {"count": 519300}},
515516
"msmarco-document/eval": {"docs": {"_ref": "msmarco-document"}, "queries": {"count": 5793}, "scoreddocs": {"count": 579300}},
@@ -535,7 +536,8 @@
535536
"msmarco-passage-v2/trec-dl-2021/judged": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 53}, "qrels": {"_ref": "msmarco-passage-v2/trec-dl-2021"}, "scoreddocs": {"count": 5300}},
536537
"msmarco-passage-v2/trec-dl-2022": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 500}, "scoreddocs": {"count": 50000}, "qrels": {"count": 386416, "fields": {"relevance": {"counts_by_value": {"0": 286459, "1": 52218, "2": 46080, "3": 1659}}}}},
537538
"msmarco-passage-v2/trec-dl-2022/judged": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 76}, "qrels": {"_ref": "msmarco-passage-v2/trec-dl-2022"}, "scoreddocs": {"count": 7600}},
538-
"msmarco-passage-v2/trec-dl-2023": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 700}, "scoreddocs": {"count": 70000}},
539+
"msmarco-passage-v2/trec-dl-2023": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 700}, "scoreddocs": {"count": 70000}, "qrels": {"count": 22327, "fields": {"relevance": {"counts_by_value": {"0": 13866, "1": 4372, "2": 2259, "3": 1830}}}}},
540+
"msmarco-passage-v2/trec-dl-2023/judged": {"docs": {"_ref": "msmarco-passage-v2"}, "queries": {"count": 82}, "qrels": {"_ref": "msmarco-passage-v2/trec-dl-2023"}, "scoreddocs": {"count": 8200}},
539541
"msmarco-passage/dev": {"docs": {"_ref": "msmarco-passage"}, "queries": {"count": 101093}, "qrels": {"count": 59273, "fields": {"relevance": {"counts_by_value": {"1": 59273}}}}},
540542
"msmarco-passage/dev/2": {"docs": {"_ref": "msmarco-passage"}, "queries": {"count": 4281}, "qrels": {"count": 4655, "fields": {"relevance": {"counts_by_value": {"1": 4655}}}}},
541543
"msmarco-passage/dev/judged": {"docs": {"_ref": "msmarco-passage"}, "queries": {"count": 55578}, "qrels": {"_ref": "msmarco-passage/dev"}},

test/integration/msmarco_document_v2.py

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -74,6 +74,11 @@ def test_queries(self):
7474
9: GenericQuery('2001686', 'good food and bad food for high cholesterol'),
7575
699: GenericQuery('3100949', 'How do birth control and hormone levels affect menstrual cycle variations?'),
7676
})
77+
self._test_queries('msmarco-document-v2/trec-dl-2023/judged', count=82, items={
78+
0: GenericQuery('2001010', 'cost comparison of funerals in australia'),
79+
9: GenericQuery('2003787', 'how to make linkedin private'),
80+
81: GenericQuery('3100922', 'What is the meaning and origin of the name Corrin?'),
81+
})
7782

7883
def test_qrels(self):
7984
self._test_qrels('msmarco-document-v2/train', count=331956, items={
@@ -131,6 +136,16 @@ def test_qrels(self):
131136
9: TrecQrel('2000511', 'msmarco_doc_00_928870217', 0, '0'),
132137
369637: TrecQrel('2056323', 'msmarco_doc_59_419476385', 1, '0'),
133138
})
139+
self._test_qrels('msmarco-document-v2/trec-dl-2023', count=18034, items={
140+
0: TrecQrel('2001010', 'msmarco_doc_00_1413647782', 0, '0'),
141+
9: TrecQrel('2001010', 'msmarco_doc_01_710280219', 0, '0'),
142+
18033: TrecQrel('3100922', 'msmarco_doc_58_1542021309', 0, '0'),
143+
})
144+
self._test_qrels('msmarco-document-v2/trec-dl-2023/judged', count=18034, items={
145+
0: TrecQrel('2001010', 'msmarco_doc_00_1413647782', 0, '0'),
146+
9: TrecQrel('2001010', 'msmarco_doc_01_710280219', 0, '0'),
147+
18033: TrecQrel('3100922', 'msmarco_doc_58_1542021309', 0, '0'),
148+
})
134149

135150
def test_scoreddocs(self):
136151
self._test_scoreddocs('msmarco-document-v2/train', count=32218809, items={
@@ -168,6 +183,11 @@ def test_scoreddocs(self):
168183
9: GenericScoredDoc('2000138', 'msmarco_doc_50_2209372409', 13.6826),
169184
69999: GenericScoredDoc('3100949', 'msmarco_doc_42_771008717', 18.752998),
170185
})
186+
self._test_scoreddocs('msmarco-document-v2/trec-dl-2023/judged', count=8200, items={
187+
0: GenericScoredDoc('2001010', 'msmarco_doc_40_566250732', 11.2369),
188+
9: GenericScoredDoc('2001010', 'msmarco_doc_41_285574609', 9.2133),
189+
8199: GenericScoredDoc('3100922', 'msmarco_doc_19_2622779171', 9.289),
190+
})
171191

172192
def test_anchor_text(self):
173193
self._test_docs("msmarco-document-v2/anchor-text", count=4821244, items={

test/integration/msmarco_passage_v2.py

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,11 @@ def test_queries(self):
6565
9: GenericQuery('2001686', 'good food and bad food for high cholesterol'),
6666
699: GenericQuery('3100949', 'How do birth control and hormone levels affect menstrual cycle variations?'),
6767
})
68+
self._test_queries('msmarco-passage-v2/trec-dl-2023/judged', count=82, items={
69+
0: GenericQuery('2001010', 'cost comparison of funerals in australia'),
70+
9: GenericQuery('2003787', 'how to make linkedin private'),
71+
81: GenericQuery('3100922', 'What is the meaning and origin of the name Corrin?'),
72+
})
6873

6974
def test_qrels(self):
7075
self._test_qrels('msmarco-passage-v2/train', count=284212, items={
@@ -102,6 +107,16 @@ def test_qrels(self):
102107
9: TrecQrel('2000511', 'msmarco_passage_00_491585086', 0, '0'),
103108
386415: TrecQrel('2056323', 'msmarco_passage_68_715747739', 1, '0'),
104109
})
110+
self._test_qrels('msmarco-passage-v2/trec-dl-2023', count=22327, items={
111+
0: TrecQrel('2001010', 'msmarco_passage_00_257661787', 0, '0'),
112+
9: TrecQrel('2001010', 'msmarco_passage_01_221183941', 0, '0'),
113+
22326: TrecQrel('3100922', 'msmarco_passage_68_194985280', 0, '0'),
114+
})
115+
self._test_qrels('msmarco-passage-v2/trec-dl-2023/judged', count=22327, items={
116+
0: TrecQrel('2001010', 'msmarco_passage_00_257661787', 0, '0'),
117+
9: TrecQrel('2001010', 'msmarco_passage_01_221183941', 0, '0'),
118+
22326: TrecQrel('3100922', 'msmarco_passage_68_194985280', 0, '0'),
119+
})
105120

106121
def test_scoreddocs(self):
107122
self._test_scoreddocs('msmarco-passage-v2/train', count=27713673, items={
@@ -139,6 +154,11 @@ def test_scoreddocs(self):
139154
9: GenericScoredDoc('2000138', 'msmarco_passage_35_358067216', 15.2805),
140155
69999: GenericScoredDoc('3100949', 'msmarco_passage_30_84437641', 18.801701),
141156
})
157+
self._test_scoreddocs('msmarco-passage-v2/trec-dl-2023/judged', count=8200, items={
158+
0: GenericScoredDoc('2001010', 'msmarco_passage_39_224640845', 11.9941),
159+
9: GenericScoredDoc('2001010', 'msmarco_passage_60_353412311', 10.992799),
160+
8199: GenericScoredDoc('3100922', 'msmarco_passage_38_636536351', 10.032599),
161+
})
142162

143163

144164
if __name__ == '__main__':

0 commit comments

Comments
 (0)