parquet-converter commited on
Commit
e3ced59
1 Parent(s): 128e206

Update parquet files

Browse files
dev.jsonl.gz → default/msmarco-passage-dev.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1a66e80d379ccda0f899ae40a1600c99ebdc044838a1a14da01188519c34ad9
3
- size 134665
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:614e9d8272c4bfede44bb2a50799b7127be44209c81c1bef50d8380c731d037a
3
+ size 241902
train.jsonl.gz → default/msmarco-passage-train-00000-of-00010.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9df43ab7a970fc63e9eca68b7a3a26036221339b6d2d7619d959306f96ec3193
3
- size 1726062395
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f8f1207260ccad5da7db331b560cc230dbc59b6ee3b571c77e95e7bbbf006ea
3
+ size 287688306
default/msmarco-passage-train-00001-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:625b7cb4abaa80f62b747e84a908335d1c56ac6c91d514fe65a4c4cac2b15857
3
+ size 287474499
default/msmarco-passage-train-00002-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09e71d52ce75095eb4fadc623fbff351bf30a5da83e58e84141941ecc2be3d3d
3
+ size 288082078
default/msmarco-passage-train-00003-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c663e910c518342448caa7b6cfcfa01912851b037e2c3e97a8b153d8e41f24
3
+ size 287894162
default/msmarco-passage-train-00004-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f6a0d67929dc076809b6c2cdb710079fcb684b9e220e93fb4a33a58f69c0a5f
3
+ size 287706978
default/msmarco-passage-train-00005-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44ffd39738f981fd86a0e6e4244a33ae3c4e39575669498b62ae35a88e028f42
3
+ size 287795682
default/msmarco-passage-train-00006-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d6d5cc32a57bc42c2f30cc38e5471bb3fe984915b78834e5145d22192ae4f5b
3
+ size 287535139
default/msmarco-passage-train-00007-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ed1c1e21180ba1635a181e1054e7d0293377c12dbf32d4fda633f337c2885a4
3
+ size 287735074
default/msmarco-passage-train-00008-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:182928baf18d0e23b35082e09699756e3ae9c4ba494f6ae69f8177f6e78ba828
3
+ size 287769241
default/msmarco-passage-train-00009-of-00010.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:675480aeb8076d4a33bcad2205517591fd880f63634fed02c54639d37aef9677
3
+ size 156008579
msmarco-passage.py DELETED
@@ -1,103 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.Wikipedia
15
-
16
- # Lint as: python3
17
- """MsMarco Passage dataset."""
18
-
19
- import json
20
-
21
- import datasets
22
-
23
- _CITATION = """
24
- @misc{bajaj2018ms,
25
- title={MS MARCO: A Human Generated MAchine Reading COmprehension Dataset},
26
- author={Payal Bajaj and Daniel Campos and Nick Craswell and Li Deng and Jianfeng Gao and Xiaodong Liu
27
- and Rangan Majumder and Andrew McNamara and Bhaskar Mitra and Tri Nguyen and Mir Rosenberg and Xia Song
28
- and Alina Stoica and Saurabh Tiwary and Tong Wang},
29
- year={2018},
30
- eprint={1611.09268},
31
- archivePrefix={arXiv},
32
- primaryClass={cs.CL}
33
- }
34
- """
35
-
36
- _DESCRIPTION = "dataset load script for MSMARCO Passage"
37
-
38
- _DATASET_URLS = {
39
- 'train': "https://huggingface.co/datasets/Tevatron/msmarco-passage/resolve/main/train.jsonl.gz",
40
- #'train': "https://www.dropbox.com/s/seqqbu90jopvtq5/msmarco_passage_train.json",
41
- 'dev': "https://huggingface.co/datasets/Tevatron/msmarco-passage/resolve/main/dev.jsonl.gz",
42
- }
43
-
44
-
45
- class MsMarcoPassage(datasets.GeneratorBasedBuilder):
46
- VERSION = datasets.Version("0.0.1")
47
-
48
- BUILDER_CONFIGS = [
49
- datasets.BuilderConfig(version=VERSION,
50
- description="MS MARCO passage train/dev datasets"),
51
- ]
52
-
53
- def _info(self):
54
- features = datasets.Features({
55
- 'query_id': datasets.Value('string'),
56
- 'query': datasets.Value('string'),
57
- 'positive_passages': [
58
- {'docid': datasets.Value('string'), 'title': datasets.Value('string'), 'text': datasets.Value('string')}
59
- ],
60
- 'negative_passages': [
61
- {'docid': datasets.Value('string'), 'title': datasets.Value('string'), 'text': datasets.Value('string')}
62
- ],
63
- })
64
- return datasets.DatasetInfo(
65
- # This is the description that will appear on the datasets page.
66
- description=_DESCRIPTION,
67
- # This defines the different columns of the dataset and their types
68
- features=features, # Here we define them above because they are different between the two configurations
69
- supervised_keys=None,
70
- # Homepage of the dataset for documentation
71
- homepage="",
72
- # License for the dataset if available
73
- license="",
74
- # Citation for the dataset
75
- citation=_CITATION,
76
- )
77
-
78
- def _split_generators(self, dl_manager):
79
- if self.config.data_files:
80
- downloaded_files = self.config.data_files
81
- else:
82
- downloaded_files = dl_manager.download_and_extract(_DATASET_URLS)
83
- splits = [
84
- datasets.SplitGenerator(
85
- name=split,
86
- gen_kwargs={
87
- "files": [downloaded_files[split]] if isinstance(downloaded_files[split], str) else downloaded_files[split],
88
- },
89
- ) for split in downloaded_files
90
- ]
91
- return splits
92
-
93
- def _generate_examples(self, files):
94
- """Yields examples."""
95
- for filepath in files:
96
- with open(filepath, encoding="utf-8") as f:
97
- for line in f:
98
- data = json.loads(line)
99
- if data.get('negative_passages') is None:
100
- data['negative_passages'] = []
101
- if data.get('positive_passages') is None:
102
- data['positive_passages'] = []
103
- yield data['query_id'], data