Datasets:

Modalities:
Tabular
Text
Formats:
csv
Libraries:
Datasets
pandas
License:
File size: 5,430 Bytes
2155f76
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
import os
import gzip
import zipfile

class ClueWeb22Api:

    """
    ClueWeb22Api adapted from source: https://github.com/lemurproject/ClueWeb22/blob/main/ClueWeb22Api.py
    """

    def __init__(self, cw22id, cw22root_path):
        self.cw22id = cw22id
        self.cw22root_path = cw22root_path

    def get_base_filename_by_id(self, cw22id, cw22root_path, file_type='html'):
        # /data/datasets/clueweb22/ClueWeb22_B/txt
        html_path = self.cw22root_path + os.sep + file_type

        # clueweb22-en00{subfolder_id}-{jjsongz_id}-{ddoc_id}"
        id_parts = cw22id.split('-')
        doc = int(id_parts[len(id_parts) - 1])

        language = id_parts[1][:2] # en 
        segment = id_parts[1][:4] # en00 
        directory = id_parts[1] # en0000
        # /data/datasets/clueweb22/ClueWeb22_B/txt + en + en00 + en0000 
        base_path = html_path + os.sep + language + os.sep + segment + os.sep + directory + os.sep
        base_filename = base_path + id_parts[1] + '-' + id_parts[2]
        return base_filename

    def get_primary_node_ids(self, annotate_html):
        annotations = annotate_html.annotations
        primary_node_ids = []
        for annotation in annotations:
            if annotation.type == AnnotateHtml.AnnotationType.Primary:
                primary_node_ids.append(int(annotation.nodeId))
        primary_node_ids.sort()
        return primary_node_ids

    def get_html_from_warc(self):
        cw22id = self.cw22id
        cw22root_path = self.cw22root_path
        base_filename = self.get_base_filename_by_id(cw22id, cw22root_path)

        warc_path = base_filename + '.warc.gz'
        offset_path = base_filename + '.warc.offset'

        id_parts = cw22id.split('-')
        doc = int(id_parts[len(id_parts) - 1])

        #Get html from warc using offset
        offset_length = len('{:010d}\n'.format(0, 0))
        with open (warc_path,'rb') as f_warc:
            with open (offset_path, 'r') as f_offset:
                f_offset.seek(int(doc) * int(offset_length))
                start_bytes = int (f_offset.read (offset_length).strip())
                end_bytes =   int (f_offset.read (offset_length).strip())
                f_warc.seek(start_bytes)
                record = f_warc.read(end_bytes - start_bytes)
                record = gzip.decompress(record).decode('utf-8')

                #Remove the WARC header to get the htmlStr
                warc_header = ''
                for line in record.splitlines():
                    warc_header += line
                    warc_header += '\r\n'
                    if len(line.strip()) == 0:
                        break
                record = record[len(warc_header):]

                return record
    


    def get_json_record(self, record_type):
        cw22id = self.cw22id
        cw22root_path = self.cw22root_path
        base_filename = self.get_base_filename_by_id(cw22id, cw22root_path, file_type=record_type)
        json_path = base_filename + '.json.gz'
        offset_path = base_filename + '.offset'

        id_parts = cw22id.split('-')
        doc = int(id_parts[len(id_parts) - 1])

        offset_length = len('{:010d}\n'.format(0, 0))
        with open (json_path,'rb') as f_json:
            with open (offset_path, 'r') as f_offset:
                f_offset.seek(int(doc) * int(offset_length))
                start_bytes = int (f_offset.read (offset_length).strip())
                end_bytes =   int (f_offset.read (offset_length).strip())
                f_json.seek(start_bytes)
                record = f_json.read(end_bytes - start_bytes)
                record = gzip.decompress(record).decode('utf-8')
                return record


    def get_clean_text(self):
        record = self.get_json_record('txt')
        return record

    def get_inlinks(self):
        record = self.get_json_record('inlink')
        return record

    def get_outlinks(self):
        record = self.get_json_record('outlink')
        return record


    def get_marcoweb_clean_text(self):
            record = self.get_marcoweb_json_record('txt')
            return record
        
    def get_marcoweb_json_record(self, record_type):
        cw22id = self.cw22id
        cw22root_path = self.cw22root_path

        def get_base_filename():
            html_path = cw22root_path + os.sep 
            id_parts = cw22id.split('-') # "clueweb22-en-{jjsongz_id}-{ddoc_id}"
            language = id_parts[1]
            base_path = html_path + os.sep + language + os.sep 
            base_filename = base_path + id_parts[1] + '-' + id_parts[2]
            return base_filename

        # custom file path 
        base_filename = get_base_filename()
        json_path = base_filename + '.json.gz'
        offset_path = base_filename + '.offset'

        id_parts = cw22id.split('-')
        doc = int(id_parts[len(id_parts) - 1])

        offset_length = len('{:010d}\n'.format(0, 0))
        with open (json_path,'rb') as f_json:
            with open (offset_path, 'r') as f_offset:
                f_offset.seek(int(doc) * int(offset_length))
                start_bytes = int (f_offset.read (offset_length).strip())
                end_bytes =   int (f_offset.read (offset_length).strip())
                f_json.seek(start_bytes)
                record = f_json.read(end_bytes - start_bytes)
                record = gzip.decompress(record).decode('utf-8')
                return record