File size: 4,168 Bytes
07119c3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
from __future__ import annotations
from dataclasses import dataclass
from typing import Generator
import numpy as np
from sentence_transformers import SentenceTransformer
from tqdm import tqdm
import torch
import argparse
from pathlib import Path
from datasets import load_dataset

parser = argparse.ArgumentParser(description="Convert datasets to embeddings")
parser.add_argument(
    "-t",
    "--target",
    type=str,
    required=True,
    choices=["data", "chunked"],
    help="target dataset, data or chunked",
)

parser.add_argument(
    "-d",
    "--debug",
    action="store_true",
    help="debug mode, use small dataset",
)
# model_name
parser.add_argument(
    "-m",
    "--model_name",
    type=str,
    required=True,
    help="huggingface model name",
)
# input_prefix
parser.add_argument(
    "-i",
    "--input_prefix",
    type=str,
    required=False,
    default="",
    help="input prefix",
)
# max_seq_length
parser.add_argument(
    "-l",
    "--max_seq_length",
    type=int,
    required=False,
    default=512,
    help="max sequence length",
)
# output_name
parser.add_argument(
    "-o",
    "--output_name",
    type=str,
    required=True,
    help="output dir",
)
args = parser.parse_args()


@dataclass
class EmbConfig:
    model_name: str
    input_prefix: str
    max_seq_length: int


args = parser.parse_args()

target_local_ds = args.target

EMB_CONFIG = EmbConfig(
    model_name=args.model_name,
    input_prefix=args.input_prefix,
    max_seq_length=args.max_seq_length,
)
embs_dir = f"embs{'_debug' if args.debug else ''}"

output_embs_path = Path("/".join([embs_dir, args.output_name, target_local_ds]))
output_embs_path.mkdir(parents=True, exist_ok=True)

print("output path:", output_embs_path)

MODEL = SentenceTransformer(EMB_CONFIG.model_name)
MODEL.max_seq_length = EMB_CONFIG.max_seq_length


def to_embs(texts: list[str], group_size=1024) -> Generator[np.ndarray, None, None]:
    group = []
    for text in texts:
        group.append(text)
        if len(group) == group_size:
            embeddings = MODEL.encode(
                group,
                normalize_embeddings=True,
                show_progress_bar=False,
            )
            yield embeddings  # type: ignore
            group = []
    if len(group) > 0:
        embeddings = MODEL.encode(
            group, normalize_embeddings=True, show_progress_bar=False
        )
        yield embeddings  # type: ignore


def _to_data_text(
    data, prefix=EMB_CONFIG.input_prefix, max_len=int(EMB_CONFIG.max_seq_length * 1.5)
):
    return (prefix + data["title"] + "\n" + data["text"])[0:max_len]


def _to_chunk_text(
    data, prefix=EMB_CONFIG.input_prefix, max_len=int(EMB_CONFIG.max_seq_length * 1.5)
):
    return (prefix + data["title"] + "\n" + data["overlap_text"] + data["text"])[
        :max_len
    ]


def ds_to_embs(
    ds,
    text_fn,
    group_size: int,
):
    texts = []
    total = len(ds)
    pbar = tqdm(total=total)
    # text は group_size 件ごとに処理する
    for i in range(0, total, group_size):
        texts = []
        for data in ds.select(range(i, min(i + group_size, total))):
            data: dict = data
            text = text_fn(data)
            texts.append(text)
        embs = []
        for group_embs in to_embs(texts):
            embs.append(group_embs)
            pbar.update(len(group_embs))
        embs = np.concatenate(embs)
        yield embs, i, pbar


if torch.cuda.is_available():
    print("use cuda")
    MODEL.to("cuda")
else:
    print("!! Warning: use cpu")

ds = load_dataset(args.target)["train"]  # type: ignore
to_text = _to_data_text if args.target == "data" else _to_chunk_text

if args.debug:
    print("debug mode")
    ds = ds.select(range(19998))  # type: ignore
    print("small dataset len:", len(ds))
    group_size = 10000
else:
    print("dataset len:", len(ds))
    group_size = 100_000

for embs, idx, pbar in ds_to_embs(ds, to_text, group_size=group_size):
    filename = f"{idx}.npz"
    filepath = output_embs_path / filename
    pbar.desc = f"saving...: {str(filepath)}"
    np.savez_compressed(filepath, embs=embs.astype(np.float16))
    pbar.desc = ""