init
Browse files- download_audio.py +8 -38
- main.sh +181 -419
- upload_audio.py +31 -0
download_audio.py
CHANGED
@@ -1,39 +1,3 @@
|
|
1 |
-
"""
|
2 |
-
enA-jaA: 718_606
|
3 |
-
# server 1
|
4 |
-
export LINE_NO_START=0
|
5 |
-
export LINE_NO_END=30000
|
6 |
-
|
7 |
-
export LINE_NO_START=30000
|
8 |
-
export LINE_NO_END=60000
|
9 |
-
|
10 |
-
export LINE_NO_START=60000
|
11 |
-
export LINE_NO_END=90000
|
12 |
-
|
13 |
-
export LINE_NO_START=90000
|
14 |
-
export LINE_NO_END=120000
|
15 |
-
|
16 |
-
export LINE_NO_START=120000
|
17 |
-
export LINE_NO_END=150000
|
18 |
-
|
19 |
-
|
20 |
-
# server 2
|
21 |
-
export LINE_NO_START=100000
|
22 |
-
export LINE_NO_END=120000
|
23 |
-
|
24 |
-
export LINE_NO_START=120000
|
25 |
-
export LINE_NO_END=140000
|
26 |
-
|
27 |
-
export LINE_NO_START=140000
|
28 |
-
export LINE_NO_END=160000
|
29 |
-
|
30 |
-
export LINE_NO_START=160000
|
31 |
-
export LINE_NO_END=180000
|
32 |
-
|
33 |
-
export LINE_NO_START=180000
|
34 |
-
export LINE_NO_END=200000
|
35 |
-
|
36 |
-
"""
|
37 |
import json
|
38 |
import os
|
39 |
import tarfile
|
@@ -97,8 +61,14 @@ def get_metadata():
|
|
97 |
meta_data_path = p_join("download", "meta", meta_data_filename)
|
98 |
if not os.path.exists(meta_data_path.replace(".gz", "")):
|
99 |
assert wget(url_metadata, output_file=meta_data_path)
|
100 |
-
|
101 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
102 |
assert len(df["direction"].unique()) == 1
|
103 |
df.pop("direction")
|
104 |
return df.sort_values(by=["line_no", "side"])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
import json
|
2 |
import os
|
3 |
import tarfile
|
|
|
61 |
meta_data_path = p_join("download", "meta", meta_data_filename)
|
62 |
if not os.path.exists(meta_data_path.replace(".gz", "")):
|
63 |
assert wget(url_metadata, output_file=meta_data_path)
|
64 |
+
if direction == "enA-jaA":
|
65 |
+
df = pd.read_csv(meta_data_path.replace(".gz", ""), sep=r'[\t\s]', header=None)
|
66 |
+
df = df[[0, 2, 3, 4, 9, 10, 11, 12]]
|
67 |
+
df.columns = ["id", "url", "duration_start", "duration_end", "laser_score", "direction", "side", "line_no"]
|
68 |
+
else:
|
69 |
+
raise NotImplementedError("")
|
70 |
+
# df = pd.read_csv(meta_data_path.replace(".gz", ""), sep='\t', header=None)[[0, 2, 4, 6, 9, 10, 11, 10]]
|
71 |
+
# df.columns = ["id", "url", "duration", "text_lid_score", "laser_score", "direction", "side", "line_no"]
|
72 |
assert len(df["direction"].unique()) == 1
|
73 |
df.pop("direction")
|
74 |
return df.sort_values(by=["line_no", "side"])
|
main.sh
CHANGED
@@ -1,419 +1,181 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
#
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
-
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
|
65 |
-
|
66 |
-
|
67 |
-
|
68 |
-
|
69 |
-
|
70 |
-
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
77 |
-
|
78 |
-
|
79 |
-
|
80 |
-
|
81 |
-
|
82 |
-
|
83 |
-
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
-
|
90 |
-
|
91 |
-
|
92 |
-
|
93 |
-
|
94 |
-
|
95 |
-
|
96 |
-
|
97 |
-
|
98 |
-
|
99 |
-
|
100 |
-
|
101 |
-
|
102 |
-
|
103 |
-
|
104 |
-
|
105 |
-
|
106 |
-
|
107 |
-
|
108 |
-
|
109 |
-
|
110 |
-
|
111 |
-
|
112 |
-
|
113 |
-
|
114 |
-
|
115 |
-
|
116 |
-
|
117 |
-
|
118 |
-
|
119 |
-
|
120 |
-
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
-
|
125 |
-
|
126 |
-
|
127 |
-
|
128 |
-
|
129 |
-
|
130 |
-
|
131 |
-
|
132 |
-
|
133 |
-
|
134 |
-
|
135 |
-
|
136 |
-
|
137 |
-
|
138 |
-
|
139 |
-
|
140 |
-
|
141 |
-
|
142 |
-
|
143 |
-
|
144 |
-
|
145 |
-
|
146 |
-
|
147 |
-
|
148 |
-
|
149 |
-
|
150 |
-
|
151 |
-
|
152 |
-
|
153 |
-
|
154 |
-
|
155 |
-
|
156 |
-
|
157 |
-
|
158 |
-
|
159 |
-
|
160 |
-
|
161 |
-
|
162 |
-
|
163 |
-
|
164 |
-
|
165 |
-
|
166 |
-
|
167 |
-
|
168 |
-
|
169 |
-
|
170 |
-
|
171 |
-
|
172 |
-
|
173 |
-
|
174 |
-
|
175 |
-
|
176 |
-
|
177 |
-
|
178 |
-
|
179 |
-
|
180 |
-
|
181 |
-
|
182 |
-
lookup.erase(found);
|
183 |
-
}
|
184 |
-
if (lookup.empty()) return;
|
185 |
-
}
|
186 |
-
// Failed to match the lines in lookup.
|
187 |
-
util::StringStream message;
|
188 |
-
for (std::pair<const uint64_t, std::vector<const Extract*> > &entry : lookup) {
|
189 |
-
for (const Extract *ext : entry.second) {
|
190 |
-
message.clear();
|
191 |
-
message << "Hash " << ext->paragraph_digest << " did not match any line in the WET";
|
192 |
-
out.Failure(ext->original_line, message.str());
|
193 |
-
}
|
194 |
-
}
|
195 |
-
}
|
196 |
-
|
197 |
-
void MatchLines(util::TokenIter<util::SingleCharacter, false> &line, const std::vector<Extract> &extracts, Output &out) {
|
198 |
-
util::TokenIter<util::SingleCharacter, false> line_start(line);
|
199 |
-
assert(!extracts.empty());
|
200 |
-
uint64_t line_counter = 0;
|
201 |
-
std::vector<Extract>::const_iterator extract = extracts.begin();
|
202 |
-
for (; line; ++line) {
|
203 |
-
// Upstream does python strip() then skips empty lines without counting them.
|
204 |
-
util::StringPiece stripped = Strip(*line);
|
205 |
-
if (stripped.empty()) continue;
|
206 |
-
while (line_counter == extract->paragraph_number) {
|
207 |
-
if (!ProcessExtract(*extract, stripped, out)) {
|
208 |
-
// A line failed to match the expected hash. Fall back to a hash join of all lines.
|
209 |
-
FallbackHashTable(line_start, extract, extracts.end(), out);
|
210 |
-
return;
|
211 |
-
}
|
212 |
-
if (++extract == extracts.end()) {
|
213 |
-
return;
|
214 |
-
}
|
215 |
-
}
|
216 |
-
++line_counter;
|
217 |
-
}
|
218 |
-
// Paragraph number exceeds number of lines.
|
219 |
-
FallbackHashTable(line_start, extract, extracts.end(), out);
|
220 |
-
}
|
221 |
-
|
222 |
-
// Extract SHA1 from header, leave at line
|
223 |
-
util::StringPiece FindSHA1(util::TokenIter<util::SingleCharacter, false> &line) {
|
224 |
-
const util::StringPiece kBlockDigest("WARC-Block-Digest: sha1:");
|
225 |
-
// Header through SHA1
|
226 |
-
for (; ; ++line) {
|
227 |
-
UTIL_THROW_IF(!line, MatchException, "Missing end of header");
|
228 |
-
if (line->starts_with(kBlockDigest)) {
|
229 |
-
UTIL_THROW_IF((*line)[line->size() - 1] != '\r', MatchException, "Expected carriage return in WARC.");
|
230 |
-
util::StringPiece ret(line->substr(kBlockDigest.size()));
|
231 |
-
ret.remove_suffix(1);
|
232 |
-
return ret;
|
233 |
-
}
|
234 |
-
UTIL_THROW_IF(line->empty(), MatchException, "No digest");
|
235 |
-
}
|
236 |
-
}
|
237 |
-
|
238 |
-
// The WARC reader calls this for every document in the WARC.
|
239 |
-
class DocumentCallback {
|
240 |
-
public:
|
241 |
-
DocumentCallback(Retrieve &retrieve, Output &out) : retrieve_(retrieve), out_(out) {}
|
242 |
-
|
243 |
-
// Return true if there's more documents to get from the same WARC.
|
244 |
-
bool operator()(const std::string &document) {
|
245 |
-
util::TokenIter<util::SingleCharacter, false> line(document, '\n');
|
246 |
-
UTIL_THROW_IF(!line, MatchException, "Blank document");
|
247 |
-
UTIL_THROW_IF(*line != "WARC/1.0\r", MatchException, "Expected WARC/1.0 header but got `" << *line << '\'');
|
248 |
-
UTIL_THROW_IF(!++line, MatchException, "Nothing after WARC/1.0 header");
|
249 |
-
if (*line == "WARC-Type: warcinfo\r") {
|
250 |
-
return true;
|
251 |
-
}
|
252 |
-
util::StringPiece sha1 = FindSHA1(line);
|
253 |
-
Retrieve::Iterator it = retrieve_.Lookup(sha1);
|
254 |
-
if (it == retrieve_.end()) return true;
|
255 |
-
const std::vector<Extract> &extracts = it->second;
|
256 |
-
assert(!extracts.empty());
|
257 |
-
// Consume rest of the header.
|
258 |
-
for (++line; ; ++line) {
|
259 |
-
UTIL_THROW_IF(!line, MatchException, "Missing end of header");
|
260 |
-
if (line->size() == 1 && (*line)[0] == '\r') break;
|
261 |
-
}
|
262 |
-
++line; // Skip blank.
|
263 |
-
MatchLines(line, extracts, out_);
|
264 |
-
retrieve_.Erase(it);
|
265 |
-
return !retrieve_.Empty();
|
266 |
-
}
|
267 |
-
|
268 |
-
private:
|
269 |
-
Retrieve &retrieve_;
|
270 |
-
Output &out_;
|
271 |
-
};
|
272 |
-
|
273 |
-
class CurlCallback {
|
274 |
-
public:
|
275 |
-
CurlCallback(Retrieve &retrieve, Output &out) : document_(retrieve, out), want_more_(true) {}
|
276 |
-
|
277 |
-
size_t operator()(void *buffer, size_t length) {
|
278 |
-
// As a C library, curl can't handle exceptions thrown by the callback.
|
279 |
-
try {
|
280 |
-
if (!warc_.GiveBytes(static_cast<const char*>(buffer), length, document_)) {
|
281 |
-
// Hang up early if all sentences from the WARC are complete.
|
282 |
-
want_more_ = false;
|
283 |
-
return 0;
|
284 |
-
}
|
285 |
-
} catch (...) {
|
286 |
-
exception_ = std::current_exception();
|
287 |
-
return 0;
|
288 |
-
}
|
289 |
-
return length;
|
290 |
-
}
|
291 |
-
|
292 |
-
bool CheckStatus() {
|
293 |
-
if (exception_) {
|
294 |
-
std::exception_ptr moved(std::move(exception_));
|
295 |
-
std::rethrow_exception(moved);
|
296 |
-
}
|
297 |
-
return want_more_;
|
298 |
-
}
|
299 |
-
private:
|
300 |
-
preprocess::WARCStream warc_;
|
301 |
-
DocumentCallback document_;
|
302 |
-
std::exception_ptr exception_;
|
303 |
-
bool want_more_;
|
304 |
-
};
|
305 |
-
|
306 |
-
class CurlWrap {
|
307 |
-
public:
|
308 |
-
CurlWrap() : curl_(curl_easy_init()) {
|
309 |
-
UTIL_THROW_IF(!curl_, MatchException, "Failed to initialize CURL");
|
310 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_ERRORBUFFER, error_buffer_), MatchException, "CURL Setting error buffer failed");
|
311 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_FOLLOWLOCATION, 1L), MatchException, "CURL Setting follow location failed " << error_buffer_);
|
312 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_WRITEFUNCTION, Incoming), MatchException, "CURL Setting function failed " << error_buffer_);
|
313 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_USERAGENT, "wet lines extraction"), MatchException, "CURL User Agent setting failed " << error_buffer_);
|
314 |
-
// TODO make timeouts configurable
|
315 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_TIMEOUT, 60L), MatchException, "CURL timeout setting failed " << error_buffer_);
|
316 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_LOW_SPEED_LIMIT, 1048576L), MatchException, "CURL low setting low speed failed " << error_buffer_);
|
317 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_LOW_SPEED_TIME, 5L), MatchException, "CURL low setting low speed time failed " << error_buffer_);
|
318 |
-
}
|
319 |
-
|
320 |
-
~CurlWrap() {
|
321 |
-
curl_easy_cleanup(curl_);
|
322 |
-
}
|
323 |
-
|
324 |
-
void Download(const char *url, CurlCallback &callback) {
|
325 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_URL, url), MatchException, "CURL Could not set URL " << error_buffer_);
|
326 |
-
UTIL_THROW_IF(CURLE_OK != curl_easy_setopt(curl_, CURLOPT_WRITEDATA, &callback), MatchException, "CURL Could not set callback " << error_buffer_);
|
327 |
-
CURLcode performed = curl_easy_perform(curl_);
|
328 |
-
// Throw any exceptions gathered during execution.
|
329 |
-
if (!callback.CheckStatus()) {
|
330 |
-
// If the code got everything it wanted then hung up, don't worry about CURL status.
|
331 |
-
return;
|
332 |
-
}
|
333 |
-
UTIL_THROW_IF(CURLE_OK != performed, MatchException, "CURL perform failed " << error_buffer_);
|
334 |
-
}
|
335 |
-
|
336 |
-
private:
|
337 |
-
static size_t Incoming(void *buffer, size_t /* one */, size_t nmemb, void *ptr) {
|
338 |
-
return (*static_cast<CurlCallback*>(ptr))(buffer, nmemb);
|
339 |
-
}
|
340 |
-
|
341 |
-
CURL *curl_;
|
342 |
-
|
343 |
-
char error_buffer_[CURL_ERROR_SIZE];
|
344 |
-
};
|
345 |
-
|
346 |
-
void ParseLine(util::StringPiece line, util::StringPiece &wet_path, util::StringPiece &sha1, Extract &extract) {
|
347 |
-
util::TokenIter<util::SingleCharacter, false> spaces(line, ' ');
|
348 |
-
UTIL_THROW_IF2(!spaces, "Metadata missing");
|
349 |
-
wet_path = *spaces;
|
350 |
-
UTIL_THROW_IF2(!++spaces, "Metadata missing sha1");
|
351 |
-
UTIL_THROW_IF2(!spaces->starts_with("sha1:"), "Expected hash starting with sha1");
|
352 |
-
sha1 = *spaces;
|
353 |
-
sha1.remove_prefix(5);
|
354 |
-
UTIL_THROW_IF2(!++spaces, "Metadata missing URL");
|
355 |
-
UTIL_THROW_IF2(!++spaces, "Metadata missing line");
|
356 |
-
std::from_chars_result r = std::from_chars(spaces->data(), spaces->data() + spaces->size(), extract.paragraph_number, 10);
|
357 |
-
UTIL_THROW_IF2(r.ec != std::errc(), "Error in number " << *spaces);
|
358 |
-
UTIL_THROW_IF2(r.ptr != spaces->end(), "Did not consume full number " << *spaces);
|
359 |
-
UTIL_THROW_IF2(!++spaces, "Metadata missing paragraph digest");
|
360 |
-
r = std::from_chars(spaces->data(), spaces->end(), extract.paragraph_digest);
|
361 |
-
UTIL_THROW_IF2(r.ec != std::errc(), "Error in number " << *spaces);
|
362 |
-
UTIL_THROW_IF2(r.ptr != spaces->end(), "Did not consume full number " << *spaces);
|
363 |
-
}
|
364 |
-
|
365 |
-
void RunWARC(const char *url, CurlWrap &curl, Retrieve &retrieve, Output &out) {
|
366 |
-
try {
|
367 |
-
CurlCallback callback(retrieve, out);
|
368 |
-
curl.Download(url, callback);
|
369 |
-
} catch (const util::Exception &e) {
|
370 |
-
for (Retrieve::Iterator i = retrieve.begin(); i != retrieve.end(); ++i) {
|
371 |
-
for (const Extract &extract : i->second) {
|
372 |
-
out.Failure(extract.original_line, e.what());
|
373 |
-
}
|
374 |
-
}
|
375 |
-
return;
|
376 |
-
}
|
377 |
-
for (Retrieve::Iterator i = retrieve.begin(); i != retrieve.end(); ++i) {
|
378 |
-
for (const Extract &extract : i->second) {
|
379 |
-
out.Failure(extract.original_line, "No error but unmatched");
|
380 |
-
}
|
381 |
-
}
|
382 |
-
out.Flush();
|
383 |
-
}
|
384 |
-
|
385 |
-
void ProcessMetadata(const util::StringPiece download_prefix, util::FilePiece &in, Output &out) {
|
386 |
-
Retrieve retrieve;
|
387 |
-
util::Pool string_pool;
|
388 |
-
CurlWrap curl;
|
389 |
-
util::StringPiece previous_wet_path;
|
390 |
-
std::string download_path(download_prefix.data(), download_prefix.size());
|
391 |
-
for (util::StringPiece line : in) {
|
392 |
-
util::StringPiece wet_path, sha1;
|
393 |
-
Extract extract;
|
394 |
-
ParseLine(line, wet_path, sha1, extract);
|
395 |
-
if (wet_path != previous_wet_path) {
|
396 |
-
// Flush existing data.
|
397 |
-
if (!previous_wet_path.empty()) {
|
398 |
-
download_path.replace(download_prefix.size(), download_path.size() - download_prefix.size(), previous_wet_path.data(), previous_wet_path.size());
|
399 |
-
RunWARC(download_path.c_str(), curl, retrieve, out);
|
400 |
-
}
|
401 |
-
retrieve.Clear();
|
402 |
-
string_pool.FreeAll();
|
403 |
-
}
|
404 |
-
// Store the full string line for use in printing later.
|
405 |
-
extract.original_line = util::StringPiece(static_cast<const char*>(memcpy(string_pool.Allocate(line.size()), line.data(), line.size())), line.size());
|
406 |
-
previous_wet_path = util::StringPiece(extract.original_line.data() + (wet_path.data() - line.data()), wet_path.size());
|
407 |
-
retrieve.Add(sha1, extract);
|
408 |
-
}
|
409 |
-
if (!previous_wet_path.empty()) {
|
410 |
-
download_path.replace(download_prefix.size(), download_path.size() - download_prefix.size(), previous_wet_path.data(), previous_wet_path.size());
|
411 |
-
RunWARC(download_path.c_str(), curl, retrieve, out);
|
412 |
-
}
|
413 |
-
}
|
414 |
-
|
415 |
-
int main() {
|
416 |
-
util::FilePiece in(0);
|
417 |
-
Output out;
|
418 |
-
ProcessMetadata("http://data.commoncrawl.org/", in, out);
|
419 |
-
}
|
|
|
1 |
+
####################
|
2 |
+
# enA-jaA: 718_606 #
|
3 |
+
####################
|
4 |
+
export DIRECTION="enA-jaA"
|
5 |
+
|
6 |
+
# server 1
|
7 |
+
export LINE_NO_START=0
|
8 |
+
export LINE_NO_END=30000
|
9 |
+
python download_audio.py
|
10 |
+
|
11 |
+
export LINE_NO_START=30000
|
12 |
+
export LINE_NO_END=60000
|
13 |
+
python download_audio.py
|
14 |
+
|
15 |
+
export LINE_NO_START=60000
|
16 |
+
export LINE_NO_END=90000
|
17 |
+
python download_audio.py
|
18 |
+
|
19 |
+
export LINE_NO_START=90000
|
20 |
+
export LINE_NO_END=120000
|
21 |
+
python download_audio.py
|
22 |
+
|
23 |
+
export LINE_NO_START=120000
|
24 |
+
export LINE_NO_END=150000
|
25 |
+
python download_audio.py
|
26 |
+
|
27 |
+
# server 2
|
28 |
+
export LINE_NO_START=150000
|
29 |
+
export LINE_NO_END=180000
|
30 |
+
python download_audio.py
|
31 |
+
|
32 |
+
export LINE_NO_START=180000
|
33 |
+
export LINE_NO_END=210000
|
34 |
+
python download_audio.py
|
35 |
+
|
36 |
+
export LINE_NO_START=210000
|
37 |
+
export LINE_NO_END=240000
|
38 |
+
python download_audio.py
|
39 |
+
|
40 |
+
export LINE_NO_START=240000
|
41 |
+
export LINE_NO_END=270000
|
42 |
+
python download_audio.py
|
43 |
+
|
44 |
+
export LINE_NO_START=270000
|
45 |
+
export LINE_NO_END=300000
|
46 |
+
python download_audio.py
|
47 |
+
|
48 |
+
# server 3
|
49 |
+
export LINE_NO_START=300000
|
50 |
+
export LINE_NO_END=330000
|
51 |
+
python download_audio.py
|
52 |
+
|
53 |
+
export LINE_NO_START=330000
|
54 |
+
export LINE_NO_END=360000
|
55 |
+
python download_audio.py
|
56 |
+
|
57 |
+
|
58 |
+
####################
|
59 |
+
# enA-jpn: 718_606 #
|
60 |
+
####################
|
61 |
+
|
62 |
+
git clone https://github.com/kpu/preprocess
|
63 |
+
cd preprocess
|
64 |
+
git checkout wet
|
65 |
+
git submodule update --init --recursive
|
66 |
+
mkdir build
|
67 |
+
cd build
|
68 |
+
cmake ..
|
69 |
+
make -j4
|
70 |
+
alias wet_lines="${PWD}/preprocess/build/bin/wet_lines"
|
71 |
+
wget https://dl.fbaipublicfiles.com/seamless/data/seamless.dataset.metadata.public.enA-jpn.withduration.tsv.gz
|
72 |
+
zcat seamless.dataset.metadata.public.enA-jpn.withduration.tsv.gz | egrep ^crawl-data | tr '\t' ' ' | wet_lines
|
73 |
+
|
74 |
+
|
75 |
+
export DIRECTION="enA-jpn"
|
76 |
+
|
77 |
+
export LINE_NO_START=0
|
78 |
+
export LINE_NO_END=30000
|
79 |
+
python download_audio.py
|
80 |
+
|
81 |
+
export LINE_NO_START=30000
|
82 |
+
export LINE_NO_END=60000
|
83 |
+
python download_audio.py
|
84 |
+
|
85 |
+
export LINE_NO_START=60000
|
86 |
+
export LINE_NO_END=90000
|
87 |
+
python download_audio.py
|
88 |
+
|
89 |
+
export LINE_NO_START=90000
|
90 |
+
export LINE_NO_END=120000
|
91 |
+
python download_audio.py
|
92 |
+
|
93 |
+
export LINE_NO_START=120000
|
94 |
+
export LINE_NO_END=150000
|
95 |
+
python download_audio.py
|
96 |
+
|
97 |
+
# server 2
|
98 |
+
export LINE_NO_START=150000
|
99 |
+
export LINE_NO_END=180000
|
100 |
+
python download_audio.py
|
101 |
+
|
102 |
+
export LINE_NO_START=180000
|
103 |
+
export LINE_NO_END=210000
|
104 |
+
python download_audio.py
|
105 |
+
|
106 |
+
export LINE_NO_START=210000
|
107 |
+
export LINE_NO_END=240000
|
108 |
+
python download_audio.py
|
109 |
+
|
110 |
+
export LINE_NO_START=240000
|
111 |
+
export LINE_NO_END=270000
|
112 |
+
python download_audio.py
|
113 |
+
|
114 |
+
export LINE_NO_START=270000
|
115 |
+
export LINE_NO_END=300000
|
116 |
+
python download_audio.py
|
117 |
+
|
118 |
+
# server 3
|
119 |
+
export LINE_NO_START=300000
|
120 |
+
export LINE_NO_END=330000
|
121 |
+
python download_audio.py
|
122 |
+
|
123 |
+
export LINE_NO_START=330000
|
124 |
+
export LINE_NO_END=360000
|
125 |
+
python download_audio.py
|
126 |
+
|
127 |
+
|
128 |
+
|
129 |
+
|
130 |
+
export LINE_NO_START=360000
|
131 |
+
export LINE_NO_END=390000
|
132 |
+
python download_audio.py
|
133 |
+
|
134 |
+
export LINE_NO_START=390000
|
135 |
+
export LINE_NO_END=420000
|
136 |
+
python download_audio.py
|
137 |
+
|
138 |
+
export LINE_NO_START=420000
|
139 |
+
export LINE_NO_END=450000
|
140 |
+
python download_audio.py
|
141 |
+
|
142 |
+
# server 4
|
143 |
+
export LINE_NO_START=450000
|
144 |
+
export LINE_NO_END=480000
|
145 |
+
python download_audio.py
|
146 |
+
|
147 |
+
export LINE_NO_START=480000
|
148 |
+
export LINE_NO_END=510000
|
149 |
+
python download_audio.py
|
150 |
+
|
151 |
+
export LINE_NO_START=510000
|
152 |
+
export LINE_NO_END=540000
|
153 |
+
python download_audio.py
|
154 |
+
|
155 |
+
export LINE_NO_START=540000
|
156 |
+
export LINE_NO_END=570000
|
157 |
+
python download_audio.py
|
158 |
+
|
159 |
+
export LINE_NO_START=570000
|
160 |
+
export LINE_NO_END=600000
|
161 |
+
python download_audio.py
|
162 |
+
|
163 |
+
|
164 |
+
# server 5
|
165 |
+
export LINE_NO_START=600000
|
166 |
+
export LINE_NO_END=630000
|
167 |
+
python download_audio.py
|
168 |
+
|
169 |
+
export LINE_NO_START=630000
|
170 |
+
export LINE_NO_END=660000
|
171 |
+
python download_audio.py
|
172 |
+
|
173 |
+
export LINE_NO_START=660000
|
174 |
+
export LINE_NO_END=690000
|
175 |
+
python download_audio.py
|
176 |
+
|
177 |
+
export LINE_NO_START=690000
|
178 |
+
export LINE_NO_END=720000
|
179 |
+
python download_audio.py
|
180 |
+
|
181 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
upload_audio.py
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import json
|
2 |
+
import os
|
3 |
+
import tarfile
|
4 |
+
import zipfile
|
5 |
+
import gzip
|
6 |
+
import subprocess
|
7 |
+
from os.path import join as p_join
|
8 |
+
from tqdm import tqdm
|
9 |
+
from multiprocessing import Pool
|
10 |
+
from typing import Optional
|
11 |
+
|
12 |
+
import pandas as pd
|
13 |
+
|
14 |
+
from datasets import Dataset, Audio
|
15 |
+
|
16 |
+
# dataset config
|
17 |
+
direction = os.getenv("DIRECTION", "enA-jaA")
|
18 |
+
sides = set(direction.split("-"))
|
19 |
+
cache_dir_audio = p_join("download", "audio", direction)
|
20 |
+
cache_dir_feature = p_join("download", "feature", direction)
|
21 |
+
os.makedirs(cache_dir_audio, exist_ok=True)
|
22 |
+
os.makedirs(cache_dir_feature, exist_ok=True)
|
23 |
+
# processor config
|
24 |
+
n_pool = int(os.getenv("N_POOL", 8))
|
25 |
+
wget_max_retry = os.getenv("MAX_RETRY", "1")
|
26 |
+
wget_timeout = os.getenv("TIMEOUT", "20")
|
27 |
+
line_no_start = int(os.getenv("LINE_NO_START", 0))
|
28 |
+
line_no_end = int(os.getenv("LINE_NO_END", 10000))
|
29 |
+
|
30 |
+
audio_dataset = Dataset.from_dict({"audio": ["path/to/audio_1", "path/to/audio_2", ..., "path/to/audio_n"]}).cast_column("audio", Audio())
|
31 |
+
audio_dataset[0]["audio"]
|