|
import os |
|
import argparse |
|
|
|
import torch |
|
import json |
|
from glob import glob |
|
|
|
from pyworld import pyworld |
|
from tqdm import tqdm |
|
from scipy.io import wavfile |
|
|
|
import cluster |
|
|
|
|
|
import logging |
|
|
|
import utils |
|
|
|
logging.getLogger('numba').setLevel(logging.WARNING) |
|
|
|
import parselmouth |
|
import librosa |
|
import numpy as np |
|
|
|
sampling_rate = 44100 |
|
hop_length = 512 |
|
|
|
|
|
def get_f0(path,p_len=None, f0_up_key=0): |
|
x, sr = librosa.load(path, sr=None) |
|
assert sr == sampling_rate |
|
if p_len is None: |
|
p_len = x.shape[0]//hop_length |
|
else: |
|
assert abs(p_len-x.shape[0]//hop_length) < 3, (path, p_len, x.shape) |
|
time_step = hop_length / sampling_rate * 1000 |
|
f0_min = 50 |
|
f0_max = 1100 |
|
f0_mel_min = 1127 * np.log(1 + f0_min / 700) |
|
f0_mel_max = 1127 * np.log(1 + f0_max / 700) |
|
|
|
f0 = parselmouth.Sound(x, sampling_rate).to_pitch_ac( |
|
time_step=time_step / 1000, voicing_threshold=0.6, |
|
pitch_floor=f0_min, pitch_ceiling=f0_max).selected_array['frequency'] |
|
|
|
pad_size=(p_len - len(f0) + 1) // 2 |
|
if(pad_size>0 or p_len - len(f0) - pad_size>0): |
|
f0 = np.pad(f0,[[pad_size,p_len - len(f0) - pad_size]], mode='constant') |
|
|
|
f0bak = f0.copy() |
|
f0 *= pow(2, f0_up_key / 12) |
|
f0_mel = 1127 * np.log(1 + f0 / 700) |
|
f0_mel[f0_mel > 0] = (f0_mel[f0_mel > 0] - f0_mel_min) * 254 / (f0_mel_max - f0_mel_min) + 1 |
|
f0_mel[f0_mel <= 1] = 1 |
|
f0_mel[f0_mel > 255] = 255 |
|
f0_coarse = np.rint(f0_mel).astype(np.int) |
|
return f0_coarse, f0bak |
|
|
|
def resize2d(x, target_len): |
|
source = np.array(x) |
|
source[source<0.001] = np.nan |
|
target = np.interp(np.arange(0, len(source)*target_len, len(source))/ target_len, np.arange(0, len(source)), source) |
|
res = np.nan_to_num(target) |
|
return res |
|
|
|
def compute_f0(path, c_len=None): |
|
x, sr = librosa.load(path, sr=None) |
|
assert sr == sampling_rate |
|
if c_len is None: |
|
c_len = x.shape[0]//hop_length |
|
|
|
f0, t = pyworld.dio( |
|
x.astype(np.double), |
|
fs=sr, |
|
f0_ceil=800, |
|
frame_period=1000 * hop_length / sr, |
|
) |
|
f0 = pyworld.stonemask(x.astype(np.double), f0, t, sampling_rate) |
|
for index, pitch in enumerate(f0): |
|
f0[index] = round(pitch, 1) |
|
assert abs(c_len - x.shape[0]//hop_length) < 3, (c_len, f0.shape) |
|
|
|
return None, resize2d(f0, c_len) |
|
|
|
|
|
def process(filename): |
|
print(filename) |
|
|
|
f0path = filename+".f0.npy" |
|
if not os.path.exists(f0path): |
|
cf0, f0 = compute_f0(filename) |
|
np.save(f0path, f0) |
|
else: |
|
f0 = np.load(f0path) |
|
c_len = f0.shape[0] |
|
save_name = filename+".discrete.npy" |
|
if not os.path.exists(save_name): |
|
devive = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
|
wav, sr = librosa.load(filename+".16k.wav",sr=None) |
|
assert sr == 16000 |
|
wav = torch.from_numpy(wav).unsqueeze(0).to(devive) |
|
c = utils.get_cn_hubert_units(hmodel, wav).cpu().squeeze(0) |
|
c = utils.repeat_expand_2d(c, c_len).numpy() |
|
|
|
c = cluster.get_cluster_result(c.transpose()) |
|
np.save(save_name,c) |
|
else: |
|
c = np.load(save_name) |
|
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
parser = argparse.ArgumentParser() |
|
parser.add_argument("--in_dir", type=str, default="dataset/", help="path to input dir") |
|
args = parser.parse_args() |
|
|
|
print("Loading hubert for content...") |
|
hmodel = utils.load_cn_model(0 if torch.cuda.is_available() else None) |
|
print("Loaded hubert.") |
|
|
|
filenames = glob(f'{args.in_dir}/*/*.wav', recursive=True) |
|
filenames = [i for i in filenames if not i.endswith(".16k.wav")] |
|
|
|
for filename in tqdm(filenames): |
|
process(filename) |
|
|