import os, random, copy import numpy as np import torch import pandas as pd import torchaudio from tqdm.notebook import tqdm import collections, json import re, sys import os, copy from pathlib import Path from typing import Optional import whisper DEVICE = "cuda" if torch.cuda.is_available() else "cpu" model = whisper.load_model('large-v2') model.eval() data = torch.load('./train_chime4.pt') data_with_speech = [] for item in data: with torch.no_grad(): ### TO FILL BY USERS: # use utterance id (item['id']) to retrieve parallel audio paths: clean_audio_path, noisy_audio_path ### extract clean audio feats clean_audio = whisper.load_audio(clean_audio_path) # clean_audio = whisper.pad_or_trim(clean_audio) # padding to 30s clean_mel = whisper.log_mel_spectrogram(clean_audio).to(model.device) clean_audio_features = model.encoder(clean_mel.unsqueeze(0))[0] # noisy audio feats noisy_audio = whisper.load_audio(noisy_audio_path) # noisy_audio = whisper.pad_or_trim(noisy_audio) # padding to 30s noisy_mel = whisper.log_mel_spectrogram(noisy_audio).to(model.device) noisy_audio_features = model.encoder(noisy_mel.unsqueeze(0))[0] item_with_speech = {**item, 'audio_features': noisy_audio_features, 'clean_audio_features': clean_audio_features} data_with_speech.append(item_with_speech) torch.save(data_with_speech, './train_chime4_with_speech.pt')