|
from models.modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel |
|
from transformers import SpeechEncoderDecoderModel, AutoConfig, AutoFeatureExtractor, AutoTokenizer |
|
from flax.traverse_util import flatten_dict, unflatten_dict |
|
import collections |
|
|
|
model_id = "sanchit-gandhi/flax-wav2vec2-2-bart-large-cv9-baseline-50k" |
|
|
|
config = AutoConfig.from_pretrained(model_id) |
|
config.encoder.use_scan = config.decoder.use_scan = False |
|
|
|
unrolled_model = FlaxSpeechEncoderDecoderModel.from_pretrained(model_id, config=config) |
|
model = FlaxSpeechEncoderDecoderModel.from_pretrained(model_id) |
|
|
|
def scanned_to_unrolled(params): |
|
new_enc_params = collections.defaultdict(dict) |
|
|
|
for key, stacked_weights in flatten_dict(params['encoder']['encoder']['layers']['FlaxWav2Vec2EncoderLayers']).items(): |
|
for layer, weights in enumerate(stacked_weights): |
|
new_key = (str(layer),) + key |
|
new_enc_params[new_key] = weights |
|
|
|
new_enc_params = unflatten_dict({('encoder', 'layers') : unflatten_dict(new_enc_params)}) |
|
|
|
|
|
new_dec_params = collections.defaultdict(dict) |
|
|
|
for key, stacked_weights in flatten_dict(params['decoder']['model']['decoder']['layers']['FlaxBartDecoderLayers']).items(): |
|
for layer, weights in enumerate(stacked_weights): |
|
new_key = (str(layer),) + key |
|
new_dec_params[new_key] = weights |
|
|
|
new_dec_params = unflatten_dict({('model', 'decoder', 'layers') : unflatten_dict(new_dec_params)}) |
|
|
|
|
|
new_params = {'encoder': new_enc_params, 'decoder': new_dec_params} |
|
new_params = flatten_dict(new_params) |
|
|
|
|
|
for k in flatten_dict(params): |
|
if 'layers' not in k or 'adapter' in k: |
|
new_params[k] = flatten_dict(params)[k] |
|
|
|
return unflatten_dict(new_params) |
|
|
|
unrolled_model.params = scanned_to_unrolled(model.params) |
|
|
|
unrolled_model.save_pretrained("./") |
|
|
|
feature_extractor = AutoFeatureExtractor.from_pretrained(model_id) |
|
feature_extractor.save_pretrained("./") |
|
tokenizer = AutoTokenizer.from_pretrained(model_id) |
|
tokenizer.save_pretrained("./") |
|
|