xtts2-gpt / config.json
mlinmg's picture
Upload 2 files
013e081 verified
raw
history blame
1.16 kB
{
"activation_function": "gelu",
"architectures": [
"XttsGPT"
],
"attn_pdrop": 0.1,
"audio_config": {
"mel_channels": 80,
"output_sample_rate": 24000,
"sample_rate": 22050
},
"auto_map": {
"AutoConfig": "AstraMindAI/xtts2-gpt--gpt_config.XTTSGPTConfig",
"AutoModelForCausalLM": "AstraMindAI/xtts2-gpt--xtts2_gpt_modeling.XttsGPT"
},
"decoder_input_dim": 1024,
"enable_redaction": false,
"gpt_batch_size": 1,
"gpt_max_audio_tokens": 605,
"hidden_size": 1024,
"initializer_range": 0.02,
"kv_cache": true,
"layer_norm_epsilon": 1e-05,
"max_audio_tokens": 605,
"max_prompt_tokens": 70,
"max_text_tokens": 402,
"model_type": "xtts_gpt",
"n_inner": 4096,
"num_attention_heads": 16,
"num_audio_tokens": 1026,
"num_hidden_layers": 30,
"number_text_tokens": 6681,
"reorder_and_upcast_attn": false,
"scale_attn_by_inverse_layer_idx": false,
"start_audio_token": 1024,
"start_text_token": null,
"stop_audio_token": 1025,
"stop_text_token": null,
"transformers_version": "4.46.0",
"use_masking_gt_prompt_approach": true,
"use_perceiver_resampler": true,
"vocab_size": 6681
}