Upload folder using huggingface_hub
Browse files- README.md +43 -0
- config.json +50 -0
- model.safetensors +3 -0
- model.safetensors.index.json +897 -0
- tokens.txt +360 -0
- vocoder_config.yaml +39 -0
- vocos.safetensors +3 -0
README.md
ADDED
|
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
tags:
|
| 5 |
+
- tts
|
| 6 |
+
- text-to-speech
|
| 7 |
+
- safetensors
|
| 8 |
+
- cake
|
| 9 |
+
license: apache-2.0
|
| 10 |
+
base_model: YatharthS/LuxTTS
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
# LuxTTS (Safetensors / FP16)
|
| 14 |
+
|
| 15 |
+
This is a converted version of [YatharthS/LuxTTS](https://huggingface.co/YatharthS/LuxTTS), a flow-matching based text-to-speech model. All credit for the original model, training, and research goes to the original authors.
|
| 16 |
+
|
| 17 |
+
## What changed
|
| 18 |
+
|
| 19 |
+
The original PyTorch checkpoint (`model.pt` and `vocoder/vocos.bin`) has been converted to **safetensors** format in **float16** precision for use with [Cake](https://github.com/evilsocket/cake). The conversion applies the following transformations:
|
| 20 |
+
|
| 21 |
+
- **Format**: `.pt` / `.bin` → `.safetensors` (safer, faster loading, memory-mappable).
|
| 22 |
+
- **Precision**: FP32 → FP16, reducing total size from ~530 MB to ~266 MB.
|
| 23 |
+
- **Key remapping**: The nested `fm_decoder.encoders.{stack}.layers.{layer}` hierarchy is flattened to `fm_decoder.layers.{flat_index}` using the stack sizes `[2, 2, 4, 4, 4]` (16 layers total). Similarly, `text_encoder.encoders.0.layers` is flattened to `text_encoder.layers`. Per-stack components (`time_emb`, `downsample`, `out_combiner`) are reorganized under `fm_decoder.stack_time_emb`, `fm_decoder.downsample`, and `fm_decoder.out_combiner` respectively.
|
| 24 |
+
- **Config**: `architectures` field and feature extraction parameters (`n_fft`, `hop_length`, `n_mels`, `sample_rate`) are added to `config.json`.
|
| 25 |
+
|
| 26 |
+
No weights were retrained or fine-tuned — this is a lossless format conversion (modulo FP32→FP16 quantization).
|
| 27 |
+
|
| 28 |
+
## Model details
|
| 29 |
+
|
| 30 |
+
| Component | File | Size |
|
| 31 |
+
|---|---|---|
|
| 32 |
+
| Main model (flow-matching decoder + text encoder) | `model.safetensors` | 235 MB |
|
| 33 |
+
| Vocoder (Vocos) | `vocos.safetensors` | 31 MB |
|
| 34 |
+
|
| 35 |
+
- **Architecture**: Flow-matching TTS with conformer-based decoder (16 layers across 5 stacks) and 4-layer text encoder
|
| 36 |
+
- **Vocoder**: Vocos (iSTFT-based, 8 layers, 512 dim)
|
| 37 |
+
- **Sample rate**: 24 kHz (with 48 kHz upsampler head)
|
| 38 |
+
- **Vocabulary**: 360 tokens (characters + punctuation)
|
| 39 |
+
|
| 40 |
+
## Original project
|
| 41 |
+
|
| 42 |
+
- **Model**: [YatharthS/LuxTTS](https://huggingface.co/YatharthS/LuxTTS)
|
| 43 |
+
- **License**: Apache 2.0
|
config.json
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model": {
|
| 3 |
+
"fm_decoder_downsampling_factor": [
|
| 4 |
+
1,
|
| 5 |
+
2,
|
| 6 |
+
4,
|
| 7 |
+
2,
|
| 8 |
+
1
|
| 9 |
+
],
|
| 10 |
+
"fm_decoder_num_layers": [
|
| 11 |
+
2,
|
| 12 |
+
2,
|
| 13 |
+
4,
|
| 14 |
+
4,
|
| 15 |
+
4
|
| 16 |
+
],
|
| 17 |
+
"fm_decoder_cnn_module_kernel": [
|
| 18 |
+
31,
|
| 19 |
+
15,
|
| 20 |
+
7,
|
| 21 |
+
15,
|
| 22 |
+
31
|
| 23 |
+
],
|
| 24 |
+
"fm_decoder_feedforward_dim": 1536,
|
| 25 |
+
"fm_decoder_num_heads": 4,
|
| 26 |
+
"fm_decoder_dim": 512,
|
| 27 |
+
"text_encoder_num_layers": 4,
|
| 28 |
+
"text_encoder_feedforward_dim": 512,
|
| 29 |
+
"text_encoder_cnn_module_kernel": 9,
|
| 30 |
+
"text_encoder_num_heads": 4,
|
| 31 |
+
"text_encoder_dim": 192,
|
| 32 |
+
"query_head_dim": 32,
|
| 33 |
+
"value_head_dim": 12,
|
| 34 |
+
"pos_head_dim": 4,
|
| 35 |
+
"pos_dim": 48,
|
| 36 |
+
"time_embed_dim": 192,
|
| 37 |
+
"text_embed_dim": 192,
|
| 38 |
+
"feat_dim": 100
|
| 39 |
+
},
|
| 40 |
+
"feature": {
|
| 41 |
+
"type": "vocos",
|
| 42 |
+
"n_fft": 1024,
|
| 43 |
+
"hop_length": 256,
|
| 44 |
+
"n_mels": 100,
|
| 45 |
+
"sample_rate": 24000
|
| 46 |
+
},
|
| 47 |
+
"architectures": [
|
| 48 |
+
"LuxTTSForTextToSpeech"
|
| 49 |
+
]
|
| 50 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1478863e12be4149f73cfae88bcda62eb06bd512b964006d7aef76b3b236be06
|
| 3 |
+
size 245508488
|
model.safetensors.index.json
ADDED
|
@@ -0,0 +1,897 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"metadata": {
|
| 3 |
+
"total_size": 245508488
|
| 4 |
+
},
|
| 5 |
+
"weight_map": {
|
| 6 |
+
"embed.weight": "model.safetensors",
|
| 7 |
+
"fm_decoder.downsample.1.bias": "model.safetensors",
|
| 8 |
+
"fm_decoder.downsample.2.bias": "model.safetensors",
|
| 9 |
+
"fm_decoder.downsample.3.bias": "model.safetensors",
|
| 10 |
+
"fm_decoder.guidance_scale_embed.weight": "model.safetensors",
|
| 11 |
+
"fm_decoder.in_proj.bias": "model.safetensors",
|
| 12 |
+
"fm_decoder.in_proj.weight": "model.safetensors",
|
| 13 |
+
"fm_decoder.layers.0.bypass.bypass_scale": "model.safetensors",
|
| 14 |
+
"fm_decoder.layers.0.bypass_mid.bypass_scale": "model.safetensors",
|
| 15 |
+
"fm_decoder.layers.0.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 16 |
+
"fm_decoder.layers.0.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 17 |
+
"fm_decoder.layers.0.conv_module1.in_proj.bias": "model.safetensors",
|
| 18 |
+
"fm_decoder.layers.0.conv_module1.in_proj.weight": "model.safetensors",
|
| 19 |
+
"fm_decoder.layers.0.conv_module1.out_proj.bias": "model.safetensors",
|
| 20 |
+
"fm_decoder.layers.0.conv_module1.out_proj.weight": "model.safetensors",
|
| 21 |
+
"fm_decoder.layers.0.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 22 |
+
"fm_decoder.layers.0.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 23 |
+
"fm_decoder.layers.0.conv_module2.in_proj.bias": "model.safetensors",
|
| 24 |
+
"fm_decoder.layers.0.conv_module2.in_proj.weight": "model.safetensors",
|
| 25 |
+
"fm_decoder.layers.0.conv_module2.out_proj.bias": "model.safetensors",
|
| 26 |
+
"fm_decoder.layers.0.conv_module2.out_proj.weight": "model.safetensors",
|
| 27 |
+
"fm_decoder.layers.0.feed_forward1.in_proj.bias": "model.safetensors",
|
| 28 |
+
"fm_decoder.layers.0.feed_forward1.in_proj.weight": "model.safetensors",
|
| 29 |
+
"fm_decoder.layers.0.feed_forward1.out_proj.bias": "model.safetensors",
|
| 30 |
+
"fm_decoder.layers.0.feed_forward1.out_proj.weight": "model.safetensors",
|
| 31 |
+
"fm_decoder.layers.0.feed_forward2.in_proj.bias": "model.safetensors",
|
| 32 |
+
"fm_decoder.layers.0.feed_forward2.in_proj.weight": "model.safetensors",
|
| 33 |
+
"fm_decoder.layers.0.feed_forward2.out_proj.bias": "model.safetensors",
|
| 34 |
+
"fm_decoder.layers.0.feed_forward2.out_proj.weight": "model.safetensors",
|
| 35 |
+
"fm_decoder.layers.0.feed_forward3.in_proj.bias": "model.safetensors",
|
| 36 |
+
"fm_decoder.layers.0.feed_forward3.in_proj.weight": "model.safetensors",
|
| 37 |
+
"fm_decoder.layers.0.feed_forward3.out_proj.bias": "model.safetensors",
|
| 38 |
+
"fm_decoder.layers.0.feed_forward3.out_proj.weight": "model.safetensors",
|
| 39 |
+
"fm_decoder.layers.0.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 40 |
+
"fm_decoder.layers.0.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 41 |
+
"fm_decoder.layers.0.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 42 |
+
"fm_decoder.layers.0.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 43 |
+
"fm_decoder.layers.0.norm.bias": "model.safetensors",
|
| 44 |
+
"fm_decoder.layers.0.norm.log_scale": "model.safetensors",
|
| 45 |
+
"fm_decoder.layers.0.self_attn1.in_proj.bias": "model.safetensors",
|
| 46 |
+
"fm_decoder.layers.0.self_attn1.in_proj.weight": "model.safetensors",
|
| 47 |
+
"fm_decoder.layers.0.self_attn1.out_proj.bias": "model.safetensors",
|
| 48 |
+
"fm_decoder.layers.0.self_attn1.out_proj.weight": "model.safetensors",
|
| 49 |
+
"fm_decoder.layers.0.self_attn2.in_proj.bias": "model.safetensors",
|
| 50 |
+
"fm_decoder.layers.0.self_attn2.in_proj.weight": "model.safetensors",
|
| 51 |
+
"fm_decoder.layers.0.self_attn2.out_proj.bias": "model.safetensors",
|
| 52 |
+
"fm_decoder.layers.0.self_attn2.out_proj.weight": "model.safetensors",
|
| 53 |
+
"fm_decoder.layers.0.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 54 |
+
"fm_decoder.layers.0.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 55 |
+
"fm_decoder.layers.0.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 56 |
+
"fm_decoder.layers.1.bypass.bypass_scale": "model.safetensors",
|
| 57 |
+
"fm_decoder.layers.1.bypass_mid.bypass_scale": "model.safetensors",
|
| 58 |
+
"fm_decoder.layers.1.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 59 |
+
"fm_decoder.layers.1.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 60 |
+
"fm_decoder.layers.1.conv_module1.in_proj.bias": "model.safetensors",
|
| 61 |
+
"fm_decoder.layers.1.conv_module1.in_proj.weight": "model.safetensors",
|
| 62 |
+
"fm_decoder.layers.1.conv_module1.out_proj.bias": "model.safetensors",
|
| 63 |
+
"fm_decoder.layers.1.conv_module1.out_proj.weight": "model.safetensors",
|
| 64 |
+
"fm_decoder.layers.1.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 65 |
+
"fm_decoder.layers.1.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 66 |
+
"fm_decoder.layers.1.conv_module2.in_proj.bias": "model.safetensors",
|
| 67 |
+
"fm_decoder.layers.1.conv_module2.in_proj.weight": "model.safetensors",
|
| 68 |
+
"fm_decoder.layers.1.conv_module2.out_proj.bias": "model.safetensors",
|
| 69 |
+
"fm_decoder.layers.1.conv_module2.out_proj.weight": "model.safetensors",
|
| 70 |
+
"fm_decoder.layers.1.feed_forward1.in_proj.bias": "model.safetensors",
|
| 71 |
+
"fm_decoder.layers.1.feed_forward1.in_proj.weight": "model.safetensors",
|
| 72 |
+
"fm_decoder.layers.1.feed_forward1.out_proj.bias": "model.safetensors",
|
| 73 |
+
"fm_decoder.layers.1.feed_forward1.out_proj.weight": "model.safetensors",
|
| 74 |
+
"fm_decoder.layers.1.feed_forward2.in_proj.bias": "model.safetensors",
|
| 75 |
+
"fm_decoder.layers.1.feed_forward2.in_proj.weight": "model.safetensors",
|
| 76 |
+
"fm_decoder.layers.1.feed_forward2.out_proj.bias": "model.safetensors",
|
| 77 |
+
"fm_decoder.layers.1.feed_forward2.out_proj.weight": "model.safetensors",
|
| 78 |
+
"fm_decoder.layers.1.feed_forward3.in_proj.bias": "model.safetensors",
|
| 79 |
+
"fm_decoder.layers.1.feed_forward3.in_proj.weight": "model.safetensors",
|
| 80 |
+
"fm_decoder.layers.1.feed_forward3.out_proj.bias": "model.safetensors",
|
| 81 |
+
"fm_decoder.layers.1.feed_forward3.out_proj.weight": "model.safetensors",
|
| 82 |
+
"fm_decoder.layers.1.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 83 |
+
"fm_decoder.layers.1.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 84 |
+
"fm_decoder.layers.1.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 85 |
+
"fm_decoder.layers.1.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 86 |
+
"fm_decoder.layers.1.norm.bias": "model.safetensors",
|
| 87 |
+
"fm_decoder.layers.1.norm.log_scale": "model.safetensors",
|
| 88 |
+
"fm_decoder.layers.1.self_attn1.in_proj.bias": "model.safetensors",
|
| 89 |
+
"fm_decoder.layers.1.self_attn1.in_proj.weight": "model.safetensors",
|
| 90 |
+
"fm_decoder.layers.1.self_attn1.out_proj.bias": "model.safetensors",
|
| 91 |
+
"fm_decoder.layers.1.self_attn1.out_proj.weight": "model.safetensors",
|
| 92 |
+
"fm_decoder.layers.1.self_attn2.in_proj.bias": "model.safetensors",
|
| 93 |
+
"fm_decoder.layers.1.self_attn2.in_proj.weight": "model.safetensors",
|
| 94 |
+
"fm_decoder.layers.1.self_attn2.out_proj.bias": "model.safetensors",
|
| 95 |
+
"fm_decoder.layers.1.self_attn2.out_proj.weight": "model.safetensors",
|
| 96 |
+
"fm_decoder.layers.1.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 97 |
+
"fm_decoder.layers.1.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 98 |
+
"fm_decoder.layers.1.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 99 |
+
"fm_decoder.layers.10.bypass.bypass_scale": "model.safetensors",
|
| 100 |
+
"fm_decoder.layers.10.bypass_mid.bypass_scale": "model.safetensors",
|
| 101 |
+
"fm_decoder.layers.10.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 102 |
+
"fm_decoder.layers.10.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 103 |
+
"fm_decoder.layers.10.conv_module1.in_proj.bias": "model.safetensors",
|
| 104 |
+
"fm_decoder.layers.10.conv_module1.in_proj.weight": "model.safetensors",
|
| 105 |
+
"fm_decoder.layers.10.conv_module1.out_proj.bias": "model.safetensors",
|
| 106 |
+
"fm_decoder.layers.10.conv_module1.out_proj.weight": "model.safetensors",
|
| 107 |
+
"fm_decoder.layers.10.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 108 |
+
"fm_decoder.layers.10.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 109 |
+
"fm_decoder.layers.10.conv_module2.in_proj.bias": "model.safetensors",
|
| 110 |
+
"fm_decoder.layers.10.conv_module2.in_proj.weight": "model.safetensors",
|
| 111 |
+
"fm_decoder.layers.10.conv_module2.out_proj.bias": "model.safetensors",
|
| 112 |
+
"fm_decoder.layers.10.conv_module2.out_proj.weight": "model.safetensors",
|
| 113 |
+
"fm_decoder.layers.10.feed_forward1.in_proj.bias": "model.safetensors",
|
| 114 |
+
"fm_decoder.layers.10.feed_forward1.in_proj.weight": "model.safetensors",
|
| 115 |
+
"fm_decoder.layers.10.feed_forward1.out_proj.bias": "model.safetensors",
|
| 116 |
+
"fm_decoder.layers.10.feed_forward1.out_proj.weight": "model.safetensors",
|
| 117 |
+
"fm_decoder.layers.10.feed_forward2.in_proj.bias": "model.safetensors",
|
| 118 |
+
"fm_decoder.layers.10.feed_forward2.in_proj.weight": "model.safetensors",
|
| 119 |
+
"fm_decoder.layers.10.feed_forward2.out_proj.bias": "model.safetensors",
|
| 120 |
+
"fm_decoder.layers.10.feed_forward2.out_proj.weight": "model.safetensors",
|
| 121 |
+
"fm_decoder.layers.10.feed_forward3.in_proj.bias": "model.safetensors",
|
| 122 |
+
"fm_decoder.layers.10.feed_forward3.in_proj.weight": "model.safetensors",
|
| 123 |
+
"fm_decoder.layers.10.feed_forward3.out_proj.bias": "model.safetensors",
|
| 124 |
+
"fm_decoder.layers.10.feed_forward3.out_proj.weight": "model.safetensors",
|
| 125 |
+
"fm_decoder.layers.10.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 126 |
+
"fm_decoder.layers.10.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 127 |
+
"fm_decoder.layers.10.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 128 |
+
"fm_decoder.layers.10.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 129 |
+
"fm_decoder.layers.10.norm.bias": "model.safetensors",
|
| 130 |
+
"fm_decoder.layers.10.norm.log_scale": "model.safetensors",
|
| 131 |
+
"fm_decoder.layers.10.self_attn1.in_proj.bias": "model.safetensors",
|
| 132 |
+
"fm_decoder.layers.10.self_attn1.in_proj.weight": "model.safetensors",
|
| 133 |
+
"fm_decoder.layers.10.self_attn1.out_proj.bias": "model.safetensors",
|
| 134 |
+
"fm_decoder.layers.10.self_attn1.out_proj.weight": "model.safetensors",
|
| 135 |
+
"fm_decoder.layers.10.self_attn2.in_proj.bias": "model.safetensors",
|
| 136 |
+
"fm_decoder.layers.10.self_attn2.in_proj.weight": "model.safetensors",
|
| 137 |
+
"fm_decoder.layers.10.self_attn2.out_proj.bias": "model.safetensors",
|
| 138 |
+
"fm_decoder.layers.10.self_attn2.out_proj.weight": "model.safetensors",
|
| 139 |
+
"fm_decoder.layers.10.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 140 |
+
"fm_decoder.layers.10.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 141 |
+
"fm_decoder.layers.10.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 142 |
+
"fm_decoder.layers.11.bypass.bypass_scale": "model.safetensors",
|
| 143 |
+
"fm_decoder.layers.11.bypass_mid.bypass_scale": "model.safetensors",
|
| 144 |
+
"fm_decoder.layers.11.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 145 |
+
"fm_decoder.layers.11.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 146 |
+
"fm_decoder.layers.11.conv_module1.in_proj.bias": "model.safetensors",
|
| 147 |
+
"fm_decoder.layers.11.conv_module1.in_proj.weight": "model.safetensors",
|
| 148 |
+
"fm_decoder.layers.11.conv_module1.out_proj.bias": "model.safetensors",
|
| 149 |
+
"fm_decoder.layers.11.conv_module1.out_proj.weight": "model.safetensors",
|
| 150 |
+
"fm_decoder.layers.11.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 151 |
+
"fm_decoder.layers.11.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 152 |
+
"fm_decoder.layers.11.conv_module2.in_proj.bias": "model.safetensors",
|
| 153 |
+
"fm_decoder.layers.11.conv_module2.in_proj.weight": "model.safetensors",
|
| 154 |
+
"fm_decoder.layers.11.conv_module2.out_proj.bias": "model.safetensors",
|
| 155 |
+
"fm_decoder.layers.11.conv_module2.out_proj.weight": "model.safetensors",
|
| 156 |
+
"fm_decoder.layers.11.feed_forward1.in_proj.bias": "model.safetensors",
|
| 157 |
+
"fm_decoder.layers.11.feed_forward1.in_proj.weight": "model.safetensors",
|
| 158 |
+
"fm_decoder.layers.11.feed_forward1.out_proj.bias": "model.safetensors",
|
| 159 |
+
"fm_decoder.layers.11.feed_forward1.out_proj.weight": "model.safetensors",
|
| 160 |
+
"fm_decoder.layers.11.feed_forward2.in_proj.bias": "model.safetensors",
|
| 161 |
+
"fm_decoder.layers.11.feed_forward2.in_proj.weight": "model.safetensors",
|
| 162 |
+
"fm_decoder.layers.11.feed_forward2.out_proj.bias": "model.safetensors",
|
| 163 |
+
"fm_decoder.layers.11.feed_forward2.out_proj.weight": "model.safetensors",
|
| 164 |
+
"fm_decoder.layers.11.feed_forward3.in_proj.bias": "model.safetensors",
|
| 165 |
+
"fm_decoder.layers.11.feed_forward3.in_proj.weight": "model.safetensors",
|
| 166 |
+
"fm_decoder.layers.11.feed_forward3.out_proj.bias": "model.safetensors",
|
| 167 |
+
"fm_decoder.layers.11.feed_forward3.out_proj.weight": "model.safetensors",
|
| 168 |
+
"fm_decoder.layers.11.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 169 |
+
"fm_decoder.layers.11.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 170 |
+
"fm_decoder.layers.11.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 171 |
+
"fm_decoder.layers.11.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 172 |
+
"fm_decoder.layers.11.norm.bias": "model.safetensors",
|
| 173 |
+
"fm_decoder.layers.11.norm.log_scale": "model.safetensors",
|
| 174 |
+
"fm_decoder.layers.11.self_attn1.in_proj.bias": "model.safetensors",
|
| 175 |
+
"fm_decoder.layers.11.self_attn1.in_proj.weight": "model.safetensors",
|
| 176 |
+
"fm_decoder.layers.11.self_attn1.out_proj.bias": "model.safetensors",
|
| 177 |
+
"fm_decoder.layers.11.self_attn1.out_proj.weight": "model.safetensors",
|
| 178 |
+
"fm_decoder.layers.11.self_attn2.in_proj.bias": "model.safetensors",
|
| 179 |
+
"fm_decoder.layers.11.self_attn2.in_proj.weight": "model.safetensors",
|
| 180 |
+
"fm_decoder.layers.11.self_attn2.out_proj.bias": "model.safetensors",
|
| 181 |
+
"fm_decoder.layers.11.self_attn2.out_proj.weight": "model.safetensors",
|
| 182 |
+
"fm_decoder.layers.11.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 183 |
+
"fm_decoder.layers.11.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 184 |
+
"fm_decoder.layers.11.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 185 |
+
"fm_decoder.layers.12.bypass.bypass_scale": "model.safetensors",
|
| 186 |
+
"fm_decoder.layers.12.bypass_mid.bypass_scale": "model.safetensors",
|
| 187 |
+
"fm_decoder.layers.12.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 188 |
+
"fm_decoder.layers.12.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 189 |
+
"fm_decoder.layers.12.conv_module1.in_proj.bias": "model.safetensors",
|
| 190 |
+
"fm_decoder.layers.12.conv_module1.in_proj.weight": "model.safetensors",
|
| 191 |
+
"fm_decoder.layers.12.conv_module1.out_proj.bias": "model.safetensors",
|
| 192 |
+
"fm_decoder.layers.12.conv_module1.out_proj.weight": "model.safetensors",
|
| 193 |
+
"fm_decoder.layers.12.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 194 |
+
"fm_decoder.layers.12.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 195 |
+
"fm_decoder.layers.12.conv_module2.in_proj.bias": "model.safetensors",
|
| 196 |
+
"fm_decoder.layers.12.conv_module2.in_proj.weight": "model.safetensors",
|
| 197 |
+
"fm_decoder.layers.12.conv_module2.out_proj.bias": "model.safetensors",
|
| 198 |
+
"fm_decoder.layers.12.conv_module2.out_proj.weight": "model.safetensors",
|
| 199 |
+
"fm_decoder.layers.12.feed_forward1.in_proj.bias": "model.safetensors",
|
| 200 |
+
"fm_decoder.layers.12.feed_forward1.in_proj.weight": "model.safetensors",
|
| 201 |
+
"fm_decoder.layers.12.feed_forward1.out_proj.bias": "model.safetensors",
|
| 202 |
+
"fm_decoder.layers.12.feed_forward1.out_proj.weight": "model.safetensors",
|
| 203 |
+
"fm_decoder.layers.12.feed_forward2.in_proj.bias": "model.safetensors",
|
| 204 |
+
"fm_decoder.layers.12.feed_forward2.in_proj.weight": "model.safetensors",
|
| 205 |
+
"fm_decoder.layers.12.feed_forward2.out_proj.bias": "model.safetensors",
|
| 206 |
+
"fm_decoder.layers.12.feed_forward2.out_proj.weight": "model.safetensors",
|
| 207 |
+
"fm_decoder.layers.12.feed_forward3.in_proj.bias": "model.safetensors",
|
| 208 |
+
"fm_decoder.layers.12.feed_forward3.in_proj.weight": "model.safetensors",
|
| 209 |
+
"fm_decoder.layers.12.feed_forward3.out_proj.bias": "model.safetensors",
|
| 210 |
+
"fm_decoder.layers.12.feed_forward3.out_proj.weight": "model.safetensors",
|
| 211 |
+
"fm_decoder.layers.12.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 212 |
+
"fm_decoder.layers.12.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 213 |
+
"fm_decoder.layers.12.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 214 |
+
"fm_decoder.layers.12.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 215 |
+
"fm_decoder.layers.12.norm.bias": "model.safetensors",
|
| 216 |
+
"fm_decoder.layers.12.norm.log_scale": "model.safetensors",
|
| 217 |
+
"fm_decoder.layers.12.self_attn1.in_proj.bias": "model.safetensors",
|
| 218 |
+
"fm_decoder.layers.12.self_attn1.in_proj.weight": "model.safetensors",
|
| 219 |
+
"fm_decoder.layers.12.self_attn1.out_proj.bias": "model.safetensors",
|
| 220 |
+
"fm_decoder.layers.12.self_attn1.out_proj.weight": "model.safetensors",
|
| 221 |
+
"fm_decoder.layers.12.self_attn2.in_proj.bias": "model.safetensors",
|
| 222 |
+
"fm_decoder.layers.12.self_attn2.in_proj.weight": "model.safetensors",
|
| 223 |
+
"fm_decoder.layers.12.self_attn2.out_proj.bias": "model.safetensors",
|
| 224 |
+
"fm_decoder.layers.12.self_attn2.out_proj.weight": "model.safetensors",
|
| 225 |
+
"fm_decoder.layers.12.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 226 |
+
"fm_decoder.layers.12.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 227 |
+
"fm_decoder.layers.12.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 228 |
+
"fm_decoder.layers.13.bypass.bypass_scale": "model.safetensors",
|
| 229 |
+
"fm_decoder.layers.13.bypass_mid.bypass_scale": "model.safetensors",
|
| 230 |
+
"fm_decoder.layers.13.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 231 |
+
"fm_decoder.layers.13.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 232 |
+
"fm_decoder.layers.13.conv_module1.in_proj.bias": "model.safetensors",
|
| 233 |
+
"fm_decoder.layers.13.conv_module1.in_proj.weight": "model.safetensors",
|
| 234 |
+
"fm_decoder.layers.13.conv_module1.out_proj.bias": "model.safetensors",
|
| 235 |
+
"fm_decoder.layers.13.conv_module1.out_proj.weight": "model.safetensors",
|
| 236 |
+
"fm_decoder.layers.13.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 237 |
+
"fm_decoder.layers.13.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 238 |
+
"fm_decoder.layers.13.conv_module2.in_proj.bias": "model.safetensors",
|
| 239 |
+
"fm_decoder.layers.13.conv_module2.in_proj.weight": "model.safetensors",
|
| 240 |
+
"fm_decoder.layers.13.conv_module2.out_proj.bias": "model.safetensors",
|
| 241 |
+
"fm_decoder.layers.13.conv_module2.out_proj.weight": "model.safetensors",
|
| 242 |
+
"fm_decoder.layers.13.feed_forward1.in_proj.bias": "model.safetensors",
|
| 243 |
+
"fm_decoder.layers.13.feed_forward1.in_proj.weight": "model.safetensors",
|
| 244 |
+
"fm_decoder.layers.13.feed_forward1.out_proj.bias": "model.safetensors",
|
| 245 |
+
"fm_decoder.layers.13.feed_forward1.out_proj.weight": "model.safetensors",
|
| 246 |
+
"fm_decoder.layers.13.feed_forward2.in_proj.bias": "model.safetensors",
|
| 247 |
+
"fm_decoder.layers.13.feed_forward2.in_proj.weight": "model.safetensors",
|
| 248 |
+
"fm_decoder.layers.13.feed_forward2.out_proj.bias": "model.safetensors",
|
| 249 |
+
"fm_decoder.layers.13.feed_forward2.out_proj.weight": "model.safetensors",
|
| 250 |
+
"fm_decoder.layers.13.feed_forward3.in_proj.bias": "model.safetensors",
|
| 251 |
+
"fm_decoder.layers.13.feed_forward3.in_proj.weight": "model.safetensors",
|
| 252 |
+
"fm_decoder.layers.13.feed_forward3.out_proj.bias": "model.safetensors",
|
| 253 |
+
"fm_decoder.layers.13.feed_forward3.out_proj.weight": "model.safetensors",
|
| 254 |
+
"fm_decoder.layers.13.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 255 |
+
"fm_decoder.layers.13.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 256 |
+
"fm_decoder.layers.13.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 257 |
+
"fm_decoder.layers.13.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 258 |
+
"fm_decoder.layers.13.norm.bias": "model.safetensors",
|
| 259 |
+
"fm_decoder.layers.13.norm.log_scale": "model.safetensors",
|
| 260 |
+
"fm_decoder.layers.13.self_attn1.in_proj.bias": "model.safetensors",
|
| 261 |
+
"fm_decoder.layers.13.self_attn1.in_proj.weight": "model.safetensors",
|
| 262 |
+
"fm_decoder.layers.13.self_attn1.out_proj.bias": "model.safetensors",
|
| 263 |
+
"fm_decoder.layers.13.self_attn1.out_proj.weight": "model.safetensors",
|
| 264 |
+
"fm_decoder.layers.13.self_attn2.in_proj.bias": "model.safetensors",
|
| 265 |
+
"fm_decoder.layers.13.self_attn2.in_proj.weight": "model.safetensors",
|
| 266 |
+
"fm_decoder.layers.13.self_attn2.out_proj.bias": "model.safetensors",
|
| 267 |
+
"fm_decoder.layers.13.self_attn2.out_proj.weight": "model.safetensors",
|
| 268 |
+
"fm_decoder.layers.13.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 269 |
+
"fm_decoder.layers.13.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 270 |
+
"fm_decoder.layers.13.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 271 |
+
"fm_decoder.layers.14.bypass.bypass_scale": "model.safetensors",
|
| 272 |
+
"fm_decoder.layers.14.bypass_mid.bypass_scale": "model.safetensors",
|
| 273 |
+
"fm_decoder.layers.14.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 274 |
+
"fm_decoder.layers.14.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 275 |
+
"fm_decoder.layers.14.conv_module1.in_proj.bias": "model.safetensors",
|
| 276 |
+
"fm_decoder.layers.14.conv_module1.in_proj.weight": "model.safetensors",
|
| 277 |
+
"fm_decoder.layers.14.conv_module1.out_proj.bias": "model.safetensors",
|
| 278 |
+
"fm_decoder.layers.14.conv_module1.out_proj.weight": "model.safetensors",
|
| 279 |
+
"fm_decoder.layers.14.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 280 |
+
"fm_decoder.layers.14.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 281 |
+
"fm_decoder.layers.14.conv_module2.in_proj.bias": "model.safetensors",
|
| 282 |
+
"fm_decoder.layers.14.conv_module2.in_proj.weight": "model.safetensors",
|
| 283 |
+
"fm_decoder.layers.14.conv_module2.out_proj.bias": "model.safetensors",
|
| 284 |
+
"fm_decoder.layers.14.conv_module2.out_proj.weight": "model.safetensors",
|
| 285 |
+
"fm_decoder.layers.14.feed_forward1.in_proj.bias": "model.safetensors",
|
| 286 |
+
"fm_decoder.layers.14.feed_forward1.in_proj.weight": "model.safetensors",
|
| 287 |
+
"fm_decoder.layers.14.feed_forward1.out_proj.bias": "model.safetensors",
|
| 288 |
+
"fm_decoder.layers.14.feed_forward1.out_proj.weight": "model.safetensors",
|
| 289 |
+
"fm_decoder.layers.14.feed_forward2.in_proj.bias": "model.safetensors",
|
| 290 |
+
"fm_decoder.layers.14.feed_forward2.in_proj.weight": "model.safetensors",
|
| 291 |
+
"fm_decoder.layers.14.feed_forward2.out_proj.bias": "model.safetensors",
|
| 292 |
+
"fm_decoder.layers.14.feed_forward2.out_proj.weight": "model.safetensors",
|
| 293 |
+
"fm_decoder.layers.14.feed_forward3.in_proj.bias": "model.safetensors",
|
| 294 |
+
"fm_decoder.layers.14.feed_forward3.in_proj.weight": "model.safetensors",
|
| 295 |
+
"fm_decoder.layers.14.feed_forward3.out_proj.bias": "model.safetensors",
|
| 296 |
+
"fm_decoder.layers.14.feed_forward3.out_proj.weight": "model.safetensors",
|
| 297 |
+
"fm_decoder.layers.14.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 298 |
+
"fm_decoder.layers.14.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 299 |
+
"fm_decoder.layers.14.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 300 |
+
"fm_decoder.layers.14.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 301 |
+
"fm_decoder.layers.14.norm.bias": "model.safetensors",
|
| 302 |
+
"fm_decoder.layers.14.norm.log_scale": "model.safetensors",
|
| 303 |
+
"fm_decoder.layers.14.self_attn1.in_proj.bias": "model.safetensors",
|
| 304 |
+
"fm_decoder.layers.14.self_attn1.in_proj.weight": "model.safetensors",
|
| 305 |
+
"fm_decoder.layers.14.self_attn1.out_proj.bias": "model.safetensors",
|
| 306 |
+
"fm_decoder.layers.14.self_attn1.out_proj.weight": "model.safetensors",
|
| 307 |
+
"fm_decoder.layers.14.self_attn2.in_proj.bias": "model.safetensors",
|
| 308 |
+
"fm_decoder.layers.14.self_attn2.in_proj.weight": "model.safetensors",
|
| 309 |
+
"fm_decoder.layers.14.self_attn2.out_proj.bias": "model.safetensors",
|
| 310 |
+
"fm_decoder.layers.14.self_attn2.out_proj.weight": "model.safetensors",
|
| 311 |
+
"fm_decoder.layers.14.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 312 |
+
"fm_decoder.layers.14.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 313 |
+
"fm_decoder.layers.14.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 314 |
+
"fm_decoder.layers.15.bypass.bypass_scale": "model.safetensors",
|
| 315 |
+
"fm_decoder.layers.15.bypass_mid.bypass_scale": "model.safetensors",
|
| 316 |
+
"fm_decoder.layers.15.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 317 |
+
"fm_decoder.layers.15.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 318 |
+
"fm_decoder.layers.15.conv_module1.in_proj.bias": "model.safetensors",
|
| 319 |
+
"fm_decoder.layers.15.conv_module1.in_proj.weight": "model.safetensors",
|
| 320 |
+
"fm_decoder.layers.15.conv_module1.out_proj.bias": "model.safetensors",
|
| 321 |
+
"fm_decoder.layers.15.conv_module1.out_proj.weight": "model.safetensors",
|
| 322 |
+
"fm_decoder.layers.15.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 323 |
+
"fm_decoder.layers.15.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 324 |
+
"fm_decoder.layers.15.conv_module2.in_proj.bias": "model.safetensors",
|
| 325 |
+
"fm_decoder.layers.15.conv_module2.in_proj.weight": "model.safetensors",
|
| 326 |
+
"fm_decoder.layers.15.conv_module2.out_proj.bias": "model.safetensors",
|
| 327 |
+
"fm_decoder.layers.15.conv_module2.out_proj.weight": "model.safetensors",
|
| 328 |
+
"fm_decoder.layers.15.feed_forward1.in_proj.bias": "model.safetensors",
|
| 329 |
+
"fm_decoder.layers.15.feed_forward1.in_proj.weight": "model.safetensors",
|
| 330 |
+
"fm_decoder.layers.15.feed_forward1.out_proj.bias": "model.safetensors",
|
| 331 |
+
"fm_decoder.layers.15.feed_forward1.out_proj.weight": "model.safetensors",
|
| 332 |
+
"fm_decoder.layers.15.feed_forward2.in_proj.bias": "model.safetensors",
|
| 333 |
+
"fm_decoder.layers.15.feed_forward2.in_proj.weight": "model.safetensors",
|
| 334 |
+
"fm_decoder.layers.15.feed_forward2.out_proj.bias": "model.safetensors",
|
| 335 |
+
"fm_decoder.layers.15.feed_forward2.out_proj.weight": "model.safetensors",
|
| 336 |
+
"fm_decoder.layers.15.feed_forward3.in_proj.bias": "model.safetensors",
|
| 337 |
+
"fm_decoder.layers.15.feed_forward3.in_proj.weight": "model.safetensors",
|
| 338 |
+
"fm_decoder.layers.15.feed_forward3.out_proj.bias": "model.safetensors",
|
| 339 |
+
"fm_decoder.layers.15.feed_forward3.out_proj.weight": "model.safetensors",
|
| 340 |
+
"fm_decoder.layers.15.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 341 |
+
"fm_decoder.layers.15.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 342 |
+
"fm_decoder.layers.15.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 343 |
+
"fm_decoder.layers.15.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 344 |
+
"fm_decoder.layers.15.norm.bias": "model.safetensors",
|
| 345 |
+
"fm_decoder.layers.15.norm.log_scale": "model.safetensors",
|
| 346 |
+
"fm_decoder.layers.15.self_attn1.in_proj.bias": "model.safetensors",
|
| 347 |
+
"fm_decoder.layers.15.self_attn1.in_proj.weight": "model.safetensors",
|
| 348 |
+
"fm_decoder.layers.15.self_attn1.out_proj.bias": "model.safetensors",
|
| 349 |
+
"fm_decoder.layers.15.self_attn1.out_proj.weight": "model.safetensors",
|
| 350 |
+
"fm_decoder.layers.15.self_attn2.in_proj.bias": "model.safetensors",
|
| 351 |
+
"fm_decoder.layers.15.self_attn2.in_proj.weight": "model.safetensors",
|
| 352 |
+
"fm_decoder.layers.15.self_attn2.out_proj.bias": "model.safetensors",
|
| 353 |
+
"fm_decoder.layers.15.self_attn2.out_proj.weight": "model.safetensors",
|
| 354 |
+
"fm_decoder.layers.15.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 355 |
+
"fm_decoder.layers.15.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 356 |
+
"fm_decoder.layers.15.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 357 |
+
"fm_decoder.layers.2.bypass.bypass_scale": "model.safetensors",
|
| 358 |
+
"fm_decoder.layers.2.bypass_mid.bypass_scale": "model.safetensors",
|
| 359 |
+
"fm_decoder.layers.2.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 360 |
+
"fm_decoder.layers.2.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 361 |
+
"fm_decoder.layers.2.conv_module1.in_proj.bias": "model.safetensors",
|
| 362 |
+
"fm_decoder.layers.2.conv_module1.in_proj.weight": "model.safetensors",
|
| 363 |
+
"fm_decoder.layers.2.conv_module1.out_proj.bias": "model.safetensors",
|
| 364 |
+
"fm_decoder.layers.2.conv_module1.out_proj.weight": "model.safetensors",
|
| 365 |
+
"fm_decoder.layers.2.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 366 |
+
"fm_decoder.layers.2.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 367 |
+
"fm_decoder.layers.2.conv_module2.in_proj.bias": "model.safetensors",
|
| 368 |
+
"fm_decoder.layers.2.conv_module2.in_proj.weight": "model.safetensors",
|
| 369 |
+
"fm_decoder.layers.2.conv_module2.out_proj.bias": "model.safetensors",
|
| 370 |
+
"fm_decoder.layers.2.conv_module2.out_proj.weight": "model.safetensors",
|
| 371 |
+
"fm_decoder.layers.2.feed_forward1.in_proj.bias": "model.safetensors",
|
| 372 |
+
"fm_decoder.layers.2.feed_forward1.in_proj.weight": "model.safetensors",
|
| 373 |
+
"fm_decoder.layers.2.feed_forward1.out_proj.bias": "model.safetensors",
|
| 374 |
+
"fm_decoder.layers.2.feed_forward1.out_proj.weight": "model.safetensors",
|
| 375 |
+
"fm_decoder.layers.2.feed_forward2.in_proj.bias": "model.safetensors",
|
| 376 |
+
"fm_decoder.layers.2.feed_forward2.in_proj.weight": "model.safetensors",
|
| 377 |
+
"fm_decoder.layers.2.feed_forward2.out_proj.bias": "model.safetensors",
|
| 378 |
+
"fm_decoder.layers.2.feed_forward2.out_proj.weight": "model.safetensors",
|
| 379 |
+
"fm_decoder.layers.2.feed_forward3.in_proj.bias": "model.safetensors",
|
| 380 |
+
"fm_decoder.layers.2.feed_forward3.in_proj.weight": "model.safetensors",
|
| 381 |
+
"fm_decoder.layers.2.feed_forward3.out_proj.bias": "model.safetensors",
|
| 382 |
+
"fm_decoder.layers.2.feed_forward3.out_proj.weight": "model.safetensors",
|
| 383 |
+
"fm_decoder.layers.2.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 384 |
+
"fm_decoder.layers.2.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 385 |
+
"fm_decoder.layers.2.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 386 |
+
"fm_decoder.layers.2.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 387 |
+
"fm_decoder.layers.2.norm.bias": "model.safetensors",
|
| 388 |
+
"fm_decoder.layers.2.norm.log_scale": "model.safetensors",
|
| 389 |
+
"fm_decoder.layers.2.self_attn1.in_proj.bias": "model.safetensors",
|
| 390 |
+
"fm_decoder.layers.2.self_attn1.in_proj.weight": "model.safetensors",
|
| 391 |
+
"fm_decoder.layers.2.self_attn1.out_proj.bias": "model.safetensors",
|
| 392 |
+
"fm_decoder.layers.2.self_attn1.out_proj.weight": "model.safetensors",
|
| 393 |
+
"fm_decoder.layers.2.self_attn2.in_proj.bias": "model.safetensors",
|
| 394 |
+
"fm_decoder.layers.2.self_attn2.in_proj.weight": "model.safetensors",
|
| 395 |
+
"fm_decoder.layers.2.self_attn2.out_proj.bias": "model.safetensors",
|
| 396 |
+
"fm_decoder.layers.2.self_attn2.out_proj.weight": "model.safetensors",
|
| 397 |
+
"fm_decoder.layers.2.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 398 |
+
"fm_decoder.layers.2.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 399 |
+
"fm_decoder.layers.2.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 400 |
+
"fm_decoder.layers.3.bypass.bypass_scale": "model.safetensors",
|
| 401 |
+
"fm_decoder.layers.3.bypass_mid.bypass_scale": "model.safetensors",
|
| 402 |
+
"fm_decoder.layers.3.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 403 |
+
"fm_decoder.layers.3.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 404 |
+
"fm_decoder.layers.3.conv_module1.in_proj.bias": "model.safetensors",
|
| 405 |
+
"fm_decoder.layers.3.conv_module1.in_proj.weight": "model.safetensors",
|
| 406 |
+
"fm_decoder.layers.3.conv_module1.out_proj.bias": "model.safetensors",
|
| 407 |
+
"fm_decoder.layers.3.conv_module1.out_proj.weight": "model.safetensors",
|
| 408 |
+
"fm_decoder.layers.3.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 409 |
+
"fm_decoder.layers.3.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 410 |
+
"fm_decoder.layers.3.conv_module2.in_proj.bias": "model.safetensors",
|
| 411 |
+
"fm_decoder.layers.3.conv_module2.in_proj.weight": "model.safetensors",
|
| 412 |
+
"fm_decoder.layers.3.conv_module2.out_proj.bias": "model.safetensors",
|
| 413 |
+
"fm_decoder.layers.3.conv_module2.out_proj.weight": "model.safetensors",
|
| 414 |
+
"fm_decoder.layers.3.feed_forward1.in_proj.bias": "model.safetensors",
|
| 415 |
+
"fm_decoder.layers.3.feed_forward1.in_proj.weight": "model.safetensors",
|
| 416 |
+
"fm_decoder.layers.3.feed_forward1.out_proj.bias": "model.safetensors",
|
| 417 |
+
"fm_decoder.layers.3.feed_forward1.out_proj.weight": "model.safetensors",
|
| 418 |
+
"fm_decoder.layers.3.feed_forward2.in_proj.bias": "model.safetensors",
|
| 419 |
+
"fm_decoder.layers.3.feed_forward2.in_proj.weight": "model.safetensors",
|
| 420 |
+
"fm_decoder.layers.3.feed_forward2.out_proj.bias": "model.safetensors",
|
| 421 |
+
"fm_decoder.layers.3.feed_forward2.out_proj.weight": "model.safetensors",
|
| 422 |
+
"fm_decoder.layers.3.feed_forward3.in_proj.bias": "model.safetensors",
|
| 423 |
+
"fm_decoder.layers.3.feed_forward3.in_proj.weight": "model.safetensors",
|
| 424 |
+
"fm_decoder.layers.3.feed_forward3.out_proj.bias": "model.safetensors",
|
| 425 |
+
"fm_decoder.layers.3.feed_forward3.out_proj.weight": "model.safetensors",
|
| 426 |
+
"fm_decoder.layers.3.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 427 |
+
"fm_decoder.layers.3.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 428 |
+
"fm_decoder.layers.3.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 429 |
+
"fm_decoder.layers.3.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 430 |
+
"fm_decoder.layers.3.norm.bias": "model.safetensors",
|
| 431 |
+
"fm_decoder.layers.3.norm.log_scale": "model.safetensors",
|
| 432 |
+
"fm_decoder.layers.3.self_attn1.in_proj.bias": "model.safetensors",
|
| 433 |
+
"fm_decoder.layers.3.self_attn1.in_proj.weight": "model.safetensors",
|
| 434 |
+
"fm_decoder.layers.3.self_attn1.out_proj.bias": "model.safetensors",
|
| 435 |
+
"fm_decoder.layers.3.self_attn1.out_proj.weight": "model.safetensors",
|
| 436 |
+
"fm_decoder.layers.3.self_attn2.in_proj.bias": "model.safetensors",
|
| 437 |
+
"fm_decoder.layers.3.self_attn2.in_proj.weight": "model.safetensors",
|
| 438 |
+
"fm_decoder.layers.3.self_attn2.out_proj.bias": "model.safetensors",
|
| 439 |
+
"fm_decoder.layers.3.self_attn2.out_proj.weight": "model.safetensors",
|
| 440 |
+
"fm_decoder.layers.3.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 441 |
+
"fm_decoder.layers.3.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 442 |
+
"fm_decoder.layers.3.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 443 |
+
"fm_decoder.layers.4.bypass.bypass_scale": "model.safetensors",
|
| 444 |
+
"fm_decoder.layers.4.bypass_mid.bypass_scale": "model.safetensors",
|
| 445 |
+
"fm_decoder.layers.4.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 446 |
+
"fm_decoder.layers.4.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 447 |
+
"fm_decoder.layers.4.conv_module1.in_proj.bias": "model.safetensors",
|
| 448 |
+
"fm_decoder.layers.4.conv_module1.in_proj.weight": "model.safetensors",
|
| 449 |
+
"fm_decoder.layers.4.conv_module1.out_proj.bias": "model.safetensors",
|
| 450 |
+
"fm_decoder.layers.4.conv_module1.out_proj.weight": "model.safetensors",
|
| 451 |
+
"fm_decoder.layers.4.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 452 |
+
"fm_decoder.layers.4.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 453 |
+
"fm_decoder.layers.4.conv_module2.in_proj.bias": "model.safetensors",
|
| 454 |
+
"fm_decoder.layers.4.conv_module2.in_proj.weight": "model.safetensors",
|
| 455 |
+
"fm_decoder.layers.4.conv_module2.out_proj.bias": "model.safetensors",
|
| 456 |
+
"fm_decoder.layers.4.conv_module2.out_proj.weight": "model.safetensors",
|
| 457 |
+
"fm_decoder.layers.4.feed_forward1.in_proj.bias": "model.safetensors",
|
| 458 |
+
"fm_decoder.layers.4.feed_forward1.in_proj.weight": "model.safetensors",
|
| 459 |
+
"fm_decoder.layers.4.feed_forward1.out_proj.bias": "model.safetensors",
|
| 460 |
+
"fm_decoder.layers.4.feed_forward1.out_proj.weight": "model.safetensors",
|
| 461 |
+
"fm_decoder.layers.4.feed_forward2.in_proj.bias": "model.safetensors",
|
| 462 |
+
"fm_decoder.layers.4.feed_forward2.in_proj.weight": "model.safetensors",
|
| 463 |
+
"fm_decoder.layers.4.feed_forward2.out_proj.bias": "model.safetensors",
|
| 464 |
+
"fm_decoder.layers.4.feed_forward2.out_proj.weight": "model.safetensors",
|
| 465 |
+
"fm_decoder.layers.4.feed_forward3.in_proj.bias": "model.safetensors",
|
| 466 |
+
"fm_decoder.layers.4.feed_forward3.in_proj.weight": "model.safetensors",
|
| 467 |
+
"fm_decoder.layers.4.feed_forward3.out_proj.bias": "model.safetensors",
|
| 468 |
+
"fm_decoder.layers.4.feed_forward3.out_proj.weight": "model.safetensors",
|
| 469 |
+
"fm_decoder.layers.4.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 470 |
+
"fm_decoder.layers.4.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 471 |
+
"fm_decoder.layers.4.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 472 |
+
"fm_decoder.layers.4.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 473 |
+
"fm_decoder.layers.4.norm.bias": "model.safetensors",
|
| 474 |
+
"fm_decoder.layers.4.norm.log_scale": "model.safetensors",
|
| 475 |
+
"fm_decoder.layers.4.self_attn1.in_proj.bias": "model.safetensors",
|
| 476 |
+
"fm_decoder.layers.4.self_attn1.in_proj.weight": "model.safetensors",
|
| 477 |
+
"fm_decoder.layers.4.self_attn1.out_proj.bias": "model.safetensors",
|
| 478 |
+
"fm_decoder.layers.4.self_attn1.out_proj.weight": "model.safetensors",
|
| 479 |
+
"fm_decoder.layers.4.self_attn2.in_proj.bias": "model.safetensors",
|
| 480 |
+
"fm_decoder.layers.4.self_attn2.in_proj.weight": "model.safetensors",
|
| 481 |
+
"fm_decoder.layers.4.self_attn2.out_proj.bias": "model.safetensors",
|
| 482 |
+
"fm_decoder.layers.4.self_attn2.out_proj.weight": "model.safetensors",
|
| 483 |
+
"fm_decoder.layers.4.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 484 |
+
"fm_decoder.layers.4.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 485 |
+
"fm_decoder.layers.4.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 486 |
+
"fm_decoder.layers.5.bypass.bypass_scale": "model.safetensors",
|
| 487 |
+
"fm_decoder.layers.5.bypass_mid.bypass_scale": "model.safetensors",
|
| 488 |
+
"fm_decoder.layers.5.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 489 |
+
"fm_decoder.layers.5.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 490 |
+
"fm_decoder.layers.5.conv_module1.in_proj.bias": "model.safetensors",
|
| 491 |
+
"fm_decoder.layers.5.conv_module1.in_proj.weight": "model.safetensors",
|
| 492 |
+
"fm_decoder.layers.5.conv_module1.out_proj.bias": "model.safetensors",
|
| 493 |
+
"fm_decoder.layers.5.conv_module1.out_proj.weight": "model.safetensors",
|
| 494 |
+
"fm_decoder.layers.5.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 495 |
+
"fm_decoder.layers.5.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 496 |
+
"fm_decoder.layers.5.conv_module2.in_proj.bias": "model.safetensors",
|
| 497 |
+
"fm_decoder.layers.5.conv_module2.in_proj.weight": "model.safetensors",
|
| 498 |
+
"fm_decoder.layers.5.conv_module2.out_proj.bias": "model.safetensors",
|
| 499 |
+
"fm_decoder.layers.5.conv_module2.out_proj.weight": "model.safetensors",
|
| 500 |
+
"fm_decoder.layers.5.feed_forward1.in_proj.bias": "model.safetensors",
|
| 501 |
+
"fm_decoder.layers.5.feed_forward1.in_proj.weight": "model.safetensors",
|
| 502 |
+
"fm_decoder.layers.5.feed_forward1.out_proj.bias": "model.safetensors",
|
| 503 |
+
"fm_decoder.layers.5.feed_forward1.out_proj.weight": "model.safetensors",
|
| 504 |
+
"fm_decoder.layers.5.feed_forward2.in_proj.bias": "model.safetensors",
|
| 505 |
+
"fm_decoder.layers.5.feed_forward2.in_proj.weight": "model.safetensors",
|
| 506 |
+
"fm_decoder.layers.5.feed_forward2.out_proj.bias": "model.safetensors",
|
| 507 |
+
"fm_decoder.layers.5.feed_forward2.out_proj.weight": "model.safetensors",
|
| 508 |
+
"fm_decoder.layers.5.feed_forward3.in_proj.bias": "model.safetensors",
|
| 509 |
+
"fm_decoder.layers.5.feed_forward3.in_proj.weight": "model.safetensors",
|
| 510 |
+
"fm_decoder.layers.5.feed_forward3.out_proj.bias": "model.safetensors",
|
| 511 |
+
"fm_decoder.layers.5.feed_forward3.out_proj.weight": "model.safetensors",
|
| 512 |
+
"fm_decoder.layers.5.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 513 |
+
"fm_decoder.layers.5.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 514 |
+
"fm_decoder.layers.5.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 515 |
+
"fm_decoder.layers.5.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 516 |
+
"fm_decoder.layers.5.norm.bias": "model.safetensors",
|
| 517 |
+
"fm_decoder.layers.5.norm.log_scale": "model.safetensors",
|
| 518 |
+
"fm_decoder.layers.5.self_attn1.in_proj.bias": "model.safetensors",
|
| 519 |
+
"fm_decoder.layers.5.self_attn1.in_proj.weight": "model.safetensors",
|
| 520 |
+
"fm_decoder.layers.5.self_attn1.out_proj.bias": "model.safetensors",
|
| 521 |
+
"fm_decoder.layers.5.self_attn1.out_proj.weight": "model.safetensors",
|
| 522 |
+
"fm_decoder.layers.5.self_attn2.in_proj.bias": "model.safetensors",
|
| 523 |
+
"fm_decoder.layers.5.self_attn2.in_proj.weight": "model.safetensors",
|
| 524 |
+
"fm_decoder.layers.5.self_attn2.out_proj.bias": "model.safetensors",
|
| 525 |
+
"fm_decoder.layers.5.self_attn2.out_proj.weight": "model.safetensors",
|
| 526 |
+
"fm_decoder.layers.5.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 527 |
+
"fm_decoder.layers.5.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 528 |
+
"fm_decoder.layers.5.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 529 |
+
"fm_decoder.layers.6.bypass.bypass_scale": "model.safetensors",
|
| 530 |
+
"fm_decoder.layers.6.bypass_mid.bypass_scale": "model.safetensors",
|
| 531 |
+
"fm_decoder.layers.6.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 532 |
+
"fm_decoder.layers.6.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 533 |
+
"fm_decoder.layers.6.conv_module1.in_proj.bias": "model.safetensors",
|
| 534 |
+
"fm_decoder.layers.6.conv_module1.in_proj.weight": "model.safetensors",
|
| 535 |
+
"fm_decoder.layers.6.conv_module1.out_proj.bias": "model.safetensors",
|
| 536 |
+
"fm_decoder.layers.6.conv_module1.out_proj.weight": "model.safetensors",
|
| 537 |
+
"fm_decoder.layers.6.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 538 |
+
"fm_decoder.layers.6.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 539 |
+
"fm_decoder.layers.6.conv_module2.in_proj.bias": "model.safetensors",
|
| 540 |
+
"fm_decoder.layers.6.conv_module2.in_proj.weight": "model.safetensors",
|
| 541 |
+
"fm_decoder.layers.6.conv_module2.out_proj.bias": "model.safetensors",
|
| 542 |
+
"fm_decoder.layers.6.conv_module2.out_proj.weight": "model.safetensors",
|
| 543 |
+
"fm_decoder.layers.6.feed_forward1.in_proj.bias": "model.safetensors",
|
| 544 |
+
"fm_decoder.layers.6.feed_forward1.in_proj.weight": "model.safetensors",
|
| 545 |
+
"fm_decoder.layers.6.feed_forward1.out_proj.bias": "model.safetensors",
|
| 546 |
+
"fm_decoder.layers.6.feed_forward1.out_proj.weight": "model.safetensors",
|
| 547 |
+
"fm_decoder.layers.6.feed_forward2.in_proj.bias": "model.safetensors",
|
| 548 |
+
"fm_decoder.layers.6.feed_forward2.in_proj.weight": "model.safetensors",
|
| 549 |
+
"fm_decoder.layers.6.feed_forward2.out_proj.bias": "model.safetensors",
|
| 550 |
+
"fm_decoder.layers.6.feed_forward2.out_proj.weight": "model.safetensors",
|
| 551 |
+
"fm_decoder.layers.6.feed_forward3.in_proj.bias": "model.safetensors",
|
| 552 |
+
"fm_decoder.layers.6.feed_forward3.in_proj.weight": "model.safetensors",
|
| 553 |
+
"fm_decoder.layers.6.feed_forward3.out_proj.bias": "model.safetensors",
|
| 554 |
+
"fm_decoder.layers.6.feed_forward3.out_proj.weight": "model.safetensors",
|
| 555 |
+
"fm_decoder.layers.6.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 556 |
+
"fm_decoder.layers.6.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 557 |
+
"fm_decoder.layers.6.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 558 |
+
"fm_decoder.layers.6.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 559 |
+
"fm_decoder.layers.6.norm.bias": "model.safetensors",
|
| 560 |
+
"fm_decoder.layers.6.norm.log_scale": "model.safetensors",
|
| 561 |
+
"fm_decoder.layers.6.self_attn1.in_proj.bias": "model.safetensors",
|
| 562 |
+
"fm_decoder.layers.6.self_attn1.in_proj.weight": "model.safetensors",
|
| 563 |
+
"fm_decoder.layers.6.self_attn1.out_proj.bias": "model.safetensors",
|
| 564 |
+
"fm_decoder.layers.6.self_attn1.out_proj.weight": "model.safetensors",
|
| 565 |
+
"fm_decoder.layers.6.self_attn2.in_proj.bias": "model.safetensors",
|
| 566 |
+
"fm_decoder.layers.6.self_attn2.in_proj.weight": "model.safetensors",
|
| 567 |
+
"fm_decoder.layers.6.self_attn2.out_proj.bias": "model.safetensors",
|
| 568 |
+
"fm_decoder.layers.6.self_attn2.out_proj.weight": "model.safetensors",
|
| 569 |
+
"fm_decoder.layers.6.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 570 |
+
"fm_decoder.layers.6.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 571 |
+
"fm_decoder.layers.6.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 572 |
+
"fm_decoder.layers.7.bypass.bypass_scale": "model.safetensors",
|
| 573 |
+
"fm_decoder.layers.7.bypass_mid.bypass_scale": "model.safetensors",
|
| 574 |
+
"fm_decoder.layers.7.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 575 |
+
"fm_decoder.layers.7.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 576 |
+
"fm_decoder.layers.7.conv_module1.in_proj.bias": "model.safetensors",
|
| 577 |
+
"fm_decoder.layers.7.conv_module1.in_proj.weight": "model.safetensors",
|
| 578 |
+
"fm_decoder.layers.7.conv_module1.out_proj.bias": "model.safetensors",
|
| 579 |
+
"fm_decoder.layers.7.conv_module1.out_proj.weight": "model.safetensors",
|
| 580 |
+
"fm_decoder.layers.7.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 581 |
+
"fm_decoder.layers.7.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 582 |
+
"fm_decoder.layers.7.conv_module2.in_proj.bias": "model.safetensors",
|
| 583 |
+
"fm_decoder.layers.7.conv_module2.in_proj.weight": "model.safetensors",
|
| 584 |
+
"fm_decoder.layers.7.conv_module2.out_proj.bias": "model.safetensors",
|
| 585 |
+
"fm_decoder.layers.7.conv_module2.out_proj.weight": "model.safetensors",
|
| 586 |
+
"fm_decoder.layers.7.feed_forward1.in_proj.bias": "model.safetensors",
|
| 587 |
+
"fm_decoder.layers.7.feed_forward1.in_proj.weight": "model.safetensors",
|
| 588 |
+
"fm_decoder.layers.7.feed_forward1.out_proj.bias": "model.safetensors",
|
| 589 |
+
"fm_decoder.layers.7.feed_forward1.out_proj.weight": "model.safetensors",
|
| 590 |
+
"fm_decoder.layers.7.feed_forward2.in_proj.bias": "model.safetensors",
|
| 591 |
+
"fm_decoder.layers.7.feed_forward2.in_proj.weight": "model.safetensors",
|
| 592 |
+
"fm_decoder.layers.7.feed_forward2.out_proj.bias": "model.safetensors",
|
| 593 |
+
"fm_decoder.layers.7.feed_forward2.out_proj.weight": "model.safetensors",
|
| 594 |
+
"fm_decoder.layers.7.feed_forward3.in_proj.bias": "model.safetensors",
|
| 595 |
+
"fm_decoder.layers.7.feed_forward3.in_proj.weight": "model.safetensors",
|
| 596 |
+
"fm_decoder.layers.7.feed_forward3.out_proj.bias": "model.safetensors",
|
| 597 |
+
"fm_decoder.layers.7.feed_forward3.out_proj.weight": "model.safetensors",
|
| 598 |
+
"fm_decoder.layers.7.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 599 |
+
"fm_decoder.layers.7.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 600 |
+
"fm_decoder.layers.7.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 601 |
+
"fm_decoder.layers.7.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 602 |
+
"fm_decoder.layers.7.norm.bias": "model.safetensors",
|
| 603 |
+
"fm_decoder.layers.7.norm.log_scale": "model.safetensors",
|
| 604 |
+
"fm_decoder.layers.7.self_attn1.in_proj.bias": "model.safetensors",
|
| 605 |
+
"fm_decoder.layers.7.self_attn1.in_proj.weight": "model.safetensors",
|
| 606 |
+
"fm_decoder.layers.7.self_attn1.out_proj.bias": "model.safetensors",
|
| 607 |
+
"fm_decoder.layers.7.self_attn1.out_proj.weight": "model.safetensors",
|
| 608 |
+
"fm_decoder.layers.7.self_attn2.in_proj.bias": "model.safetensors",
|
| 609 |
+
"fm_decoder.layers.7.self_attn2.in_proj.weight": "model.safetensors",
|
| 610 |
+
"fm_decoder.layers.7.self_attn2.out_proj.bias": "model.safetensors",
|
| 611 |
+
"fm_decoder.layers.7.self_attn2.out_proj.weight": "model.safetensors",
|
| 612 |
+
"fm_decoder.layers.7.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 613 |
+
"fm_decoder.layers.7.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 614 |
+
"fm_decoder.layers.7.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 615 |
+
"fm_decoder.layers.8.bypass.bypass_scale": "model.safetensors",
|
| 616 |
+
"fm_decoder.layers.8.bypass_mid.bypass_scale": "model.safetensors",
|
| 617 |
+
"fm_decoder.layers.8.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 618 |
+
"fm_decoder.layers.8.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 619 |
+
"fm_decoder.layers.8.conv_module1.in_proj.bias": "model.safetensors",
|
| 620 |
+
"fm_decoder.layers.8.conv_module1.in_proj.weight": "model.safetensors",
|
| 621 |
+
"fm_decoder.layers.8.conv_module1.out_proj.bias": "model.safetensors",
|
| 622 |
+
"fm_decoder.layers.8.conv_module1.out_proj.weight": "model.safetensors",
|
| 623 |
+
"fm_decoder.layers.8.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 624 |
+
"fm_decoder.layers.8.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 625 |
+
"fm_decoder.layers.8.conv_module2.in_proj.bias": "model.safetensors",
|
| 626 |
+
"fm_decoder.layers.8.conv_module2.in_proj.weight": "model.safetensors",
|
| 627 |
+
"fm_decoder.layers.8.conv_module2.out_proj.bias": "model.safetensors",
|
| 628 |
+
"fm_decoder.layers.8.conv_module2.out_proj.weight": "model.safetensors",
|
| 629 |
+
"fm_decoder.layers.8.feed_forward1.in_proj.bias": "model.safetensors",
|
| 630 |
+
"fm_decoder.layers.8.feed_forward1.in_proj.weight": "model.safetensors",
|
| 631 |
+
"fm_decoder.layers.8.feed_forward1.out_proj.bias": "model.safetensors",
|
| 632 |
+
"fm_decoder.layers.8.feed_forward1.out_proj.weight": "model.safetensors",
|
| 633 |
+
"fm_decoder.layers.8.feed_forward2.in_proj.bias": "model.safetensors",
|
| 634 |
+
"fm_decoder.layers.8.feed_forward2.in_proj.weight": "model.safetensors",
|
| 635 |
+
"fm_decoder.layers.8.feed_forward2.out_proj.bias": "model.safetensors",
|
| 636 |
+
"fm_decoder.layers.8.feed_forward2.out_proj.weight": "model.safetensors",
|
| 637 |
+
"fm_decoder.layers.8.feed_forward3.in_proj.bias": "model.safetensors",
|
| 638 |
+
"fm_decoder.layers.8.feed_forward3.in_proj.weight": "model.safetensors",
|
| 639 |
+
"fm_decoder.layers.8.feed_forward3.out_proj.bias": "model.safetensors",
|
| 640 |
+
"fm_decoder.layers.8.feed_forward3.out_proj.weight": "model.safetensors",
|
| 641 |
+
"fm_decoder.layers.8.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 642 |
+
"fm_decoder.layers.8.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 643 |
+
"fm_decoder.layers.8.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 644 |
+
"fm_decoder.layers.8.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 645 |
+
"fm_decoder.layers.8.norm.bias": "model.safetensors",
|
| 646 |
+
"fm_decoder.layers.8.norm.log_scale": "model.safetensors",
|
| 647 |
+
"fm_decoder.layers.8.self_attn1.in_proj.bias": "model.safetensors",
|
| 648 |
+
"fm_decoder.layers.8.self_attn1.in_proj.weight": "model.safetensors",
|
| 649 |
+
"fm_decoder.layers.8.self_attn1.out_proj.bias": "model.safetensors",
|
| 650 |
+
"fm_decoder.layers.8.self_attn1.out_proj.weight": "model.safetensors",
|
| 651 |
+
"fm_decoder.layers.8.self_attn2.in_proj.bias": "model.safetensors",
|
| 652 |
+
"fm_decoder.layers.8.self_attn2.in_proj.weight": "model.safetensors",
|
| 653 |
+
"fm_decoder.layers.8.self_attn2.out_proj.bias": "model.safetensors",
|
| 654 |
+
"fm_decoder.layers.8.self_attn2.out_proj.weight": "model.safetensors",
|
| 655 |
+
"fm_decoder.layers.8.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 656 |
+
"fm_decoder.layers.8.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 657 |
+
"fm_decoder.layers.8.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 658 |
+
"fm_decoder.layers.9.bypass.bypass_scale": "model.safetensors",
|
| 659 |
+
"fm_decoder.layers.9.bypass_mid.bypass_scale": "model.safetensors",
|
| 660 |
+
"fm_decoder.layers.9.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 661 |
+
"fm_decoder.layers.9.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 662 |
+
"fm_decoder.layers.9.conv_module1.in_proj.bias": "model.safetensors",
|
| 663 |
+
"fm_decoder.layers.9.conv_module1.in_proj.weight": "model.safetensors",
|
| 664 |
+
"fm_decoder.layers.9.conv_module1.out_proj.bias": "model.safetensors",
|
| 665 |
+
"fm_decoder.layers.9.conv_module1.out_proj.weight": "model.safetensors",
|
| 666 |
+
"fm_decoder.layers.9.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 667 |
+
"fm_decoder.layers.9.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 668 |
+
"fm_decoder.layers.9.conv_module2.in_proj.bias": "model.safetensors",
|
| 669 |
+
"fm_decoder.layers.9.conv_module2.in_proj.weight": "model.safetensors",
|
| 670 |
+
"fm_decoder.layers.9.conv_module2.out_proj.bias": "model.safetensors",
|
| 671 |
+
"fm_decoder.layers.9.conv_module2.out_proj.weight": "model.safetensors",
|
| 672 |
+
"fm_decoder.layers.9.feed_forward1.in_proj.bias": "model.safetensors",
|
| 673 |
+
"fm_decoder.layers.9.feed_forward1.in_proj.weight": "model.safetensors",
|
| 674 |
+
"fm_decoder.layers.9.feed_forward1.out_proj.bias": "model.safetensors",
|
| 675 |
+
"fm_decoder.layers.9.feed_forward1.out_proj.weight": "model.safetensors",
|
| 676 |
+
"fm_decoder.layers.9.feed_forward2.in_proj.bias": "model.safetensors",
|
| 677 |
+
"fm_decoder.layers.9.feed_forward2.in_proj.weight": "model.safetensors",
|
| 678 |
+
"fm_decoder.layers.9.feed_forward2.out_proj.bias": "model.safetensors",
|
| 679 |
+
"fm_decoder.layers.9.feed_forward2.out_proj.weight": "model.safetensors",
|
| 680 |
+
"fm_decoder.layers.9.feed_forward3.in_proj.bias": "model.safetensors",
|
| 681 |
+
"fm_decoder.layers.9.feed_forward3.in_proj.weight": "model.safetensors",
|
| 682 |
+
"fm_decoder.layers.9.feed_forward3.out_proj.bias": "model.safetensors",
|
| 683 |
+
"fm_decoder.layers.9.feed_forward3.out_proj.weight": "model.safetensors",
|
| 684 |
+
"fm_decoder.layers.9.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 685 |
+
"fm_decoder.layers.9.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 686 |
+
"fm_decoder.layers.9.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 687 |
+
"fm_decoder.layers.9.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 688 |
+
"fm_decoder.layers.9.norm.bias": "model.safetensors",
|
| 689 |
+
"fm_decoder.layers.9.norm.log_scale": "model.safetensors",
|
| 690 |
+
"fm_decoder.layers.9.self_attn1.in_proj.bias": "model.safetensors",
|
| 691 |
+
"fm_decoder.layers.9.self_attn1.in_proj.weight": "model.safetensors",
|
| 692 |
+
"fm_decoder.layers.9.self_attn1.out_proj.bias": "model.safetensors",
|
| 693 |
+
"fm_decoder.layers.9.self_attn1.out_proj.weight": "model.safetensors",
|
| 694 |
+
"fm_decoder.layers.9.self_attn2.in_proj.bias": "model.safetensors",
|
| 695 |
+
"fm_decoder.layers.9.self_attn2.in_proj.weight": "model.safetensors",
|
| 696 |
+
"fm_decoder.layers.9.self_attn2.out_proj.bias": "model.safetensors",
|
| 697 |
+
"fm_decoder.layers.9.self_attn2.out_proj.weight": "model.safetensors",
|
| 698 |
+
"fm_decoder.layers.9.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 699 |
+
"fm_decoder.layers.9.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 700 |
+
"fm_decoder.layers.9.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 701 |
+
"fm_decoder.out_combiner.1.bypass_scale": "model.safetensors",
|
| 702 |
+
"fm_decoder.out_combiner.2.bypass_scale": "model.safetensors",
|
| 703 |
+
"fm_decoder.out_combiner.3.bypass_scale": "model.safetensors",
|
| 704 |
+
"fm_decoder.out_proj.bias": "model.safetensors",
|
| 705 |
+
"fm_decoder.out_proj.weight": "model.safetensors",
|
| 706 |
+
"fm_decoder.stack_time_emb.0.1.bias": "model.safetensors",
|
| 707 |
+
"fm_decoder.stack_time_emb.0.1.weight": "model.safetensors",
|
| 708 |
+
"fm_decoder.stack_time_emb.1.1.bias": "model.safetensors",
|
| 709 |
+
"fm_decoder.stack_time_emb.1.1.weight": "model.safetensors",
|
| 710 |
+
"fm_decoder.stack_time_emb.2.1.bias": "model.safetensors",
|
| 711 |
+
"fm_decoder.stack_time_emb.2.1.weight": "model.safetensors",
|
| 712 |
+
"fm_decoder.stack_time_emb.3.1.bias": "model.safetensors",
|
| 713 |
+
"fm_decoder.stack_time_emb.3.1.weight": "model.safetensors",
|
| 714 |
+
"fm_decoder.stack_time_emb.4.1.bias": "model.safetensors",
|
| 715 |
+
"fm_decoder.stack_time_emb.4.1.weight": "model.safetensors",
|
| 716 |
+
"fm_decoder.time_embed.0.bias": "model.safetensors",
|
| 717 |
+
"fm_decoder.time_embed.0.weight": "model.safetensors",
|
| 718 |
+
"fm_decoder.time_embed.2.bias": "model.safetensors",
|
| 719 |
+
"fm_decoder.time_embed.2.weight": "model.safetensors",
|
| 720 |
+
"text_encoder.in_proj.bias": "model.safetensors",
|
| 721 |
+
"text_encoder.in_proj.weight": "model.safetensors",
|
| 722 |
+
"text_encoder.layers.0.bypass.bypass_scale": "model.safetensors",
|
| 723 |
+
"text_encoder.layers.0.bypass_mid.bypass_scale": "model.safetensors",
|
| 724 |
+
"text_encoder.layers.0.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 725 |
+
"text_encoder.layers.0.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 726 |
+
"text_encoder.layers.0.conv_module1.in_proj.bias": "model.safetensors",
|
| 727 |
+
"text_encoder.layers.0.conv_module1.in_proj.weight": "model.safetensors",
|
| 728 |
+
"text_encoder.layers.0.conv_module1.out_proj.bias": "model.safetensors",
|
| 729 |
+
"text_encoder.layers.0.conv_module1.out_proj.weight": "model.safetensors",
|
| 730 |
+
"text_encoder.layers.0.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 731 |
+
"text_encoder.layers.0.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 732 |
+
"text_encoder.layers.0.conv_module2.in_proj.bias": "model.safetensors",
|
| 733 |
+
"text_encoder.layers.0.conv_module2.in_proj.weight": "model.safetensors",
|
| 734 |
+
"text_encoder.layers.0.conv_module2.out_proj.bias": "model.safetensors",
|
| 735 |
+
"text_encoder.layers.0.conv_module2.out_proj.weight": "model.safetensors",
|
| 736 |
+
"text_encoder.layers.0.feed_forward1.in_proj.bias": "model.safetensors",
|
| 737 |
+
"text_encoder.layers.0.feed_forward1.in_proj.weight": "model.safetensors",
|
| 738 |
+
"text_encoder.layers.0.feed_forward1.out_proj.bias": "model.safetensors",
|
| 739 |
+
"text_encoder.layers.0.feed_forward1.out_proj.weight": "model.safetensors",
|
| 740 |
+
"text_encoder.layers.0.feed_forward2.in_proj.bias": "model.safetensors",
|
| 741 |
+
"text_encoder.layers.0.feed_forward2.in_proj.weight": "model.safetensors",
|
| 742 |
+
"text_encoder.layers.0.feed_forward2.out_proj.bias": "model.safetensors",
|
| 743 |
+
"text_encoder.layers.0.feed_forward2.out_proj.weight": "model.safetensors",
|
| 744 |
+
"text_encoder.layers.0.feed_forward3.in_proj.bias": "model.safetensors",
|
| 745 |
+
"text_encoder.layers.0.feed_forward3.in_proj.weight": "model.safetensors",
|
| 746 |
+
"text_encoder.layers.0.feed_forward3.out_proj.bias": "model.safetensors",
|
| 747 |
+
"text_encoder.layers.0.feed_forward3.out_proj.weight": "model.safetensors",
|
| 748 |
+
"text_encoder.layers.0.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 749 |
+
"text_encoder.layers.0.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 750 |
+
"text_encoder.layers.0.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 751 |
+
"text_encoder.layers.0.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 752 |
+
"text_encoder.layers.0.norm.bias": "model.safetensors",
|
| 753 |
+
"text_encoder.layers.0.norm.log_scale": "model.safetensors",
|
| 754 |
+
"text_encoder.layers.0.self_attn1.in_proj.bias": "model.safetensors",
|
| 755 |
+
"text_encoder.layers.0.self_attn1.in_proj.weight": "model.safetensors",
|
| 756 |
+
"text_encoder.layers.0.self_attn1.out_proj.bias": "model.safetensors",
|
| 757 |
+
"text_encoder.layers.0.self_attn1.out_proj.weight": "model.safetensors",
|
| 758 |
+
"text_encoder.layers.0.self_attn2.in_proj.bias": "model.safetensors",
|
| 759 |
+
"text_encoder.layers.0.self_attn2.in_proj.weight": "model.safetensors",
|
| 760 |
+
"text_encoder.layers.0.self_attn2.out_proj.bias": "model.safetensors",
|
| 761 |
+
"text_encoder.layers.0.self_attn2.out_proj.weight": "model.safetensors",
|
| 762 |
+
"text_encoder.layers.0.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 763 |
+
"text_encoder.layers.0.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 764 |
+
"text_encoder.layers.0.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 765 |
+
"text_encoder.layers.1.bypass.bypass_scale": "model.safetensors",
|
| 766 |
+
"text_encoder.layers.1.bypass_mid.bypass_scale": "model.safetensors",
|
| 767 |
+
"text_encoder.layers.1.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 768 |
+
"text_encoder.layers.1.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 769 |
+
"text_encoder.layers.1.conv_module1.in_proj.bias": "model.safetensors",
|
| 770 |
+
"text_encoder.layers.1.conv_module1.in_proj.weight": "model.safetensors",
|
| 771 |
+
"text_encoder.layers.1.conv_module1.out_proj.bias": "model.safetensors",
|
| 772 |
+
"text_encoder.layers.1.conv_module1.out_proj.weight": "model.safetensors",
|
| 773 |
+
"text_encoder.layers.1.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 774 |
+
"text_encoder.layers.1.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 775 |
+
"text_encoder.layers.1.conv_module2.in_proj.bias": "model.safetensors",
|
| 776 |
+
"text_encoder.layers.1.conv_module2.in_proj.weight": "model.safetensors",
|
| 777 |
+
"text_encoder.layers.1.conv_module2.out_proj.bias": "model.safetensors",
|
| 778 |
+
"text_encoder.layers.1.conv_module2.out_proj.weight": "model.safetensors",
|
| 779 |
+
"text_encoder.layers.1.feed_forward1.in_proj.bias": "model.safetensors",
|
| 780 |
+
"text_encoder.layers.1.feed_forward1.in_proj.weight": "model.safetensors",
|
| 781 |
+
"text_encoder.layers.1.feed_forward1.out_proj.bias": "model.safetensors",
|
| 782 |
+
"text_encoder.layers.1.feed_forward1.out_proj.weight": "model.safetensors",
|
| 783 |
+
"text_encoder.layers.1.feed_forward2.in_proj.bias": "model.safetensors",
|
| 784 |
+
"text_encoder.layers.1.feed_forward2.in_proj.weight": "model.safetensors",
|
| 785 |
+
"text_encoder.layers.1.feed_forward2.out_proj.bias": "model.safetensors",
|
| 786 |
+
"text_encoder.layers.1.feed_forward2.out_proj.weight": "model.safetensors",
|
| 787 |
+
"text_encoder.layers.1.feed_forward3.in_proj.bias": "model.safetensors",
|
| 788 |
+
"text_encoder.layers.1.feed_forward3.in_proj.weight": "model.safetensors",
|
| 789 |
+
"text_encoder.layers.1.feed_forward3.out_proj.bias": "model.safetensors",
|
| 790 |
+
"text_encoder.layers.1.feed_forward3.out_proj.weight": "model.safetensors",
|
| 791 |
+
"text_encoder.layers.1.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 792 |
+
"text_encoder.layers.1.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 793 |
+
"text_encoder.layers.1.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 794 |
+
"text_encoder.layers.1.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 795 |
+
"text_encoder.layers.1.norm.bias": "model.safetensors",
|
| 796 |
+
"text_encoder.layers.1.norm.log_scale": "model.safetensors",
|
| 797 |
+
"text_encoder.layers.1.self_attn1.in_proj.bias": "model.safetensors",
|
| 798 |
+
"text_encoder.layers.1.self_attn1.in_proj.weight": "model.safetensors",
|
| 799 |
+
"text_encoder.layers.1.self_attn1.out_proj.bias": "model.safetensors",
|
| 800 |
+
"text_encoder.layers.1.self_attn1.out_proj.weight": "model.safetensors",
|
| 801 |
+
"text_encoder.layers.1.self_attn2.in_proj.bias": "model.safetensors",
|
| 802 |
+
"text_encoder.layers.1.self_attn2.in_proj.weight": "model.safetensors",
|
| 803 |
+
"text_encoder.layers.1.self_attn2.out_proj.bias": "model.safetensors",
|
| 804 |
+
"text_encoder.layers.1.self_attn2.out_proj.weight": "model.safetensors",
|
| 805 |
+
"text_encoder.layers.1.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 806 |
+
"text_encoder.layers.1.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 807 |
+
"text_encoder.layers.1.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 808 |
+
"text_encoder.layers.2.bypass.bypass_scale": "model.safetensors",
|
| 809 |
+
"text_encoder.layers.2.bypass_mid.bypass_scale": "model.safetensors",
|
| 810 |
+
"text_encoder.layers.2.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 811 |
+
"text_encoder.layers.2.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 812 |
+
"text_encoder.layers.2.conv_module1.in_proj.bias": "model.safetensors",
|
| 813 |
+
"text_encoder.layers.2.conv_module1.in_proj.weight": "model.safetensors",
|
| 814 |
+
"text_encoder.layers.2.conv_module1.out_proj.bias": "model.safetensors",
|
| 815 |
+
"text_encoder.layers.2.conv_module1.out_proj.weight": "model.safetensors",
|
| 816 |
+
"text_encoder.layers.2.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 817 |
+
"text_encoder.layers.2.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 818 |
+
"text_encoder.layers.2.conv_module2.in_proj.bias": "model.safetensors",
|
| 819 |
+
"text_encoder.layers.2.conv_module2.in_proj.weight": "model.safetensors",
|
| 820 |
+
"text_encoder.layers.2.conv_module2.out_proj.bias": "model.safetensors",
|
| 821 |
+
"text_encoder.layers.2.conv_module2.out_proj.weight": "model.safetensors",
|
| 822 |
+
"text_encoder.layers.2.feed_forward1.in_proj.bias": "model.safetensors",
|
| 823 |
+
"text_encoder.layers.2.feed_forward1.in_proj.weight": "model.safetensors",
|
| 824 |
+
"text_encoder.layers.2.feed_forward1.out_proj.bias": "model.safetensors",
|
| 825 |
+
"text_encoder.layers.2.feed_forward1.out_proj.weight": "model.safetensors",
|
| 826 |
+
"text_encoder.layers.2.feed_forward2.in_proj.bias": "model.safetensors",
|
| 827 |
+
"text_encoder.layers.2.feed_forward2.in_proj.weight": "model.safetensors",
|
| 828 |
+
"text_encoder.layers.2.feed_forward2.out_proj.bias": "model.safetensors",
|
| 829 |
+
"text_encoder.layers.2.feed_forward2.out_proj.weight": "model.safetensors",
|
| 830 |
+
"text_encoder.layers.2.feed_forward3.in_proj.bias": "model.safetensors",
|
| 831 |
+
"text_encoder.layers.2.feed_forward3.in_proj.weight": "model.safetensors",
|
| 832 |
+
"text_encoder.layers.2.feed_forward3.out_proj.bias": "model.safetensors",
|
| 833 |
+
"text_encoder.layers.2.feed_forward3.out_proj.weight": "model.safetensors",
|
| 834 |
+
"text_encoder.layers.2.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 835 |
+
"text_encoder.layers.2.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 836 |
+
"text_encoder.layers.2.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 837 |
+
"text_encoder.layers.2.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 838 |
+
"text_encoder.layers.2.norm.bias": "model.safetensors",
|
| 839 |
+
"text_encoder.layers.2.norm.log_scale": "model.safetensors",
|
| 840 |
+
"text_encoder.layers.2.self_attn1.in_proj.bias": "model.safetensors",
|
| 841 |
+
"text_encoder.layers.2.self_attn1.in_proj.weight": "model.safetensors",
|
| 842 |
+
"text_encoder.layers.2.self_attn1.out_proj.bias": "model.safetensors",
|
| 843 |
+
"text_encoder.layers.2.self_attn1.out_proj.weight": "model.safetensors",
|
| 844 |
+
"text_encoder.layers.2.self_attn2.in_proj.bias": "model.safetensors",
|
| 845 |
+
"text_encoder.layers.2.self_attn2.in_proj.weight": "model.safetensors",
|
| 846 |
+
"text_encoder.layers.2.self_attn2.out_proj.bias": "model.safetensors",
|
| 847 |
+
"text_encoder.layers.2.self_attn2.out_proj.weight": "model.safetensors",
|
| 848 |
+
"text_encoder.layers.2.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 849 |
+
"text_encoder.layers.2.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 850 |
+
"text_encoder.layers.2.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 851 |
+
"text_encoder.layers.3.bypass.bypass_scale": "model.safetensors",
|
| 852 |
+
"text_encoder.layers.3.bypass_mid.bypass_scale": "model.safetensors",
|
| 853 |
+
"text_encoder.layers.3.conv_module1.depthwise_conv.bias": "model.safetensors",
|
| 854 |
+
"text_encoder.layers.3.conv_module1.depthwise_conv.weight": "model.safetensors",
|
| 855 |
+
"text_encoder.layers.3.conv_module1.in_proj.bias": "model.safetensors",
|
| 856 |
+
"text_encoder.layers.3.conv_module1.in_proj.weight": "model.safetensors",
|
| 857 |
+
"text_encoder.layers.3.conv_module1.out_proj.bias": "model.safetensors",
|
| 858 |
+
"text_encoder.layers.3.conv_module1.out_proj.weight": "model.safetensors",
|
| 859 |
+
"text_encoder.layers.3.conv_module2.depthwise_conv.bias": "model.safetensors",
|
| 860 |
+
"text_encoder.layers.3.conv_module2.depthwise_conv.weight": "model.safetensors",
|
| 861 |
+
"text_encoder.layers.3.conv_module2.in_proj.bias": "model.safetensors",
|
| 862 |
+
"text_encoder.layers.3.conv_module2.in_proj.weight": "model.safetensors",
|
| 863 |
+
"text_encoder.layers.3.conv_module2.out_proj.bias": "model.safetensors",
|
| 864 |
+
"text_encoder.layers.3.conv_module2.out_proj.weight": "model.safetensors",
|
| 865 |
+
"text_encoder.layers.3.feed_forward1.in_proj.bias": "model.safetensors",
|
| 866 |
+
"text_encoder.layers.3.feed_forward1.in_proj.weight": "model.safetensors",
|
| 867 |
+
"text_encoder.layers.3.feed_forward1.out_proj.bias": "model.safetensors",
|
| 868 |
+
"text_encoder.layers.3.feed_forward1.out_proj.weight": "model.safetensors",
|
| 869 |
+
"text_encoder.layers.3.feed_forward2.in_proj.bias": "model.safetensors",
|
| 870 |
+
"text_encoder.layers.3.feed_forward2.in_proj.weight": "model.safetensors",
|
| 871 |
+
"text_encoder.layers.3.feed_forward2.out_proj.bias": "model.safetensors",
|
| 872 |
+
"text_encoder.layers.3.feed_forward2.out_proj.weight": "model.safetensors",
|
| 873 |
+
"text_encoder.layers.3.feed_forward3.in_proj.bias": "model.safetensors",
|
| 874 |
+
"text_encoder.layers.3.feed_forward3.in_proj.weight": "model.safetensors",
|
| 875 |
+
"text_encoder.layers.3.feed_forward3.out_proj.bias": "model.safetensors",
|
| 876 |
+
"text_encoder.layers.3.feed_forward3.out_proj.weight": "model.safetensors",
|
| 877 |
+
"text_encoder.layers.3.nonlin_attention.in_proj.bias": "model.safetensors",
|
| 878 |
+
"text_encoder.layers.3.nonlin_attention.in_proj.weight": "model.safetensors",
|
| 879 |
+
"text_encoder.layers.3.nonlin_attention.out_proj.bias": "model.safetensors",
|
| 880 |
+
"text_encoder.layers.3.nonlin_attention.out_proj.weight": "model.safetensors",
|
| 881 |
+
"text_encoder.layers.3.norm.bias": "model.safetensors",
|
| 882 |
+
"text_encoder.layers.3.norm.log_scale": "model.safetensors",
|
| 883 |
+
"text_encoder.layers.3.self_attn1.in_proj.bias": "model.safetensors",
|
| 884 |
+
"text_encoder.layers.3.self_attn1.in_proj.weight": "model.safetensors",
|
| 885 |
+
"text_encoder.layers.3.self_attn1.out_proj.bias": "model.safetensors",
|
| 886 |
+
"text_encoder.layers.3.self_attn1.out_proj.weight": "model.safetensors",
|
| 887 |
+
"text_encoder.layers.3.self_attn2.in_proj.bias": "model.safetensors",
|
| 888 |
+
"text_encoder.layers.3.self_attn2.in_proj.weight": "model.safetensors",
|
| 889 |
+
"text_encoder.layers.3.self_attn2.out_proj.bias": "model.safetensors",
|
| 890 |
+
"text_encoder.layers.3.self_attn2.out_proj.weight": "model.safetensors",
|
| 891 |
+
"text_encoder.layers.3.self_attn_weights.in_proj.bias": "model.safetensors",
|
| 892 |
+
"text_encoder.layers.3.self_attn_weights.in_proj.weight": "model.safetensors",
|
| 893 |
+
"text_encoder.layers.3.self_attn_weights.linear_pos.weight": "model.safetensors",
|
| 894 |
+
"text_encoder.out_proj.bias": "model.safetensors",
|
| 895 |
+
"text_encoder.out_proj.weight": "model.safetensors"
|
| 896 |
+
}
|
| 897 |
+
}
|
tokens.txt
ADDED
|
@@ -0,0 +1,360 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_ 0
|
| 2 |
+
^ 1
|
| 3 |
+
$ 2
|
| 4 |
+
3
|
| 5 |
+
! 4
|
| 6 |
+
' 5
|
| 7 |
+
( 6
|
| 8 |
+
) 7
|
| 9 |
+
, 8
|
| 10 |
+
- 9
|
| 11 |
+
. 10
|
| 12 |
+
: 11
|
| 13 |
+
; 12
|
| 14 |
+
? 13
|
| 15 |
+
a 14
|
| 16 |
+
b 15
|
| 17 |
+
c 16
|
| 18 |
+
d 17
|
| 19 |
+
e 18
|
| 20 |
+
f 19
|
| 21 |
+
h 20
|
| 22 |
+
i 21
|
| 23 |
+
j 22
|
| 24 |
+
k 23
|
| 25 |
+
l 24
|
| 26 |
+
m 25
|
| 27 |
+
n 26
|
| 28 |
+
o 27
|
| 29 |
+
p 28
|
| 30 |
+
q 29
|
| 31 |
+
r 30
|
| 32 |
+
s 31
|
| 33 |
+
t 32
|
| 34 |
+
u 33
|
| 35 |
+
v 34
|
| 36 |
+
w 35
|
| 37 |
+
x 36
|
| 38 |
+
y 37
|
| 39 |
+
z 38
|
| 40 |
+
æ 39
|
| 41 |
+
ç 40
|
| 42 |
+
ð 41
|
| 43 |
+
ø 42
|
| 44 |
+
ħ 43
|
| 45 |
+
ŋ 44
|
| 46 |
+
œ 45
|
| 47 |
+
ǀ 46
|
| 48 |
+
ǁ 47
|
| 49 |
+
ǂ 48
|
| 50 |
+
ǃ 49
|
| 51 |
+
ɐ 50
|
| 52 |
+
ɑ 51
|
| 53 |
+
ɒ 52
|
| 54 |
+
ɓ 53
|
| 55 |
+
ɔ 54
|
| 56 |
+
ɕ 55
|
| 57 |
+
ɖ 56
|
| 58 |
+
ɗ 57
|
| 59 |
+
ɘ 58
|
| 60 |
+
ə 59
|
| 61 |
+
ɚ 60
|
| 62 |
+
ɛ 61
|
| 63 |
+
ɜ 62
|
| 64 |
+
ɞ 63
|
| 65 |
+
ɟ 64
|
| 66 |
+
ɠ 65
|
| 67 |
+
ɡ 66
|
| 68 |
+
ɢ 67
|
| 69 |
+
ɣ 68
|
| 70 |
+
ɤ 69
|
| 71 |
+
ɥ 70
|
| 72 |
+
ɦ 71
|
| 73 |
+
ɧ 72
|
| 74 |
+
ɨ 73
|
| 75 |
+
ɪ 74
|
| 76 |
+
ɫ 75
|
| 77 |
+
ɬ 76
|
| 78 |
+
ɭ 77
|
| 79 |
+
ɮ 78
|
| 80 |
+
ɯ 79
|
| 81 |
+
ɰ 80
|
| 82 |
+
ɱ 81
|
| 83 |
+
ɲ 82
|
| 84 |
+
ɳ 83
|
| 85 |
+
ɴ 84
|
| 86 |
+
ɵ 85
|
| 87 |
+
ɶ 86
|
| 88 |
+
ɸ 87
|
| 89 |
+
ɹ 88
|
| 90 |
+
ɺ 89
|
| 91 |
+
ɻ 90
|
| 92 |
+
ɽ 91
|
| 93 |
+
ɾ 92
|
| 94 |
+
ʀ 93
|
| 95 |
+
ʁ 94
|
| 96 |
+
ʂ 95
|
| 97 |
+
ʃ 96
|
| 98 |
+
ʄ 97
|
| 99 |
+
ʈ 98
|
| 100 |
+
ʉ 99
|
| 101 |
+
ʊ 100
|
| 102 |
+
ʋ 101
|
| 103 |
+
ʌ 102
|
| 104 |
+
ʍ 103
|
| 105 |
+
ʎ 104
|
| 106 |
+
ʏ 105
|
| 107 |
+
ʐ 106
|
| 108 |
+
ʑ 107
|
| 109 |
+
ʒ 108
|
| 110 |
+
ʔ 109
|
| 111 |
+
ʕ 110
|
| 112 |
+
ʘ 111
|
| 113 |
+
ʙ 112
|
| 114 |
+
ʛ 113
|
| 115 |
+
ʜ 114
|
| 116 |
+
ʝ 115
|
| 117 |
+
ʟ 116
|
| 118 |
+
ʡ 117
|
| 119 |
+
ʢ 118
|
| 120 |
+
ʲ 119
|
| 121 |
+
ˈ 120
|
| 122 |
+
ˌ 121
|
| 123 |
+
ː 122
|
| 124 |
+
ˑ 123
|
| 125 |
+
˞ 124
|
| 126 |
+
β 125
|
| 127 |
+
θ 126
|
| 128 |
+
χ 127
|
| 129 |
+
ᵻ 128
|
| 130 |
+
ⱱ 129
|
| 131 |
+
0 130
|
| 132 |
+
1 131
|
| 133 |
+
2 132
|
| 134 |
+
3 133
|
| 135 |
+
4 134
|
| 136 |
+
5 135
|
| 137 |
+
6 136
|
| 138 |
+
7 137
|
| 139 |
+
8 138
|
| 140 |
+
9 139
|
| 141 |
+
̧ 140
|
| 142 |
+
̃ 141
|
| 143 |
+
̪ 142
|
| 144 |
+
̯ 143
|
| 145 |
+
̩ 144
|
| 146 |
+
ʰ 145
|
| 147 |
+
ˤ 146
|
| 148 |
+
ε 147
|
| 149 |
+
↓ 148
|
| 150 |
+
# 149
|
| 151 |
+
" 150
|
| 152 |
+
↑ 151
|
| 153 |
+
̺ 152
|
| 154 |
+
̻ 153
|
| 155 |
+
g 154
|
| 156 |
+
ʦ 155
|
| 157 |
+
X 156
|
| 158 |
+
̝ 157
|
| 159 |
+
̊ 158
|
| 160 |
+
a1 159
|
| 161 |
+
a2 160
|
| 162 |
+
a3 161
|
| 163 |
+
a4 162
|
| 164 |
+
a5 163
|
| 165 |
+
ai1 164
|
| 166 |
+
ai2 165
|
| 167 |
+
ai3 166
|
| 168 |
+
ai4 167
|
| 169 |
+
ai5 168
|
| 170 |
+
an1 169
|
| 171 |
+
an2 170
|
| 172 |
+
an3 171
|
| 173 |
+
an4 172
|
| 174 |
+
an5 173
|
| 175 |
+
ang1 174
|
| 176 |
+
ang2 175
|
| 177 |
+
ang3 176
|
| 178 |
+
ang4 177
|
| 179 |
+
ang5 178
|
| 180 |
+
ao1 179
|
| 181 |
+
ao2 180
|
| 182 |
+
ao3 181
|
| 183 |
+
ao4 182
|
| 184 |
+
ao5 183
|
| 185 |
+
b0 184
|
| 186 |
+
c0 185
|
| 187 |
+
ch0 186
|
| 188 |
+
d0 187
|
| 189 |
+
e1 188
|
| 190 |
+
e2 189
|
| 191 |
+
e3 190
|
| 192 |
+
e4 191
|
| 193 |
+
e5 192
|
| 194 |
+
ei1 193
|
| 195 |
+
ei2 194
|
| 196 |
+
ei3 195
|
| 197 |
+
ei4 196
|
| 198 |
+
ei5 197
|
| 199 |
+
en1 198
|
| 200 |
+
en2 199
|
| 201 |
+
en3 200
|
| 202 |
+
en4 201
|
| 203 |
+
en5 202
|
| 204 |
+
eng1 203
|
| 205 |
+
eng2 204
|
| 206 |
+
eng3 205
|
| 207 |
+
eng4 206
|
| 208 |
+
eng5 207
|
| 209 |
+
er2 208
|
| 210 |
+
er3 209
|
| 211 |
+
er4 210
|
| 212 |
+
er5 211
|
| 213 |
+
f0 212
|
| 214 |
+
g0 213
|
| 215 |
+
g2 214
|
| 216 |
+
g3 215
|
| 217 |
+
g4 216
|
| 218 |
+
g5 217
|
| 219 |
+
h0 218
|
| 220 |
+
i1 219
|
| 221 |
+
i2 220
|
| 222 |
+
i3 221
|
| 223 |
+
i4 222
|
| 224 |
+
i5 223
|
| 225 |
+
ia1 224
|
| 226 |
+
ia2 225
|
| 227 |
+
ia3 226
|
| 228 |
+
ia4 227
|
| 229 |
+
ia5 228
|
| 230 |
+
ian1 229
|
| 231 |
+
ian2 230
|
| 232 |
+
ian3 231
|
| 233 |
+
ian4 232
|
| 234 |
+
ian5 233
|
| 235 |
+
iang1 234
|
| 236 |
+
iang2 235
|
| 237 |
+
iang3 236
|
| 238 |
+
iang4 237
|
| 239 |
+
iang5 238
|
| 240 |
+
iao1 239
|
| 241 |
+
iao2 240
|
| 242 |
+
iao3 241
|
| 243 |
+
iao4 242
|
| 244 |
+
iao5 243
|
| 245 |
+
ie1 244
|
| 246 |
+
ie2 245
|
| 247 |
+
ie3 246
|
| 248 |
+
ie4 247
|
| 249 |
+
ie5 248
|
| 250 |
+
in1 249
|
| 251 |
+
in2 250
|
| 252 |
+
in3 251
|
| 253 |
+
in4 252
|
| 254 |
+
in5 253
|
| 255 |
+
ing1 254
|
| 256 |
+
ing2 255
|
| 257 |
+
ing3 256
|
| 258 |
+
ing4 257
|
| 259 |
+
ing5 258
|
| 260 |
+
iong1 259
|
| 261 |
+
iong2 260
|
| 262 |
+
iong3 261
|
| 263 |
+
iong4 262
|
| 264 |
+
iu1 263
|
| 265 |
+
iu2 264
|
| 266 |
+
iu3 265
|
| 267 |
+
iu4 266
|
| 268 |
+
iu5 267
|
| 269 |
+
j0 268
|
| 270 |
+
k0 269
|
| 271 |
+
l0 270
|
| 272 |
+
m0 271
|
| 273 |
+
m1 272
|
| 274 |
+
m2 273
|
| 275 |
+
m4 274
|
| 276 |
+
m5 275
|
| 277 |
+
n0 276
|
| 278 |
+
n2 277
|
| 279 |
+
n3 278
|
| 280 |
+
n4 279
|
| 281 |
+
n5 280
|
| 282 |
+
ng5 281
|
| 283 |
+
o1 282
|
| 284 |
+
o2 283
|
| 285 |
+
o3 284
|
| 286 |
+
o4 285
|
| 287 |
+
o5 286
|
| 288 |
+
ong1 287
|
| 289 |
+
ong2 288
|
| 290 |
+
ong3 289
|
| 291 |
+
ong4 290
|
| 292 |
+
ong5 291
|
| 293 |
+
ou1 292
|
| 294 |
+
ou2 293
|
| 295 |
+
ou3 294
|
| 296 |
+
ou4 295
|
| 297 |
+
ou5 296
|
| 298 |
+
p0 297
|
| 299 |
+
q0 298
|
| 300 |
+
r0 299
|
| 301 |
+
s0 300
|
| 302 |
+
sh0 301
|
| 303 |
+
t0 302
|
| 304 |
+
u1 303
|
| 305 |
+
u2 304
|
| 306 |
+
u3 305
|
| 307 |
+
u4 306
|
| 308 |
+
u5 307
|
| 309 |
+
ua1 308
|
| 310 |
+
ua2 309
|
| 311 |
+
ua3 310
|
| 312 |
+
ua4 311
|
| 313 |
+
uai1 312
|
| 314 |
+
uai2 313
|
| 315 |
+
uai3 314
|
| 316 |
+
uai4 315
|
| 317 |
+
uai5 316
|
| 318 |
+
uan1 317
|
| 319 |
+
uan2 318
|
| 320 |
+
uan3 319
|
| 321 |
+
uan4 320
|
| 322 |
+
uan5 321
|
| 323 |
+
uang1 322
|
| 324 |
+
uang2 323
|
| 325 |
+
uang3 324
|
| 326 |
+
uang4 325
|
| 327 |
+
uang5 326
|
| 328 |
+
ue1 327
|
| 329 |
+
ue2 328
|
| 330 |
+
ue3 329
|
| 331 |
+
ue4 330
|
| 332 |
+
ui1 331
|
| 333 |
+
ui2 332
|
| 334 |
+
ui3 333
|
| 335 |
+
ui4 334
|
| 336 |
+
ui5 335
|
| 337 |
+
un1 336
|
| 338 |
+
un2 337
|
| 339 |
+
un3 338
|
| 340 |
+
un4 339
|
| 341 |
+
un5 340
|
| 342 |
+
uo1 341
|
| 343 |
+
uo2 342
|
| 344 |
+
uo3 343
|
| 345 |
+
uo4 344
|
| 346 |
+
uo5 345
|
| 347 |
+
v2 346
|
| 348 |
+
v3 347
|
| 349 |
+
v4 348
|
| 350 |
+
ve3 349
|
| 351 |
+
ve4 350
|
| 352 |
+
w0 351
|
| 353 |
+
x0 352
|
| 354 |
+
y0 353
|
| 355 |
+
z0 354
|
| 356 |
+
zh0 355
|
| 357 |
+
ê1 356
|
| 358 |
+
ê2 357
|
| 359 |
+
ê3 358
|
| 360 |
+
ê4 359
|
vocoder_config.yaml
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
feature_extractor:
|
| 2 |
+
class_path: vocos.feature_extractors.MelSpectrogramFeatures
|
| 3 |
+
init_args:
|
| 4 |
+
sample_rate: 24000
|
| 5 |
+
n_fft: 1024
|
| 6 |
+
hop_length: 256
|
| 7 |
+
n_mels: 100
|
| 8 |
+
padding: center
|
| 9 |
+
|
| 10 |
+
backbone:
|
| 11 |
+
class_path: vocos.models.VocosBackbone
|
| 12 |
+
init_args:
|
| 13 |
+
input_channels: 100
|
| 14 |
+
dim: 512
|
| 15 |
+
intermediate_dim: 1536
|
| 16 |
+
num_layers: 8
|
| 17 |
+
|
| 18 |
+
head:
|
| 19 |
+
class_path: vocos.heads.ISTFTHead
|
| 20 |
+
init_args:
|
| 21 |
+
dim: 512
|
| 22 |
+
n_fft: 1024
|
| 23 |
+
hop_length: 256
|
| 24 |
+
padding: center
|
| 25 |
+
|
| 26 |
+
head_48k:
|
| 27 |
+
class_path: vocos.heads.ISTFTHead
|
| 28 |
+
init_args:
|
| 29 |
+
dim: 512
|
| 30 |
+
n_fft: 1024
|
| 31 |
+
hop_length: 256
|
| 32 |
+
padding: center
|
| 33 |
+
|
| 34 |
+
upsampler:
|
| 35 |
+
class_path: linacodec.vocoder.upsampler_block.UpSamplerBlock
|
| 36 |
+
init_args:
|
| 37 |
+
in_channels: 512
|
| 38 |
+
upsample_factors: [2, 1]
|
| 39 |
+
kernel_sizes: [8, 8]
|
vocos.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2427596b2d4766b510227efc846eee12fba53b5e43a9c60acc79ada52012efd0
|
| 3 |
+
size 31980552
|