alexwengg commited on
Commit
a7fc1eb
·
verified ·
1 Parent(s): 7bf0615

Upload 101 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +212 -0
  2. compiled/.DS_Store +0 -0
  3. compiled/styletts2_decoder_1024.mlmodelc/analytics/coremldata.bin +3 -0
  4. compiled/styletts2_decoder_1024.mlmodelc/coremldata.bin +3 -0
  5. compiled/styletts2_decoder_1024.mlmodelc/metadata.json +114 -0
  6. compiled/styletts2_decoder_1024.mlmodelc/model.mil +0 -0
  7. compiled/styletts2_decoder_1024.mlmodelc/weights/weight.bin +3 -0
  8. compiled/styletts2_decoder_2048.mlmodelc/analytics/coremldata.bin +3 -0
  9. compiled/styletts2_decoder_2048.mlmodelc/coremldata.bin +3 -0
  10. compiled/styletts2_decoder_2048.mlmodelc/metadata.json +114 -0
  11. compiled/styletts2_decoder_2048.mlmodelc/model.mil +0 -0
  12. compiled/styletts2_decoder_2048.mlmodelc/weights/weight.bin +3 -0
  13. compiled/styletts2_decoder_256.mlmodelc/analytics/coremldata.bin +3 -0
  14. compiled/styletts2_decoder_256.mlmodelc/coremldata.bin +3 -0
  15. compiled/styletts2_decoder_256.mlmodelc/metadata.json +114 -0
  16. compiled/styletts2_decoder_256.mlmodelc/model.mil +0 -0
  17. compiled/styletts2_decoder_256.mlmodelc/weights/weight.bin +3 -0
  18. compiled/styletts2_decoder_4096.mlmodelc/analytics/coremldata.bin +3 -0
  19. compiled/styletts2_decoder_4096.mlmodelc/coremldata.bin +3 -0
  20. compiled/styletts2_decoder_4096.mlmodelc/metadata.json +114 -0
  21. compiled/styletts2_decoder_4096.mlmodelc/model.mil +0 -0
  22. compiled/styletts2_decoder_4096.mlmodelc/weights/weight.bin +3 -0
  23. compiled/styletts2_decoder_512.mlmodelc/analytics/coremldata.bin +3 -0
  24. compiled/styletts2_decoder_512.mlmodelc/coremldata.bin +3 -0
  25. compiled/styletts2_decoder_512.mlmodelc/metadata.json +114 -0
  26. compiled/styletts2_decoder_512.mlmodelc/model.mil +0 -0
  27. compiled/styletts2_decoder_512.mlmodelc/weights/weight.bin +3 -0
  28. compiled/styletts2_diffusion_step_512.mlmodelc/analytics/coremldata.bin +3 -0
  29. compiled/styletts2_diffusion_step_512.mlmodelc/coremldata.bin +3 -0
  30. compiled/styletts2_diffusion_step_512.mlmodelc/metadata.json +114 -0
  31. compiled/styletts2_diffusion_step_512.mlmodelc/model.mil +351 -0
  32. compiled/styletts2_diffusion_step_512.mlmodelc/weights/weight.bin +3 -0
  33. compiled/styletts2_f0n_energy.mlmodelc/analytics/coremldata.bin +3 -0
  34. compiled/styletts2_f0n_energy.mlmodelc/coremldata.bin +3 -0
  35. compiled/styletts2_f0n_energy.mlmodelc/metadata.json +98 -0
  36. compiled/styletts2_f0n_energy.mlmodelc/model.mil +396 -0
  37. compiled/styletts2_f0n_energy.mlmodelc/weights/weight.bin +3 -0
  38. compiled/styletts2_text_predictor_128.mlmodelc/analytics/coremldata.bin +3 -0
  39. compiled/styletts2_text_predictor_128.mlmodelc/coremldata.bin +3 -0
  40. compiled/styletts2_text_predictor_128.mlmodelc/metadata.json +138 -0
  41. compiled/styletts2_text_predictor_128.mlmodelc/model.mil +0 -0
  42. compiled/styletts2_text_predictor_128.mlmodelc/weights/weight.bin +3 -0
  43. compiled/styletts2_text_predictor_256.mlmodelc/analytics/coremldata.bin +3 -0
  44. compiled/styletts2_text_predictor_256.mlmodelc/coremldata.bin +3 -0
  45. compiled/styletts2_text_predictor_256.mlmodelc/metadata.json +138 -0
  46. compiled/styletts2_text_predictor_256.mlmodelc/model.mil +0 -0
  47. compiled/styletts2_text_predictor_256.mlmodelc/weights/weight.bin +3 -0
  48. compiled/styletts2_text_predictor_32.mlmodelc/analytics/coremldata.bin +3 -0
  49. compiled/styletts2_text_predictor_32.mlmodelc/coremldata.bin +3 -0
  50. compiled/styletts2_text_predictor_32.mlmodelc/metadata.json +138 -0
README.md ADDED
@@ -0,0 +1,212 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: yl4579-styletts2
4
+ license_link: https://github.com/yl4579/StyleTTS2#pre-requisites
5
+ language:
6
+ - en
7
+ library_name: coreml
8
+ tags:
9
+ - text-to-speech
10
+ - styletts2
11
+ - coreml
12
+ - apple-silicon
13
+ - libritts
14
+ - on-device
15
+ pipeline_tag: text-to-speech
16
+ inference: false
17
+ ---
18
+
19
+ # StyleTTS2 (LibriTTS) — CoreML
20
+
21
+ Apple-Silicon-optimized CoreML conversion of [yl4579/StyleTTS2](https://github.com/yl4579/StyleTTS2)
22
+ LibriTTS multi-speaker checkpoint
23
+ ([`yl4579/StyleTTS2-LibriTTS` → `Models/LibriTTS/epochs_2nd_00020.pth`](https://huggingface.co/yl4579/StyleTTS2-LibriTTS)).
24
+
25
+ Four-stage pipeline; per-stage compute-unit placement; selective int8 PTQ on
26
+ the text-and-prosody predictor; fp32 decoder.
27
+
28
+ > [!IMPORTANT]
29
+ > **These weights carry use restrictions beyond MIT. Read the License
30
+ > section before downloading.** They are not a drop-in permissively-licensed
31
+ > TTS model. If you need permissive terms, use
32
+ > [Kokoro](https://huggingface.co/hexgrad/Kokoro-82M) instead.
33
+
34
+ ## License & use restrictions
35
+
36
+ The upstream repository code is MIT, but the pre-trained LibriTTS weights
37
+ carry **two non-negotiable restrictions** declared in
38
+ [yl4579/StyleTTS2's README](https://github.com/yl4579/StyleTTS2#pre-requisites):
39
+
40
+ 1. **Synthetic-origin disclosure.** Any deployment that produces audio from
41
+ these weights must clearly disclose to listeners that the audio is
42
+ synthetic. No undisclosed synthetic-speech publishing.
43
+ 2. **Speaker consent for voice cloning.** Cloning a real person's voice
44
+ requires their consent. No unauthorized celebrity / public-figure /
45
+ non-consenting third-party voice cloning.
46
+
47
+ These restrictions ride with the weights through every redistribution,
48
+ fine-tune, and downstream derivative. Anyone downloading this repo inherits
49
+ them and must propagate them in turn.
50
+
51
+ If you cannot or will not honor these terms, **do not download these
52
+ weights**.
53
+
54
+ License-of-record: [github.com/yl4579/StyleTTS2](https://github.com/yl4579/StyleTTS2)
55
+ upstream README at the time of conversion (see *Conversion provenance* below
56
+ for the pinned commit).
57
+
58
+ ## What's in this repo
59
+
60
+ | Package | Compute unit | Precision | Buckets | Called |
61
+ |---|---|---|---|---|
62
+ | `styletts2_text_predictor_{32,64,128,256,512}.mlpackage` | ANE | fp16 | 5 token-length | 1× per utterance |
63
+ | `styletts2_diffusion_step_512.mlpackage` | CPU+GPU | fp16 | 1 (B=512 only) | ~5× per utterance |
64
+ | `styletts2_f0n_energy.mlpackage` | ANE | fp16 | dynamic | 1× per utterance |
65
+ | `styletts2_decoder_{256,512,1024,2048,4096}.mlpackage` | CPU+GPU | **fp32** | 5 mel-length | 1× per utterance |
66
+ | `constants/text_cleaner_vocab.json` | — | — | — | phoneme→id table |
67
+ | `config.json` | — | — | — | bundle runtime contract (audio/sampler/buckets) |
68
+
69
+ Total on-disk size: ~1.4 GB per format.
70
+
71
+ Both source `.mlpackage` (uncompiled, portable across Xcode versions) and
72
+ pre-compiled `.mlmodelc` (Apple Silicon, ready for `MLModel(contentsOf:)`)
73
+ are shipped. The `.mlmodelc` artifacts are under `compiled/`. Pick one:
74
+
75
+ - **`*.mlpackage`** — load via `MLModel(contentsOf:)`; the OS compiles on
76
+ first load (~5–20 s cold start the first time, cached afterward).
77
+ - **`compiled/*.mlmodelc`** — already compiled; same loader path skips the
78
+ on-device compile. Useful for shipping inside an app bundle.
79
+
80
+ The diffusion sampler loop (ADPM2 + Karras schedule + CFG) and the
81
+ hard-alignment matrix (cumsum-of-durations → one-hot → matmul) live in your
82
+ host application (Swift / Python). Per-step inference is in CoreML; control
83
+ flow is not.
84
+
85
+ ### Why the precision split looks like this
86
+
87
+ - **text_predictor is fp16.** Selective int8 PTQ was tried and dropped:
88
+ on Apple Silicon ANE the int8 path saves only ~3 MB per bucket of
89
+ weight bandwidth, has no exposed int8 GEMM, and dequantizes back to
90
+ fp16 on load. The savings did not justify the parity risk on small
91
+ projections.
92
+ - **diffusion_step stays fp16.** It runs 5 times per utterance through an
93
+ ODE-style sampler; quantization noise compounds through iterations.
94
+ Same lesson as PocketTTS issue #7.
95
+ - **f0n_energy stays fp16.** ~6 MB. No bandwidth payoff; quantizing
96
+ small projections injects audible pitch noise.
97
+ - **decoder is fp32, not fp16.** SineGen's harmonic source accumulates
98
+ phase via `cumsum × 2π × hop=300`, reaching magnitudes ~4000
99
+ mid-frame. fp16 precision at that magnitude (~4) is much larger than
100
+ the per-sample increment (~0.05 rad), which scrambles the sine output
101
+ and produces audibly robotic synthesis. fp32 is required end-to-end.
102
+
103
+ ### Why only one diffusion bucket
104
+
105
+ Empirically every observed `bert_dur` fits in B=512. The 32/64/128/256
106
+ buckets were dead weight (~192 MB) given the non-linear cost ladder
107
+ (B=32 ≈ 66 ms/step, B=512 ≈ 152 ms/step). Dropping them adds at most
108
+ ~430 ms per utterance in the worst short case.
109
+
110
+ ## Performance
111
+
112
+ - **RTFx:** 4.32× warm on M-series Mac (5-step ADPM2 sampler, all buckets
113
+ pre-warmed).
114
+ - **Log-mel cosine vs PyTorch fp32:** 0.9687.
115
+ - **ECAPA-TDNN cosine to reference clip:** 0.18 — at the model's
116
+ architectural ceiling. PyTorch fp32 itself only reaches 0.29 on the
117
+ same metric. Voice-clone fidelity is bounded by StyleTTS2's
118
+ architecture, not by this conversion.
119
+
120
+ ## How to use
121
+
122
+ ### Phonemizer
123
+
124
+ espeak-ng IPA + stress. The 178-token vocabulary in
125
+ `constants/text_cleaner_vocab.json` mirrors `text_utils.TextCleaner` from
126
+ the upstream repo: `[pad] + punctuation + ASCII letters + IPA letters`.
127
+
128
+ Pad token is `$` at id 0.
129
+
130
+ ### Inference shape
131
+
132
+ ```text
133
+ text → phonemes → token ids
134
+
135
+
136
+ text_predictor (ANE, int8)
137
+ │ ├─ d_en (1, T_dur, hidden)
138
+ │ ├─ s_pred (1, 256) (sampler init via diffusion)
139
+ │ └─ duration logits → duration → one-hot alignment matrix (host)
140
+
141
+
142
+ diffusion_step × 5 (CPU+GPU, fp16) (ADPM2 + Karras schedule + CFG)
143
+
144
+
145
+ [blend(s, ref_s) + alignment]
146
+
147
+
148
+ f0n_energy (ANE, fp16) → F0_curve, N
149
+
150
+
151
+ decoder (CPU+GPU, fp32) → 24 kHz waveform
152
+ ```
153
+
154
+ The Swift host owns the sampler loop, alignment construction, and bucket
155
+ routing. A reference Swift integration is in
156
+ [FluidInference/FluidAudio](https://github.com/FluidInference/FluidAudio).
157
+
158
+ ### Bucket routing
159
+
160
+ Round each variable-length input up to the next bucket. Pad with zeros.
161
+
162
+ | Input | Axis | Buckets |
163
+ |---|---|---|
164
+ | text_predictor `tokens` | T_tok | 32 / 64 / 128 / 256 / 512 |
165
+ | diffusion_step `embedding` | T_bert | 512 only (pad) |
166
+ | decoder `asr` | T_mel | 256 / 512 / 1024 / 2048 / 4096 |
167
+
168
+ f0n_energy is shape-flexible.
169
+
170
+ ## Conversion provenance
171
+
172
+ - **Upstream code:** [yl4579/StyleTTS2](https://github.com/yl4579/StyleTTS2)
173
+ - **Upstream weights:** [yl4579/StyleTTS2-LibriTTS](https://huggingface.co/yl4579/StyleTTS2-LibriTTS),
174
+ file `Models/LibriTTS/epochs_2nd_00020.pth`
175
+ - **Conversion scripts:** [FluidInference/mobius PR #46](https://github.com/FluidInference/mobius/pull/46)
176
+ (`models/tts/styletts2/scripts/`)
177
+ - **Quantization:** `coremltools.optimize.coreml.linear_quantize_weights`,
178
+ `mode=linear_symmetric`, `dtype=int8`, `granularity=per_channel`,
179
+ `weight_threshold=200_000`
180
+ - **Target:** `coremltools` ≥ 8.0, `minimum_deployment_target=iOS17`
181
+ (macOS 14+ / iOS 17+)
182
+
183
+ ## Known limitations
184
+
185
+ - **English (LibriTTS) only.** No multilingual support in this
186
+ checkpoint.
187
+ - **HiFi-GAN decoder, not iSTFTNet.** LibriTTS upstream uses HiFi-GAN, so
188
+ no `torch.stft` / complex tensors in the conversion path.
189
+ - **Decoder is fp32, not fp16.** Documented above. The mlpackage size
190
+ reflects this (≈210 MB per bucket).
191
+ - **Voice-clone fidelity ceiling is architectural.** ECAPA-TDNN cosine
192
+ to reference clip ≈ 0.18 here, ≈ 0.29 in PyTorch fp32. The same-speaker
193
+ threshold is ~0.30. This isn't a quantization or conversion artifact;
194
+ see PR #46 TRIALS.md Phase 5.
195
+ - **No streaming.** Whole utterance only. Add chunked streaming on the
196
+ host side if you need it.
197
+
198
+ ## Citation & acknowledgments
199
+
200
+ - Yinghao Aaron Li et al. — StyleTTS2 architecture and LibriTTS
201
+ checkpoint.
202
+ - LibriTTS authors (CC-BY-4.0 training data).
203
+ - espeak-ng — phonemization frontend.
204
+
205
+ ```bibtex
206
+ @inproceedings{li2023styletts2,
207
+ title = {StyleTTS 2: Towards Human-Level Text-to-Speech through Style Diffusion and Adversarial Training with Large Speech Language Models},
208
+ author = {Li, Yinghao Aaron and Han, Cong and Raghavan, Vinay and Mischler, Gavin and Mesgarani, Nima},
209
+ booktitle = {NeurIPS},
210
+ year = {2023}
211
+ }
212
+ ```
compiled/.DS_Store ADDED
Binary file (14.3 kB). View file
 
compiled/styletts2_decoder_1024.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c8d4484ec862aee189fe8f7b84fd9e74ad87af7f903b5e3022491f0819b6242
3
+ size 243
compiled/styletts2_decoder_1024.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:176b03f7aeb6643777e10d07e26748bc532eb92756ee547a1756b72bdd7563bf
3
+ size 447
compiled/styletts2_decoder_1024.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 HiFi-GAN decoder (LibriTTS) @ T_mel=1024. 24 kHz, hop=300.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float32",
10
+ "formattedType" : "MultiArray (Float32 1 × 1 × 614400)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 614400]",
13
+ "name" : "waveform",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float32",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.greater" : 1,
24
+ "UpsampleNearestNeighbor" : 2,
25
+ "Ios17.floor" : 1,
26
+ "Ios17.instanceNorm" : 98,
27
+ "Ios17.reshape" : 110,
28
+ "Ios17.transpose" : 8,
29
+ "Ios17.sin" : 103,
30
+ "Split" : 106,
31
+ "Ios17.convTranspose" : 5,
32
+ "Ios17.expandDims" : 4,
33
+ "Ios17.add" : 381,
34
+ "Tile" : 2,
35
+ "Ios17.sliceByIndex" : 6,
36
+ "Ios17.squeeze" : 2,
37
+ "Ios17.conv" : 119,
38
+ "Ios17.sub" : 2,
39
+ "Ios17.cast" : 1,
40
+ "Ios17.tanh" : 2,
41
+ "Ios17.pow" : 101,
42
+ "Ios17.linear" : 107,
43
+ "Ios17.leakyRelu" : 10,
44
+ "Ios17.concat" : 6,
45
+ "Ios16.cumsum" : 1,
46
+ "Ios17.mul" : 331
47
+ },
48
+ "computePrecision" : "Mixed (Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 512 × 1024)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 512, 1024]",
72
+ "name" : "asr",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1 × 2048)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 2048]",
82
+ "name" : "F0_curve",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 2048)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 2048]",
92
+ "name" : "N",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 128)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 128]",
102
+ "name" : "s",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source_dialect" : "TorchScript",
109
+ "com.github.apple.coremltools.source" : "torch==2.11.0"
110
+ },
111
+ "generatedClassName" : "styletts2_decoder_1024",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_decoder_1024.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_decoder_1024.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceacbdc62672599460d67312ac7ad25ea1cbf62ec5f251d54392b191f3003853
3
+ size 222055552
compiled/styletts2_decoder_2048.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f876eefb7dca5fd52ad07e040486cb103e7e926ffcdb23c114ca6c41493738e
3
+ size 243
compiled/styletts2_decoder_2048.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1078e4402384405983e7bafa9f6016d816fcaa846dac7deba8dbe8275a950884
3
+ size 447
compiled/styletts2_decoder_2048.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 HiFi-GAN decoder (LibriTTS) @ T_mel=2048. 24 kHz, hop=300.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float32",
10
+ "formattedType" : "MultiArray (Float32 1 × 1 × 1228800)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 1228800]",
13
+ "name" : "waveform",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float32",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.greater" : 1,
24
+ "UpsampleNearestNeighbor" : 2,
25
+ "Ios17.floor" : 1,
26
+ "Ios17.instanceNorm" : 98,
27
+ "Ios17.reshape" : 110,
28
+ "Ios17.transpose" : 8,
29
+ "Ios17.sin" : 103,
30
+ "Split" : 106,
31
+ "Ios17.convTranspose" : 5,
32
+ "Ios17.expandDims" : 4,
33
+ "Ios17.add" : 381,
34
+ "Tile" : 2,
35
+ "Ios17.sliceByIndex" : 6,
36
+ "Ios17.squeeze" : 2,
37
+ "Ios17.conv" : 119,
38
+ "Ios17.sub" : 2,
39
+ "Ios17.cast" : 1,
40
+ "Ios17.tanh" : 2,
41
+ "Ios17.pow" : 101,
42
+ "Ios17.linear" : 107,
43
+ "Ios17.leakyRelu" : 10,
44
+ "Ios17.concat" : 6,
45
+ "Ios16.cumsum" : 1,
46
+ "Ios17.mul" : 331
47
+ },
48
+ "computePrecision" : "Mixed (Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 512 × 2048)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 512, 2048]",
72
+ "name" : "asr",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1 × 4096)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 4096]",
82
+ "name" : "F0_curve",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 4096)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 4096]",
92
+ "name" : "N",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 128)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 128]",
102
+ "name" : "s",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
109
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
110
+ },
111
+ "generatedClassName" : "styletts2_decoder_2048",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_decoder_2048.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_decoder_2048.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d099eb64c9b697e596e97f0c54077e104da9d690c302576ecb624083377b75eb
3
+ size 226970752
compiled/styletts2_decoder_256.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c2984630321f205643a9417db8ed42db88d0ec2e038daa0402543a1d01aec02
3
+ size 243
compiled/styletts2_decoder_256.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822f4f79fcf843f3e6d1674a9a6870f62fca1a4c1daa3f4297d6c52a01c99809
3
+ size 446
compiled/styletts2_decoder_256.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 HiFi-GAN decoder (LibriTTS) @ T_mel=256. 24 kHz, hop=300.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float32",
10
+ "formattedType" : "MultiArray (Float32 1 × 1 × 153600)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 153600]",
13
+ "name" : "waveform",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float32",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.greater" : 1,
24
+ "UpsampleNearestNeighbor" : 2,
25
+ "Ios17.floor" : 1,
26
+ "Ios17.instanceNorm" : 98,
27
+ "Ios17.reshape" : 110,
28
+ "Ios17.transpose" : 8,
29
+ "Ios17.sin" : 103,
30
+ "Split" : 106,
31
+ "Ios17.convTranspose" : 5,
32
+ "Ios17.expandDims" : 4,
33
+ "Ios17.add" : 381,
34
+ "Tile" : 2,
35
+ "Ios17.sliceByIndex" : 6,
36
+ "Ios17.squeeze" : 2,
37
+ "Ios17.conv" : 119,
38
+ "Ios17.sub" : 2,
39
+ "Ios17.cast" : 1,
40
+ "Ios17.tanh" : 2,
41
+ "Ios17.pow" : 101,
42
+ "Ios17.linear" : 107,
43
+ "Ios17.leakyRelu" : 10,
44
+ "Ios17.concat" : 6,
45
+ "Ios16.cumsum" : 1,
46
+ "Ios17.mul" : 331
47
+ },
48
+ "computePrecision" : "Mixed (Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 512 × 256)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 512, 256]",
72
+ "name" : "asr",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1 × 512)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 512]",
82
+ "name" : "F0_curve",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 512)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 512]",
92
+ "name" : "N",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 128)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 128]",
102
+ "name" : "s",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
109
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
110
+ },
111
+ "generatedClassName" : "styletts2_decoder_256",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_decoder_256.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_decoder_256.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c21c31ce508196d7314e00dfac02b545d09a1961b529cbbfe98c77d0109b003b
3
+ size 218369152
compiled/styletts2_decoder_4096.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d33d6ef2159363ecb4586b70214f62159fc6871fe92ff7060fdd3614811e6abe
3
+ size 243
compiled/styletts2_decoder_4096.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20c4e04c0d1558b8a249986d2553654a434646c06a2d565302b360d1aebe0a24
3
+ size 448
compiled/styletts2_decoder_4096.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 HiFi-GAN decoder (LibriTTS) @ T_mel=4096. 24 kHz, hop=300.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float32",
10
+ "formattedType" : "MultiArray (Float32 1 × 1 × 2457600)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 2457600]",
13
+ "name" : "waveform",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float32",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.greater" : 1,
24
+ "UpsampleNearestNeighbor" : 2,
25
+ "Ios17.floor" : 1,
26
+ "Ios17.instanceNorm" : 98,
27
+ "Ios17.reshape" : 110,
28
+ "Ios17.transpose" : 8,
29
+ "Ios17.sin" : 103,
30
+ "Split" : 106,
31
+ "Ios17.convTranspose" : 5,
32
+ "Ios17.expandDims" : 4,
33
+ "Ios17.add" : 381,
34
+ "Tile" : 2,
35
+ "Ios17.sliceByIndex" : 6,
36
+ "Ios17.squeeze" : 2,
37
+ "Ios17.conv" : 119,
38
+ "Ios17.sub" : 2,
39
+ "Ios17.cast" : 1,
40
+ "Ios17.tanh" : 2,
41
+ "Ios17.pow" : 101,
42
+ "Ios17.linear" : 107,
43
+ "Ios17.leakyRelu" : 10,
44
+ "Ios17.concat" : 6,
45
+ "Ios16.cumsum" : 1,
46
+ "Ios17.mul" : 331
47
+ },
48
+ "computePrecision" : "Mixed (Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 512 × 4096)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 512, 4096]",
72
+ "name" : "asr",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1 × 8192)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 8192]",
82
+ "name" : "F0_curve",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 8192)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 8192]",
92
+ "name" : "N",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 128)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 128]",
102
+ "name" : "s",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
109
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
110
+ },
111
+ "generatedClassName" : "styletts2_decoder_4096",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_decoder_4096.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_decoder_4096.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e2825d97b280217d84f597c085dc5c33328c736fb5b5d63f288209ae612b4f7
3
+ size 236801152
compiled/styletts2_decoder_512.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e42c6710d2a1441e67bef87fda6ffb6af6f1feb6b06f1585a71b99086b550a0e
3
+ size 243
compiled/styletts2_decoder_512.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6abfad25a5df29fdcc4e135bcbd614a9480355904e58b24a33afae72c03bcc9a
3
+ size 446
compiled/styletts2_decoder_512.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 HiFi-GAN decoder (LibriTTS) @ T_mel=512. 24 kHz, hop=300.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float32",
10
+ "formattedType" : "MultiArray (Float32 1 × 1 × 307200)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 307200]",
13
+ "name" : "waveform",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float32",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.greater" : 1,
24
+ "UpsampleNearestNeighbor" : 2,
25
+ "Ios17.floor" : 1,
26
+ "Ios17.instanceNorm" : 98,
27
+ "Ios17.reshape" : 110,
28
+ "Ios17.transpose" : 8,
29
+ "Ios17.sin" : 103,
30
+ "Split" : 106,
31
+ "Ios17.convTranspose" : 5,
32
+ "Ios17.expandDims" : 4,
33
+ "Ios17.add" : 381,
34
+ "Tile" : 2,
35
+ "Ios17.sliceByIndex" : 6,
36
+ "Ios17.squeeze" : 2,
37
+ "Ios17.conv" : 119,
38
+ "Ios17.sub" : 2,
39
+ "Ios17.cast" : 1,
40
+ "Ios17.tanh" : 2,
41
+ "Ios17.pow" : 101,
42
+ "Ios17.linear" : 107,
43
+ "Ios17.leakyRelu" : 10,
44
+ "Ios17.concat" : 6,
45
+ "Ios16.cumsum" : 1,
46
+ "Ios17.mul" : 331
47
+ },
48
+ "computePrecision" : "Mixed (Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 512 × 512)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 512, 512]",
72
+ "name" : "asr",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1 × 1024)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 1024]",
82
+ "name" : "F0_curve",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 1024)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 1024]",
92
+ "name" : "N",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 128)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 128]",
102
+ "name" : "s",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
109
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
110
+ },
111
+ "generatedClassName" : "styletts2_decoder_512",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_decoder_512.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_decoder_512.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eaca30dac2dce98bb9341edf9739915b3f3e73ddfaa907ae523c89946dc2964
3
+ size 219597952
compiled/styletts2_diffusion_step_512.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd74ef01b9d41584bb2f2da6e5957336afff35a91f1772f83e1e8ff757928fea
3
+ size 243
compiled/styletts2_diffusion_step_512.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:686dbb5576a97eb595764e9bcb9e46aa2d560c67502a4ddf1afc683886bd1e53
3
+ size 467
compiled/styletts2_diffusion_step_512.mlmodelc/metadata.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 KDiffusion single denoise step (StyleTransformer1d) @ T_tok=512.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 1 × 256)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 1, 256]",
13
+ "name" : "denoised",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "storagePrecision" : "Float16",
18
+ "modelParameters" : [
19
+
20
+ ],
21
+ "specificationVersion" : 8,
22
+ "mlProgramOperationTypeHistogram" : {
23
+ "Ios17.cos" : 1,
24
+ "Ios17.reshape" : 20,
25
+ "Ios17.inverse" : 2,
26
+ "Ios16.softmax" : 3,
27
+ "Ios17.matmul" : 6,
28
+ "Ios17.sin" : 1,
29
+ "Ios17.transpose" : 27,
30
+ "Split" : 9,
31
+ "Ios16.reduceMean" : 1,
32
+ "Ios17.expandDims" : 2,
33
+ "Ios17.add" : 23,
34
+ "Tile" : 2,
35
+ "Ios16.reduceSum" : 1,
36
+ "Ios17.layerNorm" : 3,
37
+ "Ios17.log" : 1,
38
+ "Ios17.sqrt" : 1,
39
+ "Ios16.gelu" : 7,
40
+ "Ios17.cast" : 4,
41
+ "Ios17.conv" : 1,
42
+ "Ios17.realDiv" : 1,
43
+ "Ios17.linear" : 25,
44
+ "Stack" : 1,
45
+ "Ios17.concat" : 2,
46
+ "Ios17.mul" : 19
47
+ },
48
+ "computePrecision" : "Mixed (Float16, Float32, Int32)",
49
+ "isUpdatable" : "0",
50
+ "stateSchema" : [
51
+
52
+ ],
53
+ "availability" : {
54
+ "macOS" : "14.0",
55
+ "tvOS" : "17.0",
56
+ "visionOS" : "1.0",
57
+ "watchOS" : "10.0",
58
+ "iOS" : "17.0",
59
+ "macCatalyst" : "17.0"
60
+ },
61
+ "modelType" : {
62
+ "name" : "MLModelType_mlProgram"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float32",
69
+ "formattedType" : "MultiArray (Float32 1 × 1 × 256)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 1, 256]",
72
+ "name" : "x_noisy",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float32",
79
+ "formattedType" : "MultiArray (Float32 1)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1]",
82
+ "name" : "sigma",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float32",
89
+ "formattedType" : "MultiArray (Float32 1 × 512 × 768)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 512, 768]",
92
+ "name" : "embedding",
93
+ "type" : "MultiArray"
94
+ },
95
+ {
96
+ "hasShapeFlexibility" : "0",
97
+ "isOptional" : "0",
98
+ "dataType" : "Float32",
99
+ "formattedType" : "MultiArray (Float32 1 × 256)",
100
+ "shortDescription" : "",
101
+ "shape" : "[1, 256]",
102
+ "name" : "features",
103
+ "type" : "MultiArray"
104
+ }
105
+ ],
106
+ "userDefinedMetadata" : {
107
+ "com.github.apple.coremltools.version" : "8.3.0",
108
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
109
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
110
+ },
111
+ "generatedClassName" : "styletts2_diffusion_step_512",
112
+ "method" : "predict"
113
+ }
114
+ ]
compiled/styletts2_diffusion_step_512.mlmodelc/model.mil ADDED
@@ -0,0 +1,351 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.0)
2
+ [buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "3520.4.1"}, {"coremlc-version", "3520.5.1"}, {"coremltools-component-torch", "2.11.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})]
3
+ {
4
+ func main<ios17>(tensor<fp32, [1, 512, 768]> embedding, tensor<fp32, [1, 256]> features, tensor<fp32, [1]> sigma, tensor<fp32, [1, 1, 256]> x_noisy) {
5
+ tensor<int32, [3]> var_25 = const()[name = tensor<string, []>("op_25"), val = tensor<int32, [3]>([-1, 1, 1])];
6
+ tensor<string, []> sigma_to_fp16_dtype_0 = const()[name = tensor<string, []>("sigma_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
7
+ tensor<fp16, [1]> sigma_to_fp16 = cast(dtype = sigma_to_fp16_dtype_0, x = sigma)[name = tensor<string, []>("cast_27")];
8
+ tensor<fp16, [1, 1, 1]> s_cast_fp16 = reshape(shape = var_25, x = sigma_to_fp16)[name = tensor<string, []>("s_cast_fp16")];
9
+ tensor<fp16, [1, 1, 1]> var_27_cast_fp16 = mul(x = s_cast_fp16, y = s_cast_fp16)[name = tensor<string, []>("op_27_cast_fp16")];
10
+ tensor<fp16, []> var_29_to_fp16 = const()[name = tensor<string, []>("op_29_to_fp16"), val = tensor<fp16, []>(0x1.47cp-5)];
11
+ tensor<fp16, [1, 1, 1]> var_30_cast_fp16 = add(x = var_27_cast_fp16, y = var_29_to_fp16)[name = tensor<string, []>("op_30_cast_fp16")];
12
+ tensor<fp32, []> var_31_epsilon_0 = const()[name = tensor<string, []>("op_31_epsilon_0"), val = tensor<fp32, []>(0x1.a36e2ep-14)];
13
+ tensor<fp16, [1, 1, 1]> var_31_cast_fp16 = inverse(epsilon = var_31_epsilon_0, x = var_30_cast_fp16)[name = tensor<string, []>("op_31_cast_fp16")];
14
+ tensor<fp16, []> var_32_to_fp16 = const()[name = tensor<string, []>("op_32_to_fp16"), val = tensor<fp16, []>(0x1.47cp-5)];
15
+ tensor<fp16, [1, 1, 1]> c_skip_cast_fp16 = mul(x = var_31_cast_fp16, y = var_32_to_fp16)[name = tensor<string, []>("c_skip_cast_fp16")];
16
+ tensor<fp16, []> var_34_to_fp16 = const()[name = tensor<string, []>("op_34_to_fp16"), val = tensor<fp16, []>(0x1.998p-3)];
17
+ tensor<fp16, [1, 1, 1]> var_35_cast_fp16 = mul(x = s_cast_fp16, y = var_34_to_fp16)[name = tensor<string, []>("op_35_cast_fp16")];
18
+ tensor<fp16, [1, 1, 1]> var_40_cast_fp16 = sqrt(x = var_30_cast_fp16)[name = tensor<string, []>("op_40_cast_fp16")];
19
+ tensor<fp16, [1, 1, 1]> c_out_cast_fp16 = real_div(x = var_35_cast_fp16, y = var_40_cast_fp16)[name = tensor<string, []>("c_out_cast_fp16")];
20
+ tensor<fp32, []> var_47_epsilon_0 = const()[name = tensor<string, []>("op_47_epsilon_0"), val = tensor<fp32, []>(0x1.a36e2ep-14)];
21
+ tensor<fp16, [1, 1, 1]> var_47_cast_fp16 = inverse(epsilon = var_47_epsilon_0, x = var_40_cast_fp16)[name = tensor<string, []>("op_47_cast_fp16")];
22
+ tensor<fp32, []> var_50_epsilon_0 = const()[name = tensor<string, []>("op_50_epsilon_0"), val = tensor<fp32, []>(0x1p-149)];
23
+ tensor<fp16, [1]> var_50_cast_fp16 = log(epsilon = var_50_epsilon_0, x = sigma_to_fp16)[name = tensor<string, []>("op_50_cast_fp16")];
24
+ tensor<fp16, []> var_51_to_fp16 = const()[name = tensor<string, []>("op_51_to_fp16"), val = tensor<fp16, []>(0x1p-2)];
25
+ tensor<fp16, [1]> x_1_cast_fp16 = mul(x = var_50_cast_fp16, y = var_51_to_fp16)[name = tensor<string, []>("x_1_cast_fp16")];
26
+ tensor<string, []> x_noisy_to_fp16_dtype_0 = const()[name = tensor<string, []>("x_noisy_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
27
+ tensor<fp16, [1, 1, 256]> x_noisy_to_fp16 = cast(dtype = x_noisy_to_fp16_dtype_0, x = x_noisy)[name = tensor<string, []>("cast_26")];
28
+ tensor<fp16, [1, 1, 256]> x_11_cast_fp16 = mul(x = var_47_cast_fp16, y = x_noisy_to_fp16)[name = tensor<string, []>("x_11_cast_fp16")];
29
+ tensor<int32, []> var_55 = const()[name = tensor<string, []>("op_55"), val = tensor<int32, []>(-1)];
30
+ tensor<int32, [2]> var_67 = const()[name = tensor<string, []>("op_67"), val = tensor<int32, [2]>([1, 1])];
31
+ tensor<fp16, [1, 1]> x_5_cast_fp16 = reshape(shape = var_67, x = x_1_cast_fp16)[name = tensor<string, []>("x_5_cast_fp16")];
32
+ tensor<fp16, [1, 128]> var_75_to_fp16 = const()[name = tensor<string, []>("op_75_to_fp16"), val = tensor<fp16, [1, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
33
+ tensor<fp16, [1, 128]> var_76_cast_fp16 = mul(x = x_5_cast_fp16, y = var_75_to_fp16)[name = tensor<string, []>("op_76_cast_fp16")];
34
+ tensor<fp16, []> var_77_promoted_to_fp16 = const()[name = tensor<string, []>("op_77_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+1)];
35
+ tensor<fp16, [1, 128]> var_78_cast_fp16 = mul(x = var_76_cast_fp16, y = var_77_promoted_to_fp16)[name = tensor<string, []>("op_78_cast_fp16")];
36
+ tensor<fp16, []> var_79_to_fp16 = const()[name = tensor<string, []>("op_79_to_fp16"), val = tensor<fp16, []>(0x1.92p+1)];
37
+ tensor<fp16, [1, 128]> freqs_cast_fp16 = mul(x = var_78_cast_fp16, y = var_79_to_fp16)[name = tensor<string, []>("freqs_cast_fp16")];
38
+ tensor<fp16, [1, 128]> var_81_cast_fp16 = sin(x = freqs_cast_fp16)[name = tensor<string, []>("op_81_cast_fp16")];
39
+ tensor<fp16, [1, 128]> var_82_cast_fp16 = cos(x = freqs_cast_fp16)[name = tensor<string, []>("op_82_cast_fp16")];
40
+ tensor<bool, []> input_1_interleave_0 = const()[name = tensor<string, []>("input_1_interleave_0"), val = tensor<bool, []>(false)];
41
+ tensor<fp16, [1, 257]> input_1_cast_fp16 = concat(axis = var_55, interleave = input_1_interleave_0, values = (x_5_cast_fp16, var_81_cast_fp16, var_82_cast_fp16))[name = tensor<string, []>("input_1_cast_fp16")];
42
+ tensor<fp16, [1024, 257]> transformer_to_time_0_1_weight_to_fp16 = const()[name = tensor<string, []>("transformer_to_time_0_1_weight_to_fp16"), val = tensor<fp16, [1024, 257]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(384)))];
43
+ tensor<fp16, [1024]> transformer_to_time_0_1_bias_to_fp16 = const()[name = tensor<string, []>("transformer_to_time_0_1_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(526784)))];
44
+ tensor<fp16, [1, 1024]> linear_0_cast_fp16 = linear(bias = transformer_to_time_0_1_bias_to_fp16, weight = transformer_to_time_0_1_weight_to_fp16, x = input_1_cast_fp16)[name = tensor<string, []>("linear_0_cast_fp16")];
45
+ tensor<string, []> var_88_mode_0 = const()[name = tensor<string, []>("op_88_mode_0"), val = tensor<string, []>("EXACT")];
46
+ tensor<fp16, [1, 1024]> var_88_cast_fp16 = gelu(mode = var_88_mode_0, x = linear_0_cast_fp16)[name = tensor<string, []>("op_88_cast_fp16")];
47
+ tensor<string, []> features_to_fp16_dtype_0 = const()[name = tensor<string, []>("features_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
48
+ tensor<fp16, [1024, 256]> transformer_to_features_0_weight_to_fp16 = const()[name = tensor<string, []>("transformer_to_features_0_weight_to_fp16"), val = tensor<fp16, [1024, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(528896)))];
49
+ tensor<fp16, [1024]> transformer_to_features_0_bias_to_fp16 = const()[name = tensor<string, []>("transformer_to_features_0_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1053248)))];
50
+ tensor<fp16, [1, 256]> features_to_fp16 = cast(dtype = features_to_fp16_dtype_0, x = features)[name = tensor<string, []>("cast_25")];
51
+ tensor<fp16, [1, 1024]> linear_1_cast_fp16 = linear(bias = transformer_to_features_0_bias_to_fp16, weight = transformer_to_features_0_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_1_cast_fp16")];
52
+ tensor<string, []> var_94_mode_0 = const()[name = tensor<string, []>("op_94_mode_0"), val = tensor<string, []>("EXACT")];
53
+ tensor<fp16, [1, 1024]> var_94_cast_fp16 = gelu(mode = var_94_mode_0, x = linear_1_cast_fp16)[name = tensor<string, []>("op_94_cast_fp16")];
54
+ tensor<int32, []> x_7_axis_0 = const()[name = tensor<string, []>("x_7_axis_0"), val = tensor<int32, []>(0)];
55
+ tensor<fp16, [2, 1, 1024]> x_7_cast_fp16 = stack(axis = x_7_axis_0, values = (var_88_cast_fp16, var_94_cast_fp16))[name = tensor<string, []>("x_7_cast_fp16")];
56
+ tensor<int32, [3]> var_101 = const()[name = tensor<string, []>("op_101"), val = tensor<int32, [3]>([1, 2, 0])];
57
+ tensor<int32, [1]> input_7_axes_0 = const()[name = tensor<string, []>("input_7_axes_0"), val = tensor<int32, [1]>([2])];
58
+ tensor<bool, []> input_7_keep_dims_0 = const()[name = tensor<string, []>("input_7_keep_dims_0"), val = tensor<bool, []>(false)];
59
+ tensor<fp16, [1, 1024, 2]> x_9_cast_fp16 = transpose(perm = var_101, x = x_7_cast_fp16)[name = tensor<string, []>("transpose_41")];
60
+ tensor<fp16, [1, 1024]> input_7_cast_fp16 = reduce_sum(axes = input_7_axes_0, keep_dims = input_7_keep_dims_0, x = x_9_cast_fp16)[name = tensor<string, []>("input_7_cast_fp16")];
61
+ tensor<fp16, [1024, 1024]> transformer_to_mapping_0_weight_to_fp16 = const()[name = tensor<string, []>("transformer_to_mapping_0_weight_to_fp16"), val = tensor<fp16, [1024, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1055360)))];
62
+ tensor<fp16, [1024]> transformer_to_mapping_0_bias_to_fp16 = const()[name = tensor<string, []>("transformer_to_mapping_0_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3152576)))];
63
+ tensor<fp16, [1, 1024]> linear_2_cast_fp16 = linear(bias = transformer_to_mapping_0_bias_to_fp16, weight = transformer_to_mapping_0_weight_to_fp16, x = input_7_cast_fp16)[name = tensor<string, []>("linear_2_cast_fp16")];
64
+ tensor<string, []> input_11_mode_0 = const()[name = tensor<string, []>("input_11_mode_0"), val = tensor<string, []>("EXACT")];
65
+ tensor<fp16, [1, 1024]> input_11_cast_fp16 = gelu(mode = input_11_mode_0, x = linear_2_cast_fp16)[name = tensor<string, []>("input_11_cast_fp16")];
66
+ tensor<fp16, [1024, 1024]> transformer_to_mapping_2_weight_to_fp16 = const()[name = tensor<string, []>("transformer_to_mapping_2_weight_to_fp16"), val = tensor<fp16, [1024, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3154688)))];
67
+ tensor<fp16, [1024]> transformer_to_mapping_2_bias_to_fp16 = const()[name = tensor<string, []>("transformer_to_mapping_2_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5251904)))];
68
+ tensor<fp16, [1, 1024]> linear_3_cast_fp16 = linear(bias = transformer_to_mapping_2_bias_to_fp16, weight = transformer_to_mapping_2_weight_to_fp16, x = input_11_cast_fp16)[name = tensor<string, []>("linear_3_cast_fp16")];
69
+ tensor<string, []> mapping_1_mode_0 = const()[name = tensor<string, []>("mapping_1_mode_0"), val = tensor<string, []>("EXACT")];
70
+ tensor<fp16, [1, 1024]> mapping_1_cast_fp16 = gelu(mode = mapping_1_mode_0, x = linear_3_cast_fp16)[name = tensor<string, []>("mapping_1_cast_fp16")];
71
+ tensor<int32, [3]> var_127_reps_0 = const()[name = tensor<string, []>("op_127_reps_0"), val = tensor<int32, [3]>([1, 512, 1])];
72
+ tensor<fp16, [1, 512, 256]> var_127_cast_fp16 = tile(reps = var_127_reps_0, x = x_11_cast_fp16)[name = tensor<string, []>("op_127_cast_fp16")];
73
+ tensor<int32, []> var_129 = const()[name = tensor<string, []>("op_129"), val = tensor<int32, []>(-1)];
74
+ tensor<bool, []> x_13_interleave_0 = const()[name = tensor<string, []>("x_13_interleave_0"), val = tensor<bool, []>(false)];
75
+ tensor<string, []> embedding_to_fp16_dtype_0 = const()[name = tensor<string, []>("embedding_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
76
+ tensor<fp16, [1, 512, 768]> embedding_to_fp16 = cast(dtype = embedding_to_fp16_dtype_0, x = embedding)[name = tensor<string, []>("cast_24")];
77
+ tensor<fp16, [1, 512, 1024]> x_13_cast_fp16 = concat(axis = var_129, interleave = x_13_interleave_0, values = (var_127_cast_fp16, embedding_to_fp16))[name = tensor<string, []>("x_13_cast_fp16")];
78
+ tensor<int32, [1]> var_132_axes_0 = const()[name = tensor<string, []>("op_132_axes_0"), val = tensor<int32, [1]>([1])];
79
+ tensor<fp16, [1, 1, 1024]> var_132_cast_fp16 = expand_dims(axes = var_132_axes_0, x = mapping_1_cast_fp16)[name = tensor<string, []>("op_132_cast_fp16")];
80
+ tensor<int32, [3]> mapping_reps_0 = const()[name = tensor<string, []>("mapping_reps_0"), val = tensor<int32, [3]>([1, 512, 1])];
81
+ tensor<fp16, [1, 512, 1024]> mapping_cast_fp16 = tile(reps = mapping_reps_0, x = var_132_cast_fp16)[name = tensor<string, []>("mapping_cast_fp16")];
82
+ tensor<fp16, [1, 512, 1024]> x_15_cast_fp16 = add(x = x_13_cast_fp16, y = mapping_cast_fp16)[name = tensor<string, []>("x_15_cast_fp16")];
83
+ tensor<int32, []> var_153 = const()[name = tensor<string, []>("op_153"), val = tensor<int32, []>(-1)];
84
+ tensor<fp16, [2048, 256]> transformer_blocks_0_attention_norm_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_norm_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5254016)))];
85
+ tensor<fp16, [2048]> transformer_blocks_0_attention_norm_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_norm_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(6302656)))];
86
+ tensor<fp16, [1, 2048]> linear_4_cast_fp16 = linear(bias = transformer_blocks_0_attention_norm_fc_bias_to_fp16, weight = transformer_blocks_0_attention_norm_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_4_cast_fp16")];
87
+ tensor<int32, [3]> var_172 = const()[name = tensor<string, []>("op_172"), val = tensor<int32, [3]>([1, 2048, 1])];
88
+ tensor<fp16, [1, 2048, 1]> h_3_cast_fp16 = reshape(shape = var_172, x = linear_4_cast_fp16)[name = tensor<string, []>("h_3_cast_fp16")];
89
+ tensor<int32, [2]> var_174_split_sizes_0 = const()[name = tensor<string, []>("op_174_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
90
+ tensor<int32, []> var_174_axis_0 = const()[name = tensor<string, []>("op_174_axis_0"), val = tensor<int32, []>(1)];
91
+ tensor<fp16, [1, 1024, 1]> var_174_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_174_cast_fp16_1 = split(axis = var_174_axis_0, split_sizes = var_174_split_sizes_0, x = h_3_cast_fp16)[name = tensor<string, []>("op_174_cast_fp16")];
92
+ tensor<int32, [3]> gamma_3_perm_0 = const()[name = tensor<string, []>("gamma_3_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
93
+ tensor<int32, [3]> beta_3_perm_0 = const()[name = tensor<string, []>("beta_3_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
94
+ tensor<int32, [1]> x_19_axes_0 = const()[name = tensor<string, []>("x_19_axes_0"), val = tensor<int32, [1]>([-1])];
95
+ tensor<fp16, []> var_146_to_fp16 = const()[name = tensor<string, []>("op_146_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
96
+ tensor<fp16, [1, 512, 1024]> x_19_cast_fp16 = layer_norm(axes = x_19_axes_0, epsilon = var_146_to_fp16, x = x_15_cast_fp16)[name = tensor<string, []>("x_19_cast_fp16")];
97
+ tensor<fp16, []> var_180_promoted_to_fp16 = const()[name = tensor<string, []>("op_180_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
98
+ tensor<fp16, [1, 1, 1024]> gamma_3_cast_fp16 = transpose(perm = gamma_3_perm_0, x = var_174_cast_fp16_0)[name = tensor<string, []>("transpose_40")];
99
+ tensor<fp16, [1, 1, 1024]> var_181_cast_fp16 = add(x = gamma_3_cast_fp16, y = var_180_promoted_to_fp16)[name = tensor<string, []>("op_181_cast_fp16")];
100
+ tensor<fp16, [1, 512, 1024]> var_182_cast_fp16 = mul(x = var_181_cast_fp16, y = x_19_cast_fp16)[name = tensor<string, []>("op_182_cast_fp16")];
101
+ tensor<fp16, [1, 1, 1024]> beta_3_cast_fp16 = transpose(perm = beta_3_perm_0, x = var_174_cast_fp16_1)[name = tensor<string, []>("transpose_39")];
102
+ tensor<fp16, [1, 512, 1024]> x_21_cast_fp16 = add(x = var_182_cast_fp16, y = beta_3_cast_fp16)[name = tensor<string, []>("x_21_cast_fp16")];
103
+ tensor<fp16, [2048, 256]> transformer_blocks_0_attention_norm_context_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_norm_context_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(6306816)))];
104
+ tensor<fp16, [2048]> transformer_blocks_0_attention_norm_context_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_norm_context_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7355456)))];
105
+ tensor<fp16, [1, 2048]> linear_5_cast_fp16 = linear(bias = transformer_blocks_0_attention_norm_context_fc_bias_to_fp16, weight = transformer_blocks_0_attention_norm_context_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_5_cast_fp16")];
106
+ tensor<int32, [3]> var_194 = const()[name = tensor<string, []>("op_194"), val = tensor<int32, [3]>([1, 2048, 1])];
107
+ tensor<fp16, [1, 2048, 1]> h_7_cast_fp16 = reshape(shape = var_194, x = linear_5_cast_fp16)[name = tensor<string, []>("h_7_cast_fp16")];
108
+ tensor<int32, [2]> var_196_split_sizes_0 = const()[name = tensor<string, []>("op_196_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
109
+ tensor<int32, []> var_196_axis_0 = const()[name = tensor<string, []>("op_196_axis_0"), val = tensor<int32, []>(1)];
110
+ tensor<fp16, [1, 1024, 1]> var_196_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_196_cast_fp16_1 = split(axis = var_196_axis_0, split_sizes = var_196_split_sizes_0, x = h_7_cast_fp16)[name = tensor<string, []>("op_196_cast_fp16")];
111
+ tensor<int32, [3]> gamma_7_perm_0 = const()[name = tensor<string, []>("gamma_7_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
112
+ tensor<int32, [3]> beta_7_perm_0 = const()[name = tensor<string, []>("beta_7_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
113
+ tensor<fp16, []> var_202_promoted_to_fp16 = const()[name = tensor<string, []>("op_202_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
114
+ tensor<fp16, [1, 1, 1024]> gamma_7_cast_fp16 = transpose(perm = gamma_7_perm_0, x = var_196_cast_fp16_0)[name = tensor<string, []>("transpose_38")];
115
+ tensor<fp16, [1, 1, 1024]> var_203_cast_fp16 = add(x = gamma_7_cast_fp16, y = var_202_promoted_to_fp16)[name = tensor<string, []>("op_203_cast_fp16")];
116
+ tensor<fp16, [1, 512, 1024]> var_204_cast_fp16 = mul(x = var_203_cast_fp16, y = x_19_cast_fp16)[name = tensor<string, []>("op_204_cast_fp16")];
117
+ tensor<fp16, [1, 1, 1024]> beta_7_cast_fp16 = transpose(perm = beta_7_perm_0, x = var_196_cast_fp16_1)[name = tensor<string, []>("transpose_37")];
118
+ tensor<fp16, [1, 512, 1024]> x_27_cast_fp16 = add(x = var_204_cast_fp16, y = beta_7_cast_fp16)[name = tensor<string, []>("x_27_cast_fp16")];
119
+ tensor<fp16, [512, 1024]> transformer_blocks_0_attention_to_q_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_to_q_weight_to_fp16"), val = tensor<fp16, [512, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7359616)))];
120
+ tensor<fp16, [512]> linear_6_bias_0_to_fp16 = const()[name = tensor<string, []>("linear_6_bias_0_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8408256)))];
121
+ tensor<fp16, [1, 512, 512]> linear_6_cast_fp16 = linear(bias = linear_6_bias_0_to_fp16, weight = transformer_blocks_0_attention_to_q_weight_to_fp16, x = x_21_cast_fp16)[name = tensor<string, []>("linear_6_cast_fp16")];
122
+ tensor<fp16, [1024, 1024]> transformer_blocks_0_attention_to_kv_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_to_kv_weight_to_fp16"), val = tensor<fp16, [1024, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8409344)))];
123
+ tensor<fp16, [1024]> linear_7_bias_0_to_fp16 = const()[name = tensor<string, []>("linear_7_bias_0_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10506560)))];
124
+ tensor<fp16, [1, 512, 1024]> linear_7_cast_fp16 = linear(bias = linear_7_bias_0_to_fp16, weight = transformer_blocks_0_attention_to_kv_weight_to_fp16, x = x_27_cast_fp16)[name = tensor<string, []>("linear_7_cast_fp16")];
125
+ tensor<int32, [2]> var_212_split_sizes_0 = const()[name = tensor<string, []>("op_212_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
126
+ tensor<int32, []> var_212_axis_0 = const()[name = tensor<string, []>("op_212_axis_0"), val = tensor<int32, []>(-1)];
127
+ tensor<fp16, [1, 512, 512]> var_212_cast_fp16_0, tensor<fp16, [1, 512, 512]> var_212_cast_fp16_1 = split(axis = var_212_axis_0, split_sizes = var_212_split_sizes_0, x = linear_7_cast_fp16)[name = tensor<string, []>("op_212_cast_fp16")];
128
+ tensor<int32, [4]> var_221 = const()[name = tensor<string, []>("op_221"), val = tensor<int32, [4]>([1, 512, 8, 64])];
129
+ tensor<fp16, [1, 512, 8, 64]> x_31_cast_fp16 = reshape(shape = var_221, x = linear_6_cast_fp16)[name = tensor<string, []>("x_31_cast_fp16")];
130
+ tensor<int32, [4]> var_231 = const()[name = tensor<string, []>("op_231"), val = tensor<int32, [4]>([1, 512, 8, 64])];
131
+ tensor<fp16, [1, 512, 8, 64]> x_35_cast_fp16 = reshape(shape = var_231, x = var_212_cast_fp16_0)[name = tensor<string, []>("x_35_cast_fp16")];
132
+ tensor<int32, [4]> var_241 = const()[name = tensor<string, []>("op_241"), val = tensor<int32, [4]>([1, 512, 8, 64])];
133
+ tensor<fp16, [1, 512, 8, 64]> x_39_cast_fp16 = reshape(shape = var_241, x = var_212_cast_fp16_1)[name = tensor<string, []>("x_39_cast_fp16")];
134
+ tensor<int32, [4]> var_243 = const()[name = tensor<string, []>("op_243"), val = tensor<int32, [4]>([0, 2, 1, 3])];
135
+ tensor<bool, []> sim_1_transpose_x_0 = const()[name = tensor<string, []>("sim_1_transpose_x_0"), val = tensor<bool, []>(false)];
136
+ tensor<bool, []> sim_1_transpose_y_0 = const()[name = tensor<string, []>("sim_1_transpose_y_0"), val = tensor<bool, []>(false)];
137
+ tensor<int32, [4]> transpose_9_perm_0 = const()[name = tensor<string, []>("transpose_9_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
138
+ tensor<int32, [4]> transpose_10_perm_0 = const()[name = tensor<string, []>("transpose_10_perm_0"), val = tensor<int32, [4]>([0, 2, -1, -3])];
139
+ tensor<fp16, [1, 8, 64, 512]> transpose_10 = transpose(perm = transpose_10_perm_0, x = x_35_cast_fp16)[name = tensor<string, []>("transpose_34")];
140
+ tensor<fp16, [1, 8, 512, 64]> transpose_9 = transpose(perm = transpose_9_perm_0, x = x_31_cast_fp16)[name = tensor<string, []>("transpose_35")];
141
+ tensor<fp16, [1, 8, 512, 512]> sim_1_cast_fp16 = matmul(transpose_x = sim_1_transpose_x_0, transpose_y = sim_1_transpose_y_0, x = transpose_9, y = transpose_10)[name = tensor<string, []>("sim_1_cast_fp16")];
142
+ tensor<fp16, []> var_247_to_fp16 = const()[name = tensor<string, []>("op_247_to_fp16"), val = tensor<fp16, []>(0x1p-3)];
143
+ tensor<fp16, [1, 8, 512, 512]> sim_3_cast_fp16 = mul(x = sim_1_cast_fp16, y = var_247_to_fp16)[name = tensor<string, []>("sim_3_cast_fp16")];
144
+ tensor<fp16, [1, 8, 512, 512]> attn_1_cast_fp16 = softmax(axis = var_153, x = sim_3_cast_fp16)[name = tensor<string, []>("attn_1_cast_fp16")];
145
+ tensor<bool, []> x_41_transpose_x_0 = const()[name = tensor<string, []>("x_41_transpose_x_0"), val = tensor<bool, []>(false)];
146
+ tensor<bool, []> x_41_transpose_y_0 = const()[name = tensor<string, []>("x_41_transpose_y_0"), val = tensor<bool, []>(false)];
147
+ tensor<fp16, [1, 8, 512, 64]> v_1_cast_fp16 = transpose(perm = var_243, x = x_39_cast_fp16)[name = tensor<string, []>("transpose_36")];
148
+ tensor<fp16, [1, 8, 512, 64]> x_41_cast_fp16 = matmul(transpose_x = x_41_transpose_x_0, transpose_y = x_41_transpose_y_0, x = attn_1_cast_fp16, y = v_1_cast_fp16)[name = tensor<string, []>("x_41_cast_fp16")];
149
+ tensor<int32, [4]> var_269 = const()[name = tensor<string, []>("op_269"), val = tensor<int32, [4]>([0, 2, 1, 3])];
150
+ tensor<int32, [3]> var_271 = const()[name = tensor<string, []>("op_271"), val = tensor<int32, [3]>([1, 512, 512])];
151
+ tensor<fp16, [1, 512, 8, 64]> x_43_cast_fp16 = transpose(perm = var_269, x = x_41_cast_fp16)[name = tensor<string, []>("transpose_33")];
152
+ tensor<fp16, [1, 512, 512]> input_23_cast_fp16 = reshape(shape = var_271, x = x_43_cast_fp16)[name = tensor<string, []>("input_23_cast_fp16")];
153
+ tensor<fp16, [1024, 512]> transformer_blocks_0_attention_attention_to_out_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_attention_to_out_weight_to_fp16"), val = tensor<fp16, [1024, 512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10508672)))];
154
+ tensor<fp16, [1024]> transformer_blocks_0_attention_attention_to_out_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_attention_attention_to_out_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(11557312)))];
155
+ tensor<fp16, [1, 512, 1024]> linear_8_cast_fp16 = linear(bias = transformer_blocks_0_attention_attention_to_out_bias_to_fp16, weight = transformer_blocks_0_attention_attention_to_out_weight_to_fp16, x = input_23_cast_fp16)[name = tensor<string, []>("linear_8_cast_fp16")];
156
+ tensor<fp16, [1, 512, 1024]> input_25_cast_fp16 = add(x = linear_8_cast_fp16, y = x_15_cast_fp16)[name = tensor<string, []>("input_25_cast_fp16")];
157
+ tensor<fp16, [2048, 1024]> transformer_blocks_0_feed_forward_0_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_feed_forward_0_weight_to_fp16"), val = tensor<fp16, [2048, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(11559424)))];
158
+ tensor<fp16, [2048]> transformer_blocks_0_feed_forward_0_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_feed_forward_0_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15753792)))];
159
+ tensor<fp16, [1, 512, 2048]> linear_9_cast_fp16 = linear(bias = transformer_blocks_0_feed_forward_0_bias_to_fp16, weight = transformer_blocks_0_feed_forward_0_weight_to_fp16, x = input_25_cast_fp16)[name = tensor<string, []>("linear_9_cast_fp16")];
160
+ tensor<string, []> input_29_mode_0 = const()[name = tensor<string, []>("input_29_mode_0"), val = tensor<string, []>("EXACT")];
161
+ tensor<fp16, [1, 512, 2048]> input_29_cast_fp16 = gelu(mode = input_29_mode_0, x = linear_9_cast_fp16)[name = tensor<string, []>("input_29_cast_fp16")];
162
+ tensor<fp16, [1024, 2048]> transformer_blocks_0_feed_forward_2_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_feed_forward_2_weight_to_fp16"), val = tensor<fp16, [1024, 2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15757952)))];
163
+ tensor<fp16, [1024]> transformer_blocks_0_feed_forward_2_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_0_feed_forward_2_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(19952320)))];
164
+ tensor<fp16, [1, 512, 1024]> linear_10_cast_fp16 = linear(bias = transformer_blocks_0_feed_forward_2_bias_to_fp16, weight = transformer_blocks_0_feed_forward_2_weight_to_fp16, x = input_29_cast_fp16)[name = tensor<string, []>("linear_10_cast_fp16")];
165
+ tensor<fp16, [1, 512, 1024]> x_45_cast_fp16 = add(x = linear_10_cast_fp16, y = input_25_cast_fp16)[name = tensor<string, []>("x_45_cast_fp16")];
166
+ tensor<fp16, [1, 512, 1024]> x_47_cast_fp16 = add(x = x_45_cast_fp16, y = mapping_cast_fp16)[name = tensor<string, []>("x_47_cast_fp16")];
167
+ tensor<int32, []> var_298 = const()[name = tensor<string, []>("op_298"), val = tensor<int32, []>(-1)];
168
+ tensor<fp16, [2048, 256]> transformer_blocks_1_attention_norm_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_norm_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(19954432)))];
169
+ tensor<fp16, [2048]> transformer_blocks_1_attention_norm_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_norm_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(21003072)))];
170
+ tensor<fp16, [1, 2048]> linear_11_cast_fp16 = linear(bias = transformer_blocks_1_attention_norm_fc_bias_to_fp16, weight = transformer_blocks_1_attention_norm_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_11_cast_fp16")];
171
+ tensor<int32, [3]> var_317 = const()[name = tensor<string, []>("op_317"), val = tensor<int32, [3]>([1, 2048, 1])];
172
+ tensor<fp16, [1, 2048, 1]> h_11_cast_fp16 = reshape(shape = var_317, x = linear_11_cast_fp16)[name = tensor<string, []>("h_11_cast_fp16")];
173
+ tensor<int32, [2]> var_319_split_sizes_0 = const()[name = tensor<string, []>("op_319_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
174
+ tensor<int32, []> var_319_axis_0 = const()[name = tensor<string, []>("op_319_axis_0"), val = tensor<int32, []>(1)];
175
+ tensor<fp16, [1, 1024, 1]> var_319_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_319_cast_fp16_1 = split(axis = var_319_axis_0, split_sizes = var_319_split_sizes_0, x = h_11_cast_fp16)[name = tensor<string, []>("op_319_cast_fp16")];
176
+ tensor<int32, [3]> gamma_11_perm_0 = const()[name = tensor<string, []>("gamma_11_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
177
+ tensor<int32, [3]> beta_11_perm_0 = const()[name = tensor<string, []>("beta_11_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
178
+ tensor<int32, [1]> x_51_axes_0 = const()[name = tensor<string, []>("x_51_axes_0"), val = tensor<int32, [1]>([-1])];
179
+ tensor<fp16, []> var_291_to_fp16 = const()[name = tensor<string, []>("op_291_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
180
+ tensor<fp16, [1, 512, 1024]> x_51_cast_fp16 = layer_norm(axes = x_51_axes_0, epsilon = var_291_to_fp16, x = x_47_cast_fp16)[name = tensor<string, []>("x_51_cast_fp16")];
181
+ tensor<fp16, []> var_325_promoted_to_fp16 = const()[name = tensor<string, []>("op_325_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
182
+ tensor<fp16, [1, 1, 1024]> gamma_11_cast_fp16 = transpose(perm = gamma_11_perm_0, x = var_319_cast_fp16_0)[name = tensor<string, []>("transpose_32")];
183
+ tensor<fp16, [1, 1, 1024]> var_326_cast_fp16 = add(x = gamma_11_cast_fp16, y = var_325_promoted_to_fp16)[name = tensor<string, []>("op_326_cast_fp16")];
184
+ tensor<fp16, [1, 512, 1024]> var_327_cast_fp16 = mul(x = var_326_cast_fp16, y = x_51_cast_fp16)[name = tensor<string, []>("op_327_cast_fp16")];
185
+ tensor<fp16, [1, 1, 1024]> beta_11_cast_fp16 = transpose(perm = beta_11_perm_0, x = var_319_cast_fp16_1)[name = tensor<string, []>("transpose_31")];
186
+ tensor<fp16, [1, 512, 1024]> x_53_cast_fp16 = add(x = var_327_cast_fp16, y = beta_11_cast_fp16)[name = tensor<string, []>("x_53_cast_fp16")];
187
+ tensor<fp16, [2048, 256]> transformer_blocks_1_attention_norm_context_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_norm_context_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(21007232)))];
188
+ tensor<fp16, [2048]> transformer_blocks_1_attention_norm_context_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_norm_context_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(22055872)))];
189
+ tensor<fp16, [1, 2048]> linear_12_cast_fp16 = linear(bias = transformer_blocks_1_attention_norm_context_fc_bias_to_fp16, weight = transformer_blocks_1_attention_norm_context_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_12_cast_fp16")];
190
+ tensor<int32, [3]> var_339 = const()[name = tensor<string, []>("op_339"), val = tensor<int32, [3]>([1, 2048, 1])];
191
+ tensor<fp16, [1, 2048, 1]> h_15_cast_fp16 = reshape(shape = var_339, x = linear_12_cast_fp16)[name = tensor<string, []>("h_15_cast_fp16")];
192
+ tensor<int32, [2]> var_341_split_sizes_0 = const()[name = tensor<string, []>("op_341_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
193
+ tensor<int32, []> var_341_axis_0 = const()[name = tensor<string, []>("op_341_axis_0"), val = tensor<int32, []>(1)];
194
+ tensor<fp16, [1, 1024, 1]> var_341_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_341_cast_fp16_1 = split(axis = var_341_axis_0, split_sizes = var_341_split_sizes_0, x = h_15_cast_fp16)[name = tensor<string, []>("op_341_cast_fp16")];
195
+ tensor<int32, [3]> gamma_15_perm_0 = const()[name = tensor<string, []>("gamma_15_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
196
+ tensor<int32, [3]> beta_15_perm_0 = const()[name = tensor<string, []>("beta_15_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
197
+ tensor<fp16, []> var_347_promoted_to_fp16 = const()[name = tensor<string, []>("op_347_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
198
+ tensor<fp16, [1, 1, 1024]> gamma_15_cast_fp16 = transpose(perm = gamma_15_perm_0, x = var_341_cast_fp16_0)[name = tensor<string, []>("transpose_30")];
199
+ tensor<fp16, [1, 1, 1024]> var_348_cast_fp16 = add(x = gamma_15_cast_fp16, y = var_347_promoted_to_fp16)[name = tensor<string, []>("op_348_cast_fp16")];
200
+ tensor<fp16, [1, 512, 1024]> var_349_cast_fp16 = mul(x = var_348_cast_fp16, y = x_51_cast_fp16)[name = tensor<string, []>("op_349_cast_fp16")];
201
+ tensor<fp16, [1, 1, 1024]> beta_15_cast_fp16 = transpose(perm = beta_15_perm_0, x = var_341_cast_fp16_1)[name = tensor<string, []>("transpose_29")];
202
+ tensor<fp16, [1, 512, 1024]> x_59_cast_fp16 = add(x = var_349_cast_fp16, y = beta_15_cast_fp16)[name = tensor<string, []>("x_59_cast_fp16")];
203
+ tensor<fp16, [512, 1024]> transformer_blocks_1_attention_to_q_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_to_q_weight_to_fp16"), val = tensor<fp16, [512, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(22060032)))];
204
+ tensor<fp16, [1, 512, 512]> linear_13_cast_fp16 = linear(bias = linear_6_bias_0_to_fp16, weight = transformer_blocks_1_attention_to_q_weight_to_fp16, x = x_53_cast_fp16)[name = tensor<string, []>("linear_13_cast_fp16")];
205
+ tensor<fp16, [1024, 1024]> transformer_blocks_1_attention_to_kv_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_to_kv_weight_to_fp16"), val = tensor<fp16, [1024, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(23108672)))];
206
+ tensor<fp16, [1, 512, 1024]> linear_14_cast_fp16 = linear(bias = linear_7_bias_0_to_fp16, weight = transformer_blocks_1_attention_to_kv_weight_to_fp16, x = x_59_cast_fp16)[name = tensor<string, []>("linear_14_cast_fp16")];
207
+ tensor<int32, [2]> var_357_split_sizes_0 = const()[name = tensor<string, []>("op_357_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
208
+ tensor<int32, []> var_357_axis_0 = const()[name = tensor<string, []>("op_357_axis_0"), val = tensor<int32, []>(-1)];
209
+ tensor<fp16, [1, 512, 512]> var_357_cast_fp16_0, tensor<fp16, [1, 512, 512]> var_357_cast_fp16_1 = split(axis = var_357_axis_0, split_sizes = var_357_split_sizes_0, x = linear_14_cast_fp16)[name = tensor<string, []>("op_357_cast_fp16")];
210
+ tensor<int32, [4]> var_366 = const()[name = tensor<string, []>("op_366"), val = tensor<int32, [4]>([1, 512, 8, 64])];
211
+ tensor<fp16, [1, 512, 8, 64]> x_63_cast_fp16 = reshape(shape = var_366, x = linear_13_cast_fp16)[name = tensor<string, []>("x_63_cast_fp16")];
212
+ tensor<int32, [4]> var_376 = const()[name = tensor<string, []>("op_376"), val = tensor<int32, [4]>([1, 512, 8, 64])];
213
+ tensor<fp16, [1, 512, 8, 64]> x_67_cast_fp16 = reshape(shape = var_376, x = var_357_cast_fp16_0)[name = tensor<string, []>("x_67_cast_fp16")];
214
+ tensor<int32, [4]> var_386 = const()[name = tensor<string, []>("op_386"), val = tensor<int32, [4]>([1, 512, 8, 64])];
215
+ tensor<fp16, [1, 512, 8, 64]> x_71_cast_fp16 = reshape(shape = var_386, x = var_357_cast_fp16_1)[name = tensor<string, []>("x_71_cast_fp16")];
216
+ tensor<int32, [4]> var_388 = const()[name = tensor<string, []>("op_388"), val = tensor<int32, [4]>([0, 2, 1, 3])];
217
+ tensor<bool, []> sim_5_transpose_x_0 = const()[name = tensor<string, []>("sim_5_transpose_x_0"), val = tensor<bool, []>(false)];
218
+ tensor<bool, []> sim_5_transpose_y_0 = const()[name = tensor<string, []>("sim_5_transpose_y_0"), val = tensor<bool, []>(false)];
219
+ tensor<int32, [4]> transpose_11_perm_0 = const()[name = tensor<string, []>("transpose_11_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
220
+ tensor<int32, [4]> transpose_12_perm_0 = const()[name = tensor<string, []>("transpose_12_perm_0"), val = tensor<int32, [4]>([0, 2, -1, -3])];
221
+ tensor<fp16, [1, 8, 64, 512]> transpose_12 = transpose(perm = transpose_12_perm_0, x = x_67_cast_fp16)[name = tensor<string, []>("transpose_26")];
222
+ tensor<fp16, [1, 8, 512, 64]> transpose_11 = transpose(perm = transpose_11_perm_0, x = x_63_cast_fp16)[name = tensor<string, []>("transpose_27")];
223
+ tensor<fp16, [1, 8, 512, 512]> sim_5_cast_fp16 = matmul(transpose_x = sim_5_transpose_x_0, transpose_y = sim_5_transpose_y_0, x = transpose_11, y = transpose_12)[name = tensor<string, []>("sim_5_cast_fp16")];
224
+ tensor<fp16, []> var_392_to_fp16 = const()[name = tensor<string, []>("op_392_to_fp16"), val = tensor<fp16, []>(0x1p-3)];
225
+ tensor<fp16, [1, 8, 512, 512]> sim_7_cast_fp16 = mul(x = sim_5_cast_fp16, y = var_392_to_fp16)[name = tensor<string, []>("sim_7_cast_fp16")];
226
+ tensor<fp16, [1, 8, 512, 512]> attn_3_cast_fp16 = softmax(axis = var_298, x = sim_7_cast_fp16)[name = tensor<string, []>("attn_3_cast_fp16")];
227
+ tensor<bool, []> x_73_transpose_x_0 = const()[name = tensor<string, []>("x_73_transpose_x_0"), val = tensor<bool, []>(false)];
228
+ tensor<bool, []> x_73_transpose_y_0 = const()[name = tensor<string, []>("x_73_transpose_y_0"), val = tensor<bool, []>(false)];
229
+ tensor<fp16, [1, 8, 512, 64]> v_3_cast_fp16 = transpose(perm = var_388, x = x_71_cast_fp16)[name = tensor<string, []>("transpose_28")];
230
+ tensor<fp16, [1, 8, 512, 64]> x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = attn_3_cast_fp16, y = v_3_cast_fp16)[name = tensor<string, []>("x_73_cast_fp16")];
231
+ tensor<int32, [4]> var_414 = const()[name = tensor<string, []>("op_414"), val = tensor<int32, [4]>([0, 2, 1, 3])];
232
+ tensor<int32, [3]> var_416 = const()[name = tensor<string, []>("op_416"), val = tensor<int32, [3]>([1, 512, 512])];
233
+ tensor<fp16, [1, 512, 8, 64]> x_75_cast_fp16 = transpose(perm = var_414, x = x_73_cast_fp16)[name = tensor<string, []>("transpose_25")];
234
+ tensor<fp16, [1, 512, 512]> input_39_cast_fp16 = reshape(shape = var_416, x = x_75_cast_fp16)[name = tensor<string, []>("input_39_cast_fp16")];
235
+ tensor<fp16, [1024, 512]> transformer_blocks_1_attention_attention_to_out_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_attention_to_out_weight_to_fp16"), val = tensor<fp16, [1024, 512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(25205888)))];
236
+ tensor<fp16, [1024]> transformer_blocks_1_attention_attention_to_out_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_attention_attention_to_out_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(26254528)))];
237
+ tensor<fp16, [1, 512, 1024]> linear_15_cast_fp16 = linear(bias = transformer_blocks_1_attention_attention_to_out_bias_to_fp16, weight = transformer_blocks_1_attention_attention_to_out_weight_to_fp16, x = input_39_cast_fp16)[name = tensor<string, []>("linear_15_cast_fp16")];
238
+ tensor<fp16, [1, 512, 1024]> input_41_cast_fp16 = add(x = linear_15_cast_fp16, y = x_47_cast_fp16)[name = tensor<string, []>("input_41_cast_fp16")];
239
+ tensor<fp16, [2048, 1024]> transformer_blocks_1_feed_forward_0_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_feed_forward_0_weight_to_fp16"), val = tensor<fp16, [2048, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(26256640)))];
240
+ tensor<fp16, [2048]> transformer_blocks_1_feed_forward_0_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_feed_forward_0_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(30451008)))];
241
+ tensor<fp16, [1, 512, 2048]> linear_16_cast_fp16 = linear(bias = transformer_blocks_1_feed_forward_0_bias_to_fp16, weight = transformer_blocks_1_feed_forward_0_weight_to_fp16, x = input_41_cast_fp16)[name = tensor<string, []>("linear_16_cast_fp16")];
242
+ tensor<string, []> input_45_mode_0 = const()[name = tensor<string, []>("input_45_mode_0"), val = tensor<string, []>("EXACT")];
243
+ tensor<fp16, [1, 512, 2048]> input_45_cast_fp16 = gelu(mode = input_45_mode_0, x = linear_16_cast_fp16)[name = tensor<string, []>("input_45_cast_fp16")];
244
+ tensor<fp16, [1024, 2048]> transformer_blocks_1_feed_forward_2_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_feed_forward_2_weight_to_fp16"), val = tensor<fp16, [1024, 2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(30455168)))];
245
+ tensor<fp16, [1024]> transformer_blocks_1_feed_forward_2_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_1_feed_forward_2_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(34649536)))];
246
+ tensor<fp16, [1, 512, 1024]> linear_17_cast_fp16 = linear(bias = transformer_blocks_1_feed_forward_2_bias_to_fp16, weight = transformer_blocks_1_feed_forward_2_weight_to_fp16, x = input_45_cast_fp16)[name = tensor<string, []>("linear_17_cast_fp16")];
247
+ tensor<fp16, [1, 512, 1024]> x_77_cast_fp16 = add(x = linear_17_cast_fp16, y = input_41_cast_fp16)[name = tensor<string, []>("x_77_cast_fp16")];
248
+ tensor<fp16, [1, 512, 1024]> x_79_cast_fp16 = add(x = x_77_cast_fp16, y = mapping_cast_fp16)[name = tensor<string, []>("x_79_cast_fp16")];
249
+ tensor<int32, []> var_443 = const()[name = tensor<string, []>("op_443"), val = tensor<int32, []>(-1)];
250
+ tensor<fp16, [2048, 256]> transformer_blocks_2_attention_norm_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_norm_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(34651648)))];
251
+ tensor<fp16, [2048]> transformer_blocks_2_attention_norm_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_norm_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(35700288)))];
252
+ tensor<fp16, [1, 2048]> linear_18_cast_fp16 = linear(bias = transformer_blocks_2_attention_norm_fc_bias_to_fp16, weight = transformer_blocks_2_attention_norm_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_18_cast_fp16")];
253
+ tensor<int32, [3]> var_462 = const()[name = tensor<string, []>("op_462"), val = tensor<int32, [3]>([1, 2048, 1])];
254
+ tensor<fp16, [1, 2048, 1]> h_19_cast_fp16 = reshape(shape = var_462, x = linear_18_cast_fp16)[name = tensor<string, []>("h_19_cast_fp16")];
255
+ tensor<int32, [2]> var_464_split_sizes_0 = const()[name = tensor<string, []>("op_464_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
256
+ tensor<int32, []> var_464_axis_0 = const()[name = tensor<string, []>("op_464_axis_0"), val = tensor<int32, []>(1)];
257
+ tensor<fp16, [1, 1024, 1]> var_464_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_464_cast_fp16_1 = split(axis = var_464_axis_0, split_sizes = var_464_split_sizes_0, x = h_19_cast_fp16)[name = tensor<string, []>("op_464_cast_fp16")];
258
+ tensor<int32, [3]> gamma_19_perm_0 = const()[name = tensor<string, []>("gamma_19_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
259
+ tensor<int32, [3]> beta_19_perm_0 = const()[name = tensor<string, []>("beta_19_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
260
+ tensor<int32, [1]> x_83_axes_0 = const()[name = tensor<string, []>("x_83_axes_0"), val = tensor<int32, [1]>([-1])];
261
+ tensor<fp16, []> var_436_to_fp16 = const()[name = tensor<string, []>("op_436_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
262
+ tensor<fp16, [1, 512, 1024]> x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, epsilon = var_436_to_fp16, x = x_79_cast_fp16)[name = tensor<string, []>("x_83_cast_fp16")];
263
+ tensor<fp16, []> var_470_promoted_to_fp16 = const()[name = tensor<string, []>("op_470_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
264
+ tensor<fp16, [1, 1, 1024]> gamma_19_cast_fp16 = transpose(perm = gamma_19_perm_0, x = var_464_cast_fp16_0)[name = tensor<string, []>("transpose_24")];
265
+ tensor<fp16, [1, 1, 1024]> var_471_cast_fp16 = add(x = gamma_19_cast_fp16, y = var_470_promoted_to_fp16)[name = tensor<string, []>("op_471_cast_fp16")];
266
+ tensor<fp16, [1, 512, 1024]> var_472_cast_fp16 = mul(x = var_471_cast_fp16, y = x_83_cast_fp16)[name = tensor<string, []>("op_472_cast_fp16")];
267
+ tensor<fp16, [1, 1, 1024]> beta_19_cast_fp16 = transpose(perm = beta_19_perm_0, x = var_464_cast_fp16_1)[name = tensor<string, []>("transpose_23")];
268
+ tensor<fp16, [1, 512, 1024]> x_85_cast_fp16 = add(x = var_472_cast_fp16, y = beta_19_cast_fp16)[name = tensor<string, []>("x_85_cast_fp16")];
269
+ tensor<fp16, [2048, 256]> transformer_blocks_2_attention_norm_context_fc_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_norm_context_fc_weight_to_fp16"), val = tensor<fp16, [2048, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(35704448)))];
270
+ tensor<fp16, [2048]> transformer_blocks_2_attention_norm_context_fc_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_norm_context_fc_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(36753088)))];
271
+ tensor<fp16, [1, 2048]> linear_19_cast_fp16 = linear(bias = transformer_blocks_2_attention_norm_context_fc_bias_to_fp16, weight = transformer_blocks_2_attention_norm_context_fc_weight_to_fp16, x = features_to_fp16)[name = tensor<string, []>("linear_19_cast_fp16")];
272
+ tensor<int32, [3]> var_484 = const()[name = tensor<string, []>("op_484"), val = tensor<int32, [3]>([1, 2048, 1])];
273
+ tensor<fp16, [1, 2048, 1]> h_cast_fp16 = reshape(shape = var_484, x = linear_19_cast_fp16)[name = tensor<string, []>("h_cast_fp16")];
274
+ tensor<int32, [2]> var_486_split_sizes_0 = const()[name = tensor<string, []>("op_486_split_sizes_0"), val = tensor<int32, [2]>([1024, 1024])];
275
+ tensor<int32, []> var_486_axis_0 = const()[name = tensor<string, []>("op_486_axis_0"), val = tensor<int32, []>(1)];
276
+ tensor<fp16, [1, 1024, 1]> var_486_cast_fp16_0, tensor<fp16, [1, 1024, 1]> var_486_cast_fp16_1 = split(axis = var_486_axis_0, split_sizes = var_486_split_sizes_0, x = h_cast_fp16)[name = tensor<string, []>("op_486_cast_fp16")];
277
+ tensor<int32, [3]> gamma_perm_0 = const()[name = tensor<string, []>("gamma_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
278
+ tensor<int32, [3]> beta_perm_0 = const()[name = tensor<string, []>("beta_perm_0"), val = tensor<int32, [3]>([0, -1, 1])];
279
+ tensor<fp16, []> var_492_promoted_to_fp16 = const()[name = tensor<string, []>("op_492_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
280
+ tensor<fp16, [1, 1, 1024]> gamma_cast_fp16 = transpose(perm = gamma_perm_0, x = var_486_cast_fp16_0)[name = tensor<string, []>("transpose_22")];
281
+ tensor<fp16, [1, 1, 1024]> var_493_cast_fp16 = add(x = gamma_cast_fp16, y = var_492_promoted_to_fp16)[name = tensor<string, []>("op_493_cast_fp16")];
282
+ tensor<fp16, [1, 512, 1024]> var_494_cast_fp16 = mul(x = var_493_cast_fp16, y = x_83_cast_fp16)[name = tensor<string, []>("op_494_cast_fp16")];
283
+ tensor<fp16, [1, 1, 1024]> beta_cast_fp16 = transpose(perm = beta_perm_0, x = var_486_cast_fp16_1)[name = tensor<string, []>("transpose_21")];
284
+ tensor<fp16, [1, 512, 1024]> x_91_cast_fp16 = add(x = var_494_cast_fp16, y = beta_cast_fp16)[name = tensor<string, []>("x_91_cast_fp16")];
285
+ tensor<fp16, [512, 1024]> transformer_blocks_2_attention_to_q_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_to_q_weight_to_fp16"), val = tensor<fp16, [512, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(36757248)))];
286
+ tensor<fp16, [1, 512, 512]> linear_20_cast_fp16 = linear(bias = linear_6_bias_0_to_fp16, weight = transformer_blocks_2_attention_to_q_weight_to_fp16, x = x_85_cast_fp16)[name = tensor<string, []>("linear_20_cast_fp16")];
287
+ tensor<fp16, [1024, 1024]> transformer_blocks_2_attention_to_kv_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_to_kv_weight_to_fp16"), val = tensor<fp16, [1024, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(37805888)))];
288
+ tensor<fp16, [1, 512, 1024]> linear_21_cast_fp16 = linear(bias = linear_7_bias_0_to_fp16, weight = transformer_blocks_2_attention_to_kv_weight_to_fp16, x = x_91_cast_fp16)[name = tensor<string, []>("linear_21_cast_fp16")];
289
+ tensor<int32, [2]> var_502_split_sizes_0 = const()[name = tensor<string, []>("op_502_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
290
+ tensor<int32, []> var_502_axis_0 = const()[name = tensor<string, []>("op_502_axis_0"), val = tensor<int32, []>(-1)];
291
+ tensor<fp16, [1, 512, 512]> var_502_cast_fp16_0, tensor<fp16, [1, 512, 512]> var_502_cast_fp16_1 = split(axis = var_502_axis_0, split_sizes = var_502_split_sizes_0, x = linear_21_cast_fp16)[name = tensor<string, []>("op_502_cast_fp16")];
292
+ tensor<int32, [4]> var_511 = const()[name = tensor<string, []>("op_511"), val = tensor<int32, [4]>([1, 512, 8, 64])];
293
+ tensor<fp16, [1, 512, 8, 64]> x_95_cast_fp16 = reshape(shape = var_511, x = linear_20_cast_fp16)[name = tensor<string, []>("x_95_cast_fp16")];
294
+ tensor<int32, [4]> var_521 = const()[name = tensor<string, []>("op_521"), val = tensor<int32, [4]>([1, 512, 8, 64])];
295
+ tensor<fp16, [1, 512, 8, 64]> x_99_cast_fp16 = reshape(shape = var_521, x = var_502_cast_fp16_0)[name = tensor<string, []>("x_99_cast_fp16")];
296
+ tensor<int32, [4]> var_531 = const()[name = tensor<string, []>("op_531"), val = tensor<int32, [4]>([1, 512, 8, 64])];
297
+ tensor<fp16, [1, 512, 8, 64]> x_103_cast_fp16 = reshape(shape = var_531, x = var_502_cast_fp16_1)[name = tensor<string, []>("x_103_cast_fp16")];
298
+ tensor<int32, [4]> var_533 = const()[name = tensor<string, []>("op_533"), val = tensor<int32, [4]>([0, 2, 1, 3])];
299
+ tensor<bool, []> sim_9_transpose_x_0 = const()[name = tensor<string, []>("sim_9_transpose_x_0"), val = tensor<bool, []>(false)];
300
+ tensor<bool, []> sim_9_transpose_y_0 = const()[name = tensor<string, []>("sim_9_transpose_y_0"), val = tensor<bool, []>(false)];
301
+ tensor<int32, [4]> transpose_13_perm_0 = const()[name = tensor<string, []>("transpose_13_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
302
+ tensor<int32, [4]> transpose_14_perm_0 = const()[name = tensor<string, []>("transpose_14_perm_0"), val = tensor<int32, [4]>([0, 2, -1, -3])];
303
+ tensor<fp16, [1, 8, 64, 512]> transpose_14 = transpose(perm = transpose_14_perm_0, x = x_99_cast_fp16)[name = tensor<string, []>("transpose_18")];
304
+ tensor<fp16, [1, 8, 512, 64]> transpose_13 = transpose(perm = transpose_13_perm_0, x = x_95_cast_fp16)[name = tensor<string, []>("transpose_19")];
305
+ tensor<fp16, [1, 8, 512, 512]> sim_9_cast_fp16 = matmul(transpose_x = sim_9_transpose_x_0, transpose_y = sim_9_transpose_y_0, x = transpose_13, y = transpose_14)[name = tensor<string, []>("sim_9_cast_fp16")];
306
+ tensor<fp16, []> var_537_to_fp16 = const()[name = tensor<string, []>("op_537_to_fp16"), val = tensor<fp16, []>(0x1p-3)];
307
+ tensor<fp16, [1, 8, 512, 512]> sim_cast_fp16 = mul(x = sim_9_cast_fp16, y = var_537_to_fp16)[name = tensor<string, []>("sim_cast_fp16")];
308
+ tensor<fp16, [1, 8, 512, 512]> attn_cast_fp16 = softmax(axis = var_443, x = sim_cast_fp16)[name = tensor<string, []>("attn_cast_fp16")];
309
+ tensor<bool, []> x_105_transpose_x_0 = const()[name = tensor<string, []>("x_105_transpose_x_0"), val = tensor<bool, []>(false)];
310
+ tensor<bool, []> x_105_transpose_y_0 = const()[name = tensor<string, []>("x_105_transpose_y_0"), val = tensor<bool, []>(false)];
311
+ tensor<fp16, [1, 8, 512, 64]> v_cast_fp16 = transpose(perm = var_533, x = x_103_cast_fp16)[name = tensor<string, []>("transpose_20")];
312
+ tensor<fp16, [1, 8, 512, 64]> x_105_cast_fp16 = matmul(transpose_x = x_105_transpose_x_0, transpose_y = x_105_transpose_y_0, x = attn_cast_fp16, y = v_cast_fp16)[name = tensor<string, []>("x_105_cast_fp16")];
313
+ tensor<int32, [4]> var_559 = const()[name = tensor<string, []>("op_559"), val = tensor<int32, [4]>([0, 2, 1, 3])];
314
+ tensor<int32, [3]> var_561 = const()[name = tensor<string, []>("op_561"), val = tensor<int32, [3]>([1, 512, 512])];
315
+ tensor<fp16, [1, 512, 8, 64]> x_107_cast_fp16 = transpose(perm = var_559, x = x_105_cast_fp16)[name = tensor<string, []>("transpose_17")];
316
+ tensor<fp16, [1, 512, 512]> input_55_cast_fp16 = reshape(shape = var_561, x = x_107_cast_fp16)[name = tensor<string, []>("input_55_cast_fp16")];
317
+ tensor<fp16, [1024, 512]> transformer_blocks_2_attention_attention_to_out_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_attention_to_out_weight_to_fp16"), val = tensor<fp16, [1024, 512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(39903104)))];
318
+ tensor<fp16, [1024]> transformer_blocks_2_attention_attention_to_out_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_attention_attention_to_out_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(40951744)))];
319
+ tensor<fp16, [1, 512, 1024]> linear_22_cast_fp16 = linear(bias = transformer_blocks_2_attention_attention_to_out_bias_to_fp16, weight = transformer_blocks_2_attention_attention_to_out_weight_to_fp16, x = input_55_cast_fp16)[name = tensor<string, []>("linear_22_cast_fp16")];
320
+ tensor<fp16, [1, 512, 1024]> input_57_cast_fp16 = add(x = linear_22_cast_fp16, y = x_79_cast_fp16)[name = tensor<string, []>("input_57_cast_fp16")];
321
+ tensor<fp16, [2048, 1024]> transformer_blocks_2_feed_forward_0_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_feed_forward_0_weight_to_fp16"), val = tensor<fp16, [2048, 1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(40953856)))];
322
+ tensor<fp16, [2048]> transformer_blocks_2_feed_forward_0_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_feed_forward_0_bias_to_fp16"), val = tensor<fp16, [2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(45148224)))];
323
+ tensor<fp16, [1, 512, 2048]> linear_23_cast_fp16 = linear(bias = transformer_blocks_2_feed_forward_0_bias_to_fp16, weight = transformer_blocks_2_feed_forward_0_weight_to_fp16, x = input_57_cast_fp16)[name = tensor<string, []>("linear_23_cast_fp16")];
324
+ tensor<string, []> input_61_mode_0 = const()[name = tensor<string, []>("input_61_mode_0"), val = tensor<string, []>("EXACT")];
325
+ tensor<fp16, [1, 512, 2048]> input_61_cast_fp16 = gelu(mode = input_61_mode_0, x = linear_23_cast_fp16)[name = tensor<string, []>("input_61_cast_fp16")];
326
+ tensor<fp16, [1024, 2048]> transformer_blocks_2_feed_forward_2_weight_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_feed_forward_2_weight_to_fp16"), val = tensor<fp16, [1024, 2048]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(45152384)))];
327
+ tensor<fp16, [1024]> transformer_blocks_2_feed_forward_2_bias_to_fp16 = const()[name = tensor<string, []>("transformer_blocks_2_feed_forward_2_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(49346752)))];
328
+ tensor<fp16, [1, 512, 1024]> linear_24_cast_fp16 = linear(bias = transformer_blocks_2_feed_forward_2_bias_to_fp16, weight = transformer_blocks_2_feed_forward_2_weight_to_fp16, x = input_61_cast_fp16)[name = tensor<string, []>("linear_24_cast_fp16")];
329
+ tensor<fp16, [1, 512, 1024]> x_109_cast_fp16 = add(x = linear_24_cast_fp16, y = input_57_cast_fp16)[name = tensor<string, []>("x_109_cast_fp16")];
330
+ tensor<int32, [1]> var_581_axes_0 = const()[name = tensor<string, []>("op_581_axes_0"), val = tensor<int32, [1]>([1])];
331
+ tensor<bool, []> var_581_keep_dims_0 = const()[name = tensor<string, []>("op_581_keep_dims_0"), val = tensor<bool, []>(false)];
332
+ tensor<fp16, [1, 1024]> var_581_cast_fp16 = reduce_mean(axes = var_581_axes_0, keep_dims = var_581_keep_dims_0, x = x_109_cast_fp16)[name = tensor<string, []>("op_581_cast_fp16")];
333
+ tensor<int32, [1]> x_111_axes_0 = const()[name = tensor<string, []>("x_111_axes_0"), val = tensor<int32, [1]>([1])];
334
+ tensor<fp16, [1, 1, 1024]> x_111_cast_fp16 = expand_dims(axes = x_111_axes_0, x = var_581_cast_fp16)[name = tensor<string, []>("x_111_cast_fp16")];
335
+ tensor<int32, [3]> var_590 = const()[name = tensor<string, []>("op_590"), val = tensor<int32, [3]>([0, 2, 1])];
336
+ tensor<string, []> x_pad_type_0 = const()[name = tensor<string, []>("x_pad_type_0"), val = tensor<string, []>("valid")];
337
+ tensor<int32, [1]> x_strides_0 = const()[name = tensor<string, []>("x_strides_0"), val = tensor<int32, [1]>([1])];
338
+ tensor<int32, [2]> x_pad_0 = const()[name = tensor<string, []>("x_pad_0"), val = tensor<int32, [2]>([0, 0])];
339
+ tensor<int32, [1]> x_dilations_0 = const()[name = tensor<string, []>("x_dilations_0"), val = tensor<int32, [1]>([1])];
340
+ tensor<int32, []> x_groups_0 = const()[name = tensor<string, []>("x_groups_0"), val = tensor<int32, []>(1)];
341
+ tensor<fp16, [256, 1024, 1]> transformer_to_out_1_weight_to_fp16 = const()[name = tensor<string, []>("transformer_to_out_1_weight_to_fp16"), val = tensor<fp16, [256, 1024, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(49348864)))];
342
+ tensor<fp16, [256]> transformer_to_out_1_bias_to_fp16 = const()[name = tensor<string, []>("transformer_to_out_1_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(49873216)))];
343
+ tensor<fp16, [1, 1024, 1]> input_cast_fp16 = transpose(perm = var_590, x = x_111_cast_fp16)[name = tensor<string, []>("transpose_16")];
344
+ tensor<fp16, [1, 256, 1]> x_cast_fp16 = conv(bias = transformer_to_out_1_bias_to_fp16, dilations = x_dilations_0, groups = x_groups_0, pad = x_pad_0, pad_type = x_pad_type_0, strides = x_strides_0, weight = transformer_to_out_1_weight_to_fp16, x = input_cast_fp16)[name = tensor<string, []>("x_cast_fp16")];
345
+ tensor<int32, [3]> x_pred_perm_0 = const()[name = tensor<string, []>("x_pred_perm_0"), val = tensor<int32, [3]>([0, -1, -2])];
346
+ tensor<fp16, [1, 1, 256]> var_602_cast_fp16 = mul(x = c_skip_cast_fp16, y = x_noisy_to_fp16)[name = tensor<string, []>("op_602_cast_fp16")];
347
+ tensor<fp16, [1, 1, 256]> x_pred_cast_fp16 = transpose(perm = x_pred_perm_0, x = x_cast_fp16)[name = tensor<string, []>("transpose_15")];
348
+ tensor<fp16, [1, 1, 256]> var_603_cast_fp16 = mul(x = c_out_cast_fp16, y = x_pred_cast_fp16)[name = tensor<string, []>("op_603_cast_fp16")];
349
+ tensor<fp16, [1, 1, 256]> denoised = add(x = var_602_cast_fp16, y = var_603_cast_fp16)[name = tensor<string, []>("op_605_cast_fp16")];
350
+ } -> (denoised);
351
+ }
compiled/styletts2_diffusion_step_512.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17ba48a8bdc68851289a23593b223573aaddd1b445e8c77765f5350feed8a251
3
+ size 49873792
compiled/styletts2_f0n_energy.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5467b4fad22d57b3e2e115d3b52104d6d7a0288ae95fe2f7d0789c79cda9bb9d
3
+ size 243
compiled/styletts2_f0n_energy.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:121cb6ca7a0aedd3d608362b9dc12ba3b15a09b81c827b5578b20cf4364ee5d3
3
+ size 427
compiled/styletts2_f0n_energy.mlmodelc/metadata.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 F0Ntrain (LibriTTS). ANE-eligible.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16)",
11
+ "shortDescription" : "",
12
+ "shape" : "[]",
13
+ "name" : "F0",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16)",
21
+ "shortDescription" : "",
22
+ "shape" : "[]",
23
+ "name" : "N",
24
+ "type" : "MultiArray"
25
+ }
26
+ ],
27
+ "storagePrecision" : "Float16",
28
+ "modelParameters" : [
29
+
30
+ ],
31
+ "specificationVersion" : 8,
32
+ "mlProgramOperationTypeHistogram" : {
33
+ "Ios17.squeeze" : 4,
34
+ "Ios17.mul" : 18,
35
+ "Ios17.linear" : 12,
36
+ "Ios17.transpose" : 3,
37
+ "Ios17.conv" : 16,
38
+ "Ios17.leakyRelu" : 12,
39
+ "Ios17.add" : 30,
40
+ "Ios17.convTranspose" : 2,
41
+ "Ios17.lstm" : 1,
42
+ "Ios17.sliceByIndex" : 2,
43
+ "UpsampleNearestNeighbor" : 2,
44
+ "Ios17.expandDims" : 2,
45
+ "Ios17.instanceNorm" : 11,
46
+ "Ios17.reshape" : 12,
47
+ "Ios17.cast" : 2,
48
+ "Split" : 12
49
+ },
50
+ "computePrecision" : "Mixed (Float16, Float32, Int32)",
51
+ "isUpdatable" : "0",
52
+ "stateSchema" : [
53
+
54
+ ],
55
+ "availability" : {
56
+ "macOS" : "14.0",
57
+ "tvOS" : "17.0",
58
+ "visionOS" : "1.0",
59
+ "watchOS" : "10.0",
60
+ "iOS" : "17.0",
61
+ "macCatalyst" : "17.0"
62
+ },
63
+ "modelType" : {
64
+ "name" : "MLModelType_mlProgram"
65
+ },
66
+ "inputSchema" : [
67
+ {
68
+ "shortDescription" : "",
69
+ "dataType" : "Float32",
70
+ "hasShapeFlexibility" : "1",
71
+ "isOptional" : "0",
72
+ "shapeFlexibility" : "1 × 640 × 4096 | 1 × 640 × 256 | 1 × 640 × 512 | 1 × 640 × 1024 | 1 × 640 × 2048",
73
+ "formattedType" : "MultiArray (Float32 1 × 640 × 4096)",
74
+ "type" : "MultiArray",
75
+ "shape" : "[1, 640, 4096]",
76
+ "name" : "en",
77
+ "enumeratedShapes" : "[[1, 640, 4096], [1, 640, 256], [1, 640, 512], [1, 640, 1024], [1, 640, 2048]]"
78
+ },
79
+ {
80
+ "hasShapeFlexibility" : "0",
81
+ "isOptional" : "0",
82
+ "dataType" : "Float32",
83
+ "formattedType" : "MultiArray (Float32 1 × 128)",
84
+ "shortDescription" : "",
85
+ "shape" : "[1, 128]",
86
+ "name" : "s",
87
+ "type" : "MultiArray"
88
+ }
89
+ ],
90
+ "userDefinedMetadata" : {
91
+ "com.github.apple.coremltools.version" : "8.3.0",
92
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
93
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
94
+ },
95
+ "generatedClassName" : "styletts2_f0n_energy",
96
+ "method" : "predict"
97
+ }
98
+ ]
compiled/styletts2_f0n_energy.mlmodelc/model.mil ADDED
@@ -0,0 +1,396 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.0)
2
+ [buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "3520.4.1"}, {"coremlc-version", "3520.5.1"}, {"coremltools-component-torch", "2.11.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})]
3
+ {
4
+ func main<ios17>(tensor<fp32, [1, 640, ?]> en, tensor<fp32, [1, 128]> s) [FlexibleShapeInformation = tuple<tuple<tensor<string, []>, dict<tensor<string, []>, tensor<int32, [?]>>>, tuple<tensor<string, []>, dict<tensor<string, []>, dict<tensor<string, []>, tensor<int32, [?]>>>>>((("DefaultShapes", {{"en", [1, 640, 4096]}}), ("EnumeratedShapes", {{"en_1_1_1_640_1024_s_1_1_1_1_128_", {{"en", [1, 640, 1024]}, {"s", [1, 128]}}}, {"en_1_1_1_640_2048_s_1_1_1_1_128_", {{"en", [1, 640, 2048]}, {"s", [1, 128]}}}, {"en_1_1_1_640_256_s_1_1_1_1_128_", {{"en", [1, 640, 256]}, {"s", [1, 128]}}}, {"en_1_1_1_640_4096_s_1_1_1_1_128_", {{"en", [1, 640, 4096]}, {"s", [1, 128]}}}, {"en_1_1_1_640_512_s_1_1_1_1_128_", {{"en", [1, 640, 512]}, {"s", [1, 128]}}}})))] {
5
+ tensor<int32, [3]> transpose_0_perm_0 = const()[name = tensor<string, []>("transpose_0_perm_0"), val = tensor<int32, [3]>([-1, 0, -2])];
6
+ tensor<string, []> en_to_fp16_dtype_0 = const()[name = tensor<string, []>("en_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
7
+ tensor<string, []> x_batch_first_direction_0 = const()[name = tensor<string, []>("x_batch_first_direction_0"), val = tensor<string, []>("bidirectional")];
8
+ tensor<bool, []> x_batch_first_output_sequence_0 = const()[name = tensor<string, []>("x_batch_first_output_sequence_0"), val = tensor<bool, []>(true)];
9
+ tensor<string, []> x_batch_first_recurrent_activation_0 = const()[name = tensor<string, []>("x_batch_first_recurrent_activation_0"), val = tensor<string, []>("sigmoid")];
10
+ tensor<string, []> x_batch_first_cell_activation_0 = const()[name = tensor<string, []>("x_batch_first_cell_activation_0"), val = tensor<string, []>("tanh")];
11
+ tensor<string, []> x_batch_first_activation_0 = const()[name = tensor<string, []>("x_batch_first_activation_0"), val = tensor<string, []>("tanh")];
12
+ tensor<fp16, [1, 512]> x_batch_first_lstm_h0_reshaped_to_fp16 = const()[name = tensor<string, []>("x_batch_first_lstm_h0_reshaped_to_fp16"), val = tensor<fp16, [1, 512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
13
+ tensor<fp16, [1024, 640]> concat_4_to_fp16 = const()[name = tensor<string, []>("concat_4_to_fp16"), val = tensor<fp16, [1024, 640]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1152)))];
14
+ tensor<fp16, [1024, 256]> concat_5_to_fp16 = const()[name = tensor<string, []>("concat_5_to_fp16"), val = tensor<fp16, [1024, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1311936)))];
15
+ tensor<fp16, [1024]> add_1_to_fp16 = const()[name = tensor<string, []>("add_1_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1836288)))];
16
+ tensor<fp16, [1024, 640]> concat_6_to_fp16 = const()[name = tensor<string, []>("concat_6_to_fp16"), val = tensor<fp16, [1024, 640]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(1838400)))];
17
+ tensor<fp16, [1024, 256]> concat_7_to_fp16 = const()[name = tensor<string, []>("concat_7_to_fp16"), val = tensor<fp16, [1024, 256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3149184)))];
18
+ tensor<fp16, [1024]> add_2_to_fp16 = const()[name = tensor<string, []>("add_2_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3673536)))];
19
+ tensor<fp16, [1, 640, ?]> en_to_fp16 = cast(dtype = en_to_fp16_dtype_0, x = en)[name = tensor<string, []>("cast_27")];
20
+ tensor<fp16, [?, 1, 640]> transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = en_to_fp16)[name = tensor<string, []>("transpose_3")];
21
+ tensor<fp16, [?, 1, 512]> x_batch_first_cast_fp16_0, tensor<fp16, [1, 512]> x_batch_first_cast_fp16_1, tensor<fp16, [1, 512]> x_batch_first_cast_fp16_2 = lstm(activation = x_batch_first_activation_0, bias = add_1_to_fp16, bias_back = add_2_to_fp16, cell_activation = x_batch_first_cell_activation_0, direction = x_batch_first_direction_0, initial_c = x_batch_first_lstm_h0_reshaped_to_fp16, initial_h = x_batch_first_lstm_h0_reshaped_to_fp16, output_sequence = x_batch_first_output_sequence_0, recurrent_activation = x_batch_first_recurrent_activation_0, weight_hh = concat_5_to_fp16, weight_hh_back = concat_7_to_fp16, weight_ih = concat_4_to_fp16, weight_ih_back = concat_6_to_fp16, x = transpose_0_cast_fp16)[name = tensor<string, []>("x_batch_first_cast_fp16")];
22
+ tensor<int32, [3]> x_perm_0 = const()[name = tensor<string, []>("x_perm_0"), val = tensor<int32, [3]>([1, 0, 2])];
23
+ tensor<int32, [3]> input_3_perm_0 = const()[name = tensor<string, []>("input_3_perm_0"), val = tensor<int32, [3]>([0, -1, -2])];
24
+ tensor<fp32, []> var_53 = const()[name = tensor<string, []>("op_53"), val = tensor<fp32, []>(0x1.99999ap-3)];
25
+ tensor<string, []> s_to_fp16_dtype_0 = const()[name = tensor<string, []>("s_to_fp16_dtype_0"), val = tensor<string, []>("fp16")];
26
+ tensor<fp16, [1024, 128]> F0_0_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_0_norm1_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3675648)))];
27
+ tensor<fp16, [1024]> F0_0_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_0_norm1_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3937856)))];
28
+ tensor<fp16, [1, 128]> s_to_fp16 = cast(dtype = s_to_fp16_dtype_0, x = s)[name = tensor<string, []>("cast_26")];
29
+ tensor<fp16, [1, 1024]> linear_0_cast_fp16 = linear(bias = F0_0_norm1_fc_bias_to_fp16, weight = F0_0_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_0_cast_fp16")];
30
+ tensor<int32, [3]> var_78 = const()[name = tensor<string, []>("op_78"), val = tensor<int32, [3]>([1, 1024, 1])];
31
+ tensor<fp16, [1, 1024, 1]> h_3_cast_fp16 = reshape(shape = var_78, x = linear_0_cast_fp16)[name = tensor<string, []>("h_3_cast_fp16")];
32
+ tensor<int32, [2]> var_80_split_sizes_0 = const()[name = tensor<string, []>("op_80_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
33
+ tensor<int32, []> var_80_axis_0 = const()[name = tensor<string, []>("op_80_axis_0"), val = tensor<int32, []>(1)];
34
+ tensor<fp16, [1, 512, 1]> var_80_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_80_cast_fp16_1 = split(axis = var_80_axis_0, split_sizes = var_80_split_sizes_0, x = h_3_cast_fp16)[name = tensor<string, []>("op_80_cast_fp16")];
35
+ tensor<fp16, []> var_82_promoted_to_fp16 = const()[name = tensor<string, []>("op_82_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
36
+ tensor<fp16, [1, 512, 1]> var_83_cast_fp16 = add(x = var_80_cast_fp16_0, y = var_82_promoted_to_fp16)[name = tensor<string, []>("op_83_cast_fp16")];
37
+ tensor<fp16, []> var_56_to_fp16 = const()[name = tensor<string, []>("op_56_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
38
+ tensor<fp16, [1, ?, 512]> x_cast_fp16 = transpose(perm = x_perm_0, x = x_batch_first_cast_fp16_0)[name = tensor<string, []>("transpose_2")];
39
+ tensor<fp16, [1, 512, ?]> input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = x_cast_fp16)[name = tensor<string, []>("transpose_1")];
40
+ tensor<fp16, [1, 512, ?]> var_84_cast_fp16 = instance_norm(epsilon = var_56_to_fp16, x = input_3_cast_fp16)[name = tensor<string, []>("op_84_cast_fp16")];
41
+ tensor<fp16, [1, 512, ?]> var_85_cast_fp16 = mul(x = var_83_cast_fp16, y = var_84_cast_fp16)[name = tensor<string, []>("op_85_cast_fp16")];
42
+ tensor<fp16, [1, 512, ?]> input_5_cast_fp16 = add(x = var_85_cast_fp16, y = var_80_cast_fp16_1)[name = tensor<string, []>("input_5_cast_fp16")];
43
+ tensor<fp16, [1, 512, ?]> input_7_cast_fp16 = leaky_relu(alpha = var_53, x = input_5_cast_fp16)[name = tensor<string, []>("input_7_cast_fp16")];
44
+ tensor<string, []> input_11_pad_type_0 = const()[name = tensor<string, []>("input_11_pad_type_0"), val = tensor<string, []>("custom")];
45
+ tensor<int32, [2]> input_11_pad_0 = const()[name = tensor<string, []>("input_11_pad_0"), val = tensor<int32, [2]>([1, 1])];
46
+ tensor<int32, [1]> input_11_strides_0 = const()[name = tensor<string, []>("input_11_strides_0"), val = tensor<int32, [1]>([1])];
47
+ tensor<int32, [1]> input_11_dilations_0 = const()[name = tensor<string, []>("input_11_dilations_0"), val = tensor<int32, [1]>([1])];
48
+ tensor<int32, []> input_11_groups_0 = const()[name = tensor<string, []>("input_11_groups_0"), val = tensor<int32, []>(1)];
49
+ tensor<fp16, [512, 512, 3]> weight_1_to_fp16 = const()[name = tensor<string, []>("weight_1_to_fp16"), val = tensor<fp16, [512, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(3939968)))];
50
+ tensor<fp16, [512]> F0_0_conv1_bias_to_fp16 = const()[name = tensor<string, []>("F0_0_conv1_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5512896)))];
51
+ tensor<fp16, [1, 512, ?]> input_11_cast_fp16 = conv(bias = F0_0_conv1_bias_to_fp16, dilations = input_11_dilations_0, groups = input_11_groups_0, pad = input_11_pad_0, pad_type = input_11_pad_type_0, strides = input_11_strides_0, weight = weight_1_to_fp16, x = input_7_cast_fp16)[name = tensor<string, []>("input_11_cast_fp16")];
52
+ tensor<fp16, [1024, 128]> F0_0_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_0_norm2_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5513984)))];
53
+ tensor<fp16, [1024]> F0_0_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_0_norm2_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5776192)))];
54
+ tensor<fp16, [1, 1024]> linear_1_cast_fp16 = linear(bias = F0_0_norm2_fc_bias_to_fp16, weight = F0_0_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_1_cast_fp16")];
55
+ tensor<int32, [3]> var_102 = const()[name = tensor<string, []>("op_102"), val = tensor<int32, [3]>([1, 1024, 1])];
56
+ tensor<fp16, [1, 1024, 1]> h_7_cast_fp16 = reshape(shape = var_102, x = linear_1_cast_fp16)[name = tensor<string, []>("h_7_cast_fp16")];
57
+ tensor<int32, [2]> var_104_split_sizes_0 = const()[name = tensor<string, []>("op_104_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
58
+ tensor<int32, []> var_104_axis_0 = const()[name = tensor<string, []>("op_104_axis_0"), val = tensor<int32, []>(1)];
59
+ tensor<fp16, [1, 512, 1]> var_104_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_104_cast_fp16_1 = split(axis = var_104_axis_0, split_sizes = var_104_split_sizes_0, x = h_7_cast_fp16)[name = tensor<string, []>("op_104_cast_fp16")];
60
+ tensor<fp16, []> var_106_promoted_to_fp16 = const()[name = tensor<string, []>("op_106_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
61
+ tensor<fp16, [1, 512, 1]> var_107_cast_fp16 = add(x = var_104_cast_fp16_0, y = var_106_promoted_to_fp16)[name = tensor<string, []>("op_107_cast_fp16")];
62
+ tensor<fp16, [1, 512, ?]> var_108_cast_fp16 = instance_norm(epsilon = var_56_to_fp16, x = input_11_cast_fp16)[name = tensor<string, []>("op_108_cast_fp16")];
63
+ tensor<fp16, [1, 512, ?]> var_109_cast_fp16 = mul(x = var_107_cast_fp16, y = var_108_cast_fp16)[name = tensor<string, []>("op_109_cast_fp16")];
64
+ tensor<fp16, [1, 512, ?]> input_13_cast_fp16 = add(x = var_109_cast_fp16, y = var_104_cast_fp16_1)[name = tensor<string, []>("input_13_cast_fp16")];
65
+ tensor<fp16, [1, 512, ?]> input_15_cast_fp16 = leaky_relu(alpha = var_53, x = input_13_cast_fp16)[name = tensor<string, []>("input_15_cast_fp16")];
66
+ tensor<string, []> out_1_pad_type_0 = const()[name = tensor<string, []>("out_1_pad_type_0"), val = tensor<string, []>("custom")];
67
+ tensor<int32, [2]> out_1_pad_0 = const()[name = tensor<string, []>("out_1_pad_0"), val = tensor<int32, [2]>([1, 1])];
68
+ tensor<int32, [1]> out_1_strides_0 = const()[name = tensor<string, []>("out_1_strides_0"), val = tensor<int32, [1]>([1])];
69
+ tensor<int32, [1]> out_1_dilations_0 = const()[name = tensor<string, []>("out_1_dilations_0"), val = tensor<int32, [1]>([1])];
70
+ tensor<int32, []> out_1_groups_0 = const()[name = tensor<string, []>("out_1_groups_0"), val = tensor<int32, []>(1)];
71
+ tensor<fp16, [512, 512, 3]> weight_3_to_fp16 = const()[name = tensor<string, []>("weight_3_to_fp16"), val = tensor<fp16, [512, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(5778304)))];
72
+ tensor<fp16, [512]> F0_0_conv2_bias_to_fp16 = const()[name = tensor<string, []>("F0_0_conv2_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7351232)))];
73
+ tensor<fp16, [1, 512, ?]> out_1_cast_fp16 = conv(bias = F0_0_conv2_bias_to_fp16, dilations = out_1_dilations_0, groups = out_1_groups_0, pad = out_1_pad_0, pad_type = out_1_pad_type_0, strides = out_1_strides_0, weight = weight_3_to_fp16, x = input_15_cast_fp16)[name = tensor<string, []>("out_1_cast_fp16")];
74
+ tensor<fp16, [1, 512, ?]> var_120_cast_fp16 = add(x = out_1_cast_fp16, y = input_3_cast_fp16)[name = tensor<string, []>("op_120_cast_fp16")];
75
+ tensor<fp16, []> _inversed_input_19_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_19_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
76
+ tensor<fp16, [1, 512, ?]> _inversed_input_19_cast_fp16 = mul(x = var_120_cast_fp16, y = _inversed_input_19_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_19_cast_fp16")];
77
+ tensor<fp32, []> var_126 = const()[name = tensor<string, []>("op_126"), val = tensor<fp32, []>(0x1.99999ap-3)];
78
+ tensor<fp16, [1024, 128]> F0_1_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_1_norm1_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7352320)))];
79
+ tensor<fp16, [1024]> F0_1_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_1_norm1_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7614528)))];
80
+ tensor<fp16, [1, 1024]> linear_2_cast_fp16 = linear(bias = F0_1_norm1_fc_bias_to_fp16, weight = F0_1_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_2_cast_fp16")];
81
+ tensor<int32, [3]> var_161 = const()[name = tensor<string, []>("op_161"), val = tensor<int32, [3]>([1, 1024, 1])];
82
+ tensor<fp16, [1, 1024, 1]> h_11_cast_fp16 = reshape(shape = var_161, x = linear_2_cast_fp16)[name = tensor<string, []>("h_11_cast_fp16")];
83
+ tensor<int32, [2]> var_163_split_sizes_0 = const()[name = tensor<string, []>("op_163_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
84
+ tensor<int32, []> var_163_axis_0 = const()[name = tensor<string, []>("op_163_axis_0"), val = tensor<int32, []>(1)];
85
+ tensor<fp16, [1, 512, 1]> var_163_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_163_cast_fp16_1 = split(axis = var_163_axis_0, split_sizes = var_163_split_sizes_0, x = h_11_cast_fp16)[name = tensor<string, []>("op_163_cast_fp16")];
86
+ tensor<fp16, []> var_165_promoted_to_fp16 = const()[name = tensor<string, []>("op_165_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
87
+ tensor<fp16, [1, 512, 1]> var_166_cast_fp16 = add(x = var_163_cast_fp16_0, y = var_165_promoted_to_fp16)[name = tensor<string, []>("op_166_cast_fp16")];
88
+ tensor<fp16, []> var_130_to_fp16 = const()[name = tensor<string, []>("op_130_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
89
+ tensor<fp16, [1, 512, ?]> var_167_cast_fp16 = instance_norm(epsilon = var_130_to_fp16, x = _inversed_input_19_cast_fp16)[name = tensor<string, []>("op_167_cast_fp16")];
90
+ tensor<fp16, [1, 512, ?]> var_168_cast_fp16 = mul(x = var_166_cast_fp16, y = var_167_cast_fp16)[name = tensor<string, []>("op_168_cast_fp16")];
91
+ tensor<fp16, [1, 512, ?]> input_21_cast_fp16 = add(x = var_168_cast_fp16, y = var_163_cast_fp16_1)[name = tensor<string, []>("input_21_cast_fp16")];
92
+ tensor<fp16, [1, 512, ?]> input_23_cast_fp16 = leaky_relu(alpha = var_126, x = input_21_cast_fp16)[name = tensor<string, []>("input_23_cast_fp16")];
93
+ tensor<string, []> conv_transpose_0_pad_type_0 = const()[name = tensor<string, []>("conv_transpose_0_pad_type_0"), val = tensor<string, []>("custom")];
94
+ tensor<int32, [2]> conv_transpose_0_pad_0 = const()[name = tensor<string, []>("conv_transpose_0_pad_0"), val = tensor<int32, [2]>([0, 0])];
95
+ tensor<int32, [1]> conv_transpose_0_strides_0 = const()[name = tensor<string, []>("conv_transpose_0_strides_0"), val = tensor<int32, [1]>([2])];
96
+ tensor<int32, []> conv_transpose_0_groups_0 = const()[name = tensor<string, []>("conv_transpose_0_groups_0"), val = tensor<int32, []>(512)];
97
+ tensor<int32, [1]> conv_transpose_0_dilations_0 = const()[name = tensor<string, []>("conv_transpose_0_dilations_0"), val = tensor<int32, [1]>([1])];
98
+ tensor<fp16, [512, 1, 3]> var_171_to_fp16 = const()[name = tensor<string, []>("op_171_to_fp16"), val = tensor<fp16, [512, 1, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7616640)))];
99
+ tensor<fp16, [512]> F0_1_pool_bias_to_fp16 = const()[name = tensor<string, []>("F0_1_pool_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7619776)))];
100
+ tensor<fp16, [1, 512, ?]> conv_transpose_0_cast_fp16 = conv_transpose(bias = F0_1_pool_bias_to_fp16, dilations = conv_transpose_0_dilations_0, groups = conv_transpose_0_groups_0, pad = conv_transpose_0_pad_0, pad_type = conv_transpose_0_pad_type_0, strides = conv_transpose_0_strides_0, weight = var_171_to_fp16, x = input_23_cast_fp16)[name = tensor<string, []>("conv_transpose_0_cast_fp16")];
101
+ tensor<int32, [3]> input_25_begin_0 = const()[name = tensor<string, []>("input_25_begin_0"), val = tensor<int32, [3]>([0, 0, 1])];
102
+ tensor<int32, [3]> input_25_end_0 = const()[name = tensor<string, []>("input_25_end_0"), val = tensor<int32, [3]>([0, 0, 0])];
103
+ tensor<bool, [3]> input_25_begin_mask_0 = const()[name = tensor<string, []>("input_25_begin_mask_0"), val = tensor<bool, [3]>([true, true, false])];
104
+ tensor<bool, [3]> input_25_end_mask_0 = const()[name = tensor<string, []>("input_25_end_mask_0"), val = tensor<bool, [3]>([true, true, true])];
105
+ tensor<fp16, [1, 512, ?]> input_25_cast_fp16 = slice_by_index(begin = input_25_begin_0, begin_mask = input_25_begin_mask_0, end = input_25_end_0, end_mask = input_25_end_mask_0, x = conv_transpose_0_cast_fp16)[name = tensor<string, []>("input_25_cast_fp16")];
106
+ tensor<string, []> input_29_pad_type_0 = const()[name = tensor<string, []>("input_29_pad_type_0"), val = tensor<string, []>("custom")];
107
+ tensor<int32, [2]> input_29_pad_0 = const()[name = tensor<string, []>("input_29_pad_0"), val = tensor<int32, [2]>([1, 1])];
108
+ tensor<int32, [1]> input_29_strides_0 = const()[name = tensor<string, []>("input_29_strides_0"), val = tensor<int32, [1]>([1])];
109
+ tensor<int32, [1]> input_29_dilations_0 = const()[name = tensor<string, []>("input_29_dilations_0"), val = tensor<int32, [1]>([1])];
110
+ tensor<int32, []> input_29_groups_0 = const()[name = tensor<string, []>("input_29_groups_0"), val = tensor<int32, []>(1)];
111
+ tensor<fp16, [256, 512, 3]> weight_5_to_fp16 = const()[name = tensor<string, []>("weight_5_to_fp16"), val = tensor<fp16, [256, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(7620864)))];
112
+ tensor<fp16, [256]> F0_1_conv1_bias_to_fp16 = const()[name = tensor<string, []>("F0_1_conv1_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8407360)))];
113
+ tensor<fp16, [1, 256, ?]> input_29_cast_fp16 = conv(bias = F0_1_conv1_bias_to_fp16, dilations = input_29_dilations_0, groups = input_29_groups_0, pad = input_29_pad_0, pad_type = input_29_pad_type_0, strides = input_29_strides_0, weight = weight_5_to_fp16, x = input_25_cast_fp16)[name = tensor<string, []>("input_29_cast_fp16")];
114
+ tensor<fp16, [512, 128]> F0_1_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_1_norm2_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8407936)))];
115
+ tensor<fp16, [512]> F0_1_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_1_norm2_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8539072)))];
116
+ tensor<fp16, [1, 512]> linear_3_cast_fp16 = linear(bias = F0_1_norm2_fc_bias_to_fp16, weight = F0_1_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_3_cast_fp16")];
117
+ tensor<int32, [3]> var_192 = const()[name = tensor<string, []>("op_192"), val = tensor<int32, [3]>([1, 512, 1])];
118
+ tensor<fp16, [1, 512, 1]> h_15_cast_fp16 = reshape(shape = var_192, x = linear_3_cast_fp16)[name = tensor<string, []>("h_15_cast_fp16")];
119
+ tensor<int32, [2]> var_194_split_sizes_0 = const()[name = tensor<string, []>("op_194_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
120
+ tensor<int32, []> var_194_axis_0 = const()[name = tensor<string, []>("op_194_axis_0"), val = tensor<int32, []>(1)];
121
+ tensor<fp16, [1, 256, 1]> var_194_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_194_cast_fp16_1 = split(axis = var_194_axis_0, split_sizes = var_194_split_sizes_0, x = h_15_cast_fp16)[name = tensor<string, []>("op_194_cast_fp16")];
122
+ tensor<fp16, []> var_196_promoted_to_fp16 = const()[name = tensor<string, []>("op_196_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
123
+ tensor<fp16, [1, 256, 1]> var_197_cast_fp16 = add(x = var_194_cast_fp16_0, y = var_196_promoted_to_fp16)[name = tensor<string, []>("op_197_cast_fp16")];
124
+ tensor<fp16, [1, 256, ?]> var_198_cast_fp16 = instance_norm(epsilon = var_130_to_fp16, x = input_29_cast_fp16)[name = tensor<string, []>("op_198_cast_fp16")];
125
+ tensor<fp16, [1, 256, ?]> var_199_cast_fp16 = mul(x = var_197_cast_fp16, y = var_198_cast_fp16)[name = tensor<string, []>("op_199_cast_fp16")];
126
+ tensor<fp16, [1, 256, ?]> input_31_cast_fp16 = add(x = var_199_cast_fp16, y = var_194_cast_fp16_1)[name = tensor<string, []>("input_31_cast_fp16")];
127
+ tensor<fp16, [1, 256, ?]> input_33_cast_fp16 = leaky_relu(alpha = var_126, x = input_31_cast_fp16)[name = tensor<string, []>("input_33_cast_fp16")];
128
+ tensor<string, []> out_3_pad_type_0 = const()[name = tensor<string, []>("out_3_pad_type_0"), val = tensor<string, []>("custom")];
129
+ tensor<int32, [2]> out_3_pad_0 = const()[name = tensor<string, []>("out_3_pad_0"), val = tensor<int32, [2]>([1, 1])];
130
+ tensor<int32, [1]> out_3_strides_0 = const()[name = tensor<string, []>("out_3_strides_0"), val = tensor<int32, [1]>([1])];
131
+ tensor<int32, [1]> out_3_dilations_0 = const()[name = tensor<string, []>("out_3_dilations_0"), val = tensor<int32, [1]>([1])];
132
+ tensor<int32, []> out_3_groups_0 = const()[name = tensor<string, []>("out_3_groups_0"), val = tensor<int32, []>(1)];
133
+ tensor<fp16, [256, 256, 3]> weight_7_to_fp16 = const()[name = tensor<string, []>("weight_7_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8540160)))];
134
+ tensor<fp16, [256]> F0_1_conv2_bias_to_fp16 = const()[name = tensor<string, []>("F0_1_conv2_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8933440)))];
135
+ tensor<fp16, [1, 256, ?]> out_3_cast_fp16 = conv(bias = F0_1_conv2_bias_to_fp16, dilations = out_3_dilations_0, groups = out_3_groups_0, pad = out_3_pad_0, pad_type = out_3_pad_type_0, strides = out_3_strides_0, weight = weight_7_to_fp16, x = input_33_cast_fp16)[name = tensor<string, []>("out_3_cast_fp16")];
136
+ tensor<int32, [1]> expand_dims_0_axes_0 = const()[name = tensor<string, []>("expand_dims_0_axes_0"), val = tensor<int32, [1]>([3])];
137
+ tensor<fp16, [1, 512, ?, 1]> expand_dims_0_cast_fp16 = expand_dims(axes = expand_dims_0_axes_0, x = _inversed_input_19_cast_fp16)[name = tensor<string, []>("expand_dims_0_cast_fp16")];
138
+ tensor<int32, []> upsample_nearest_neighbor_0_scale_factor_height_0 = const()[name = tensor<string, []>("upsample_nearest_neighbor_0_scale_factor_height_0"), val = tensor<int32, []>(2)];
139
+ tensor<int32, []> upsample_nearest_neighbor_0_scale_factor_width_0 = const()[name = tensor<string, []>("upsample_nearest_neighbor_0_scale_factor_width_0"), val = tensor<int32, []>(1)];
140
+ tensor<fp16, [1, 512, ?, 1]> upsample_nearest_neighbor_0_cast_fp16 = upsample_nearest_neighbor(scale_factor_height = upsample_nearest_neighbor_0_scale_factor_height_0, scale_factor_width = upsample_nearest_neighbor_0_scale_factor_width_0, x = expand_dims_0_cast_fp16)[name = tensor<string, []>("upsample_nearest_neighbor_0_cast_fp16")];
141
+ tensor<int32, [1]> input_37_axes_0 = const()[name = tensor<string, []>("input_37_axes_0"), val = tensor<int32, [1]>([3])];
142
+ tensor<fp16, [1, 512, ?]> input_37_cast_fp16 = squeeze(axes = input_37_axes_0, x = upsample_nearest_neighbor_0_cast_fp16)[name = tensor<string, []>("input_37_cast_fp16")];
143
+ tensor<string, []> var_217_pad_type_0 = const()[name = tensor<string, []>("op_217_pad_type_0"), val = tensor<string, []>("valid")];
144
+ tensor<int32, [1]> var_217_strides_0 = const()[name = tensor<string, []>("op_217_strides_0"), val = tensor<int32, [1]>([1])];
145
+ tensor<int32, [2]> var_217_pad_0 = const()[name = tensor<string, []>("op_217_pad_0"), val = tensor<int32, [2]>([0, 0])];
146
+ tensor<int32, [1]> var_217_dilations_0 = const()[name = tensor<string, []>("op_217_dilations_0"), val = tensor<int32, [1]>([1])];
147
+ tensor<int32, []> var_217_groups_0 = const()[name = tensor<string, []>("op_217_groups_0"), val = tensor<int32, []>(1)];
148
+ tensor<fp16, [256, 512, 1]> weight_9_to_fp16 = const()[name = tensor<string, []>("weight_9_to_fp16"), val = tensor<fp16, [256, 512, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(8934016)))];
149
+ tensor<fp16, [1, 256, ?]> var_217_cast_fp16 = conv(dilations = var_217_dilations_0, groups = var_217_groups_0, pad = var_217_pad_0, pad_type = var_217_pad_type_0, strides = var_217_strides_0, weight = weight_9_to_fp16, x = input_37_cast_fp16)[name = tensor<string, []>("op_217_cast_fp16")];
150
+ tensor<fp16, [1, 256, ?]> var_218_cast_fp16 = add(x = out_3_cast_fp16, y = var_217_cast_fp16)[name = tensor<string, []>("op_218_cast_fp16")];
151
+ tensor<fp16, []> _inversed_input_39_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_39_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
152
+ tensor<fp16, [1, 256, ?]> _inversed_input_39_cast_fp16 = mul(x = var_218_cast_fp16, y = _inversed_input_39_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_39_cast_fp16")];
153
+ tensor<fp32, []> var_222 = const()[name = tensor<string, []>("op_222"), val = tensor<fp32, []>(0x1.99999ap-3)];
154
+ tensor<fp16, [512, 128]> F0_2_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_2_norm1_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9196224)))];
155
+ tensor<fp16, [512]> F0_2_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_2_norm1_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9327360)))];
156
+ tensor<fp16, [1, 512]> linear_4_cast_fp16 = linear(bias = F0_2_norm1_fc_bias_to_fp16, weight = F0_2_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_4_cast_fp16")];
157
+ tensor<int32, [3]> var_247 = const()[name = tensor<string, []>("op_247"), val = tensor<int32, [3]>([1, 512, 1])];
158
+ tensor<fp16, [1, 512, 1]> h_19_cast_fp16 = reshape(shape = var_247, x = linear_4_cast_fp16)[name = tensor<string, []>("h_19_cast_fp16")];
159
+ tensor<int32, [2]> var_249_split_sizes_0 = const()[name = tensor<string, []>("op_249_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
160
+ tensor<int32, []> var_249_axis_0 = const()[name = tensor<string, []>("op_249_axis_0"), val = tensor<int32, []>(1)];
161
+ tensor<fp16, [1, 256, 1]> var_249_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_249_cast_fp16_1 = split(axis = var_249_axis_0, split_sizes = var_249_split_sizes_0, x = h_19_cast_fp16)[name = tensor<string, []>("op_249_cast_fp16")];
162
+ tensor<fp16, []> var_251_promoted_to_fp16 = const()[name = tensor<string, []>("op_251_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
163
+ tensor<fp16, [1, 256, 1]> var_252_cast_fp16 = add(x = var_249_cast_fp16_0, y = var_251_promoted_to_fp16)[name = tensor<string, []>("op_252_cast_fp16")];
164
+ tensor<fp16, []> var_225_to_fp16 = const()[name = tensor<string, []>("op_225_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
165
+ tensor<fp16, [1, 256, ?]> var_253_cast_fp16 = instance_norm(epsilon = var_225_to_fp16, x = _inversed_input_39_cast_fp16)[name = tensor<string, []>("op_253_cast_fp16")];
166
+ tensor<fp16, [1, 256, ?]> var_254_cast_fp16 = mul(x = var_252_cast_fp16, y = var_253_cast_fp16)[name = tensor<string, []>("op_254_cast_fp16")];
167
+ tensor<fp16, [1, 256, ?]> input_41_cast_fp16 = add(x = var_254_cast_fp16, y = var_249_cast_fp16_1)[name = tensor<string, []>("input_41_cast_fp16")];
168
+ tensor<fp16, [1, 256, ?]> input_43_cast_fp16 = leaky_relu(alpha = var_222, x = input_41_cast_fp16)[name = tensor<string, []>("input_43_cast_fp16")];
169
+ tensor<string, []> input_47_pad_type_0 = const()[name = tensor<string, []>("input_47_pad_type_0"), val = tensor<string, []>("custom")];
170
+ tensor<int32, [2]> input_47_pad_0 = const()[name = tensor<string, []>("input_47_pad_0"), val = tensor<int32, [2]>([1, 1])];
171
+ tensor<int32, [1]> input_47_strides_0 = const()[name = tensor<string, []>("input_47_strides_0"), val = tensor<int32, [1]>([1])];
172
+ tensor<int32, [1]> input_47_dilations_0 = const()[name = tensor<string, []>("input_47_dilations_0"), val = tensor<int32, [1]>([1])];
173
+ tensor<int32, []> input_47_groups_0 = const()[name = tensor<string, []>("input_47_groups_0"), val = tensor<int32, []>(1)];
174
+ tensor<fp16, [256, 256, 3]> weight_11_to_fp16 = const()[name = tensor<string, []>("weight_11_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9328448)))];
175
+ tensor<fp16, [256]> F0_2_conv1_bias_to_fp16 = const()[name = tensor<string, []>("F0_2_conv1_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9721728)))];
176
+ tensor<fp16, [1, 256, ?]> input_47_cast_fp16 = conv(bias = F0_2_conv1_bias_to_fp16, dilations = input_47_dilations_0, groups = input_47_groups_0, pad = input_47_pad_0, pad_type = input_47_pad_type_0, strides = input_47_strides_0, weight = weight_11_to_fp16, x = input_43_cast_fp16)[name = tensor<string, []>("input_47_cast_fp16")];
177
+ tensor<fp16, [512, 128]> F0_2_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("F0_2_norm2_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9722304)))];
178
+ tensor<fp16, [512]> F0_2_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("F0_2_norm2_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9853440)))];
179
+ tensor<fp16, [1, 512]> linear_5_cast_fp16 = linear(bias = F0_2_norm2_fc_bias_to_fp16, weight = F0_2_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_5_cast_fp16")];
180
+ tensor<int32, [3]> var_271 = const()[name = tensor<string, []>("op_271"), val = tensor<int32, [3]>([1, 512, 1])];
181
+ tensor<fp16, [1, 512, 1]> h_23_cast_fp16 = reshape(shape = var_271, x = linear_5_cast_fp16)[name = tensor<string, []>("h_23_cast_fp16")];
182
+ tensor<int32, [2]> var_273_split_sizes_0 = const()[name = tensor<string, []>("op_273_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
183
+ tensor<int32, []> var_273_axis_0 = const()[name = tensor<string, []>("op_273_axis_0"), val = tensor<int32, []>(1)];
184
+ tensor<fp16, [1, 256, 1]> var_273_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_273_cast_fp16_1 = split(axis = var_273_axis_0, split_sizes = var_273_split_sizes_0, x = h_23_cast_fp16)[name = tensor<string, []>("op_273_cast_fp16")];
185
+ tensor<fp16, []> var_275_promoted_to_fp16 = const()[name = tensor<string, []>("op_275_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
186
+ tensor<fp16, [1, 256, 1]> var_276_cast_fp16 = add(x = var_273_cast_fp16_0, y = var_275_promoted_to_fp16)[name = tensor<string, []>("op_276_cast_fp16")];
187
+ tensor<fp16, [1, 256, ?]> var_277_cast_fp16 = instance_norm(epsilon = var_225_to_fp16, x = input_47_cast_fp16)[name = tensor<string, []>("op_277_cast_fp16")];
188
+ tensor<fp16, [1, 256, ?]> var_278_cast_fp16 = mul(x = var_276_cast_fp16, y = var_277_cast_fp16)[name = tensor<string, []>("op_278_cast_fp16")];
189
+ tensor<fp16, [1, 256, ?]> input_49_cast_fp16 = add(x = var_278_cast_fp16, y = var_273_cast_fp16_1)[name = tensor<string, []>("input_49_cast_fp16")];
190
+ tensor<fp16, [1, 256, ?]> input_51_cast_fp16 = leaky_relu(alpha = var_222, x = input_49_cast_fp16)[name = tensor<string, []>("input_51_cast_fp16")];
191
+ tensor<string, []> out_5_pad_type_0 = const()[name = tensor<string, []>("out_5_pad_type_0"), val = tensor<string, []>("custom")];
192
+ tensor<int32, [2]> out_5_pad_0 = const()[name = tensor<string, []>("out_5_pad_0"), val = tensor<int32, [2]>([1, 1])];
193
+ tensor<int32, [1]> out_5_strides_0 = const()[name = tensor<string, []>("out_5_strides_0"), val = tensor<int32, [1]>([1])];
194
+ tensor<int32, [1]> out_5_dilations_0 = const()[name = tensor<string, []>("out_5_dilations_0"), val = tensor<int32, [1]>([1])];
195
+ tensor<int32, []> out_5_groups_0 = const()[name = tensor<string, []>("out_5_groups_0"), val = tensor<int32, []>(1)];
196
+ tensor<fp16, [256, 256, 3]> weight_13_to_fp16 = const()[name = tensor<string, []>("weight_13_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(9854528)))];
197
+ tensor<fp16, [256]> F0_2_conv2_bias_to_fp16 = const()[name = tensor<string, []>("F0_2_conv2_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10247808)))];
198
+ tensor<fp16, [1, 256, ?]> out_5_cast_fp16 = conv(bias = F0_2_conv2_bias_to_fp16, dilations = out_5_dilations_0, groups = out_5_groups_0, pad = out_5_pad_0, pad_type = out_5_pad_type_0, strides = out_5_strides_0, weight = weight_13_to_fp16, x = input_51_cast_fp16)[name = tensor<string, []>("out_5_cast_fp16")];
199
+ tensor<fp16, [1, 256, ?]> var_289_cast_fp16 = add(x = out_5_cast_fp16, y = _inversed_input_39_cast_fp16)[name = tensor<string, []>("op_289_cast_fp16")];
200
+ tensor<fp16, []> _inversed_input_55_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_55_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
201
+ tensor<fp16, [1, 256, ?]> _inversed_input_55_cast_fp16 = mul(x = var_289_cast_fp16, y = _inversed_input_55_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_55_cast_fp16")];
202
+ tensor<string, []> var_302_pad_type_0 = const()[name = tensor<string, []>("op_302_pad_type_0"), val = tensor<string, []>("valid")];
203
+ tensor<int32, [1]> var_302_strides_0 = const()[name = tensor<string, []>("op_302_strides_0"), val = tensor<int32, [1]>([1])];
204
+ tensor<int32, [2]> var_302_pad_0 = const()[name = tensor<string, []>("op_302_pad_0"), val = tensor<int32, [2]>([0, 0])];
205
+ tensor<int32, [1]> var_302_dilations_0 = const()[name = tensor<string, []>("op_302_dilations_0"), val = tensor<int32, [1]>([1])];
206
+ tensor<int32, []> var_302_groups_0 = const()[name = tensor<string, []>("op_302_groups_0"), val = tensor<int32, []>(1)];
207
+ tensor<fp16, [1, 256, 1]> F0_proj_weight_to_fp16 = const()[name = tensor<string, []>("F0_proj_weight_to_fp16"), val = tensor<fp16, [1, 256, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10248384)))];
208
+ tensor<fp16, [1]> F0_proj_bias_to_fp16 = const()[name = tensor<string, []>("F0_proj_bias_to_fp16"), val = tensor<fp16, [1]>([0x1.adp-4])];
209
+ tensor<fp16, [1, 1, ?]> var_302_cast_fp16 = conv(bias = F0_proj_bias_to_fp16, dilations = var_302_dilations_0, groups = var_302_groups_0, pad = var_302_pad_0, pad_type = var_302_pad_type_0, strides = var_302_strides_0, weight = F0_proj_weight_to_fp16, x = _inversed_input_55_cast_fp16)[name = tensor<string, []>("op_302_cast_fp16")];
210
+ tensor<int32, [1]> var_304_axes_0 = const()[name = tensor<string, []>("op_304_axes_0"), val = tensor<int32, [1]>([1])];
211
+ tensor<fp16, [1, ?]> F0 = squeeze(axes = var_304_axes_0, x = var_302_cast_fp16)[name = tensor<string, []>("op_304_cast_fp16")];
212
+ tensor<fp32, []> var_309 = const()[name = tensor<string, []>("op_309"), val = tensor<fp32, []>(0x1.99999ap-3)];
213
+ tensor<fp16, [1024, 128]> N_0_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_0_norm1_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10248960)))];
214
+ tensor<fp16, [1024]> N_0_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_0_norm1_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10511168)))];
215
+ tensor<fp16, [1, 1024]> linear_6_cast_fp16 = linear(bias = N_0_norm1_fc_bias_to_fp16, weight = N_0_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_6_cast_fp16")];
216
+ tensor<int32, [3]> var_334 = const()[name = tensor<string, []>("op_334"), val = tensor<int32, [3]>([1, 1024, 1])];
217
+ tensor<fp16, [1, 1024, 1]> h_27_cast_fp16 = reshape(shape = var_334, x = linear_6_cast_fp16)[name = tensor<string, []>("h_27_cast_fp16")];
218
+ tensor<int32, [2]> var_336_split_sizes_0 = const()[name = tensor<string, []>("op_336_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
219
+ tensor<int32, []> var_336_axis_0 = const()[name = tensor<string, []>("op_336_axis_0"), val = tensor<int32, []>(1)];
220
+ tensor<fp16, [1, 512, 1]> var_336_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_336_cast_fp16_1 = split(axis = var_336_axis_0, split_sizes = var_336_split_sizes_0, x = h_27_cast_fp16)[name = tensor<string, []>("op_336_cast_fp16")];
221
+ tensor<fp16, []> var_338_promoted_to_fp16 = const()[name = tensor<string, []>("op_338_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
222
+ tensor<fp16, [1, 512, 1]> var_339_cast_fp16 = add(x = var_336_cast_fp16_0, y = var_338_promoted_to_fp16)[name = tensor<string, []>("op_339_cast_fp16")];
223
+ tensor<fp16, [1, 512, ?]> var_341_cast_fp16 = mul(x = var_339_cast_fp16, y = var_84_cast_fp16)[name = tensor<string, []>("op_341_cast_fp16")];
224
+ tensor<fp16, [1, 512, ?]> input_59_cast_fp16 = add(x = var_341_cast_fp16, y = var_336_cast_fp16_1)[name = tensor<string, []>("input_59_cast_fp16")];
225
+ tensor<fp16, [1, 512, ?]> input_61_cast_fp16 = leaky_relu(alpha = var_309, x = input_59_cast_fp16)[name = tensor<string, []>("input_61_cast_fp16")];
226
+ tensor<string, []> input_65_pad_type_0 = const()[name = tensor<string, []>("input_65_pad_type_0"), val = tensor<string, []>("custom")];
227
+ tensor<int32, [2]> input_65_pad_0 = const()[name = tensor<string, []>("input_65_pad_0"), val = tensor<int32, [2]>([1, 1])];
228
+ tensor<int32, [1]> input_65_strides_0 = const()[name = tensor<string, []>("input_65_strides_0"), val = tensor<int32, [1]>([1])];
229
+ tensor<int32, [1]> input_65_dilations_0 = const()[name = tensor<string, []>("input_65_dilations_0"), val = tensor<int32, [1]>([1])];
230
+ tensor<int32, []> input_65_groups_0 = const()[name = tensor<string, []>("input_65_groups_0"), val = tensor<int32, []>(1)];
231
+ tensor<fp16, [512, 512, 3]> weight_17_to_fp16 = const()[name = tensor<string, []>("weight_17_to_fp16"), val = tensor<fp16, [512, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(10513280)))];
232
+ tensor<fp16, [512]> N_0_conv1_bias_to_fp16 = const()[name = tensor<string, []>("N_0_conv1_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(12086208)))];
233
+ tensor<fp16, [1, 512, ?]> input_65_cast_fp16 = conv(bias = N_0_conv1_bias_to_fp16, dilations = input_65_dilations_0, groups = input_65_groups_0, pad = input_65_pad_0, pad_type = input_65_pad_type_0, strides = input_65_strides_0, weight = weight_17_to_fp16, x = input_61_cast_fp16)[name = tensor<string, []>("input_65_cast_fp16")];
234
+ tensor<fp16, [1024, 128]> N_0_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_0_norm2_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(12087296)))];
235
+ tensor<fp16, [1024]> N_0_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_0_norm2_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(12349504)))];
236
+ tensor<fp16, [1, 1024]> linear_7_cast_fp16 = linear(bias = N_0_norm2_fc_bias_to_fp16, weight = N_0_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_7_cast_fp16")];
237
+ tensor<int32, [3]> var_358 = const()[name = tensor<string, []>("op_358"), val = tensor<int32, [3]>([1, 1024, 1])];
238
+ tensor<fp16, [1, 1024, 1]> h_31_cast_fp16 = reshape(shape = var_358, x = linear_7_cast_fp16)[name = tensor<string, []>("h_31_cast_fp16")];
239
+ tensor<int32, [2]> var_360_split_sizes_0 = const()[name = tensor<string, []>("op_360_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
240
+ tensor<int32, []> var_360_axis_0 = const()[name = tensor<string, []>("op_360_axis_0"), val = tensor<int32, []>(1)];
241
+ tensor<fp16, [1, 512, 1]> var_360_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_360_cast_fp16_1 = split(axis = var_360_axis_0, split_sizes = var_360_split_sizes_0, x = h_31_cast_fp16)[name = tensor<string, []>("op_360_cast_fp16")];
242
+ tensor<fp16, []> var_362_promoted_to_fp16 = const()[name = tensor<string, []>("op_362_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
243
+ tensor<fp16, [1, 512, 1]> var_363_cast_fp16 = add(x = var_360_cast_fp16_0, y = var_362_promoted_to_fp16)[name = tensor<string, []>("op_363_cast_fp16")];
244
+ tensor<fp16, []> var_312_to_fp16 = const()[name = tensor<string, []>("op_312_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
245
+ tensor<fp16, [1, 512, ?]> var_364_cast_fp16 = instance_norm(epsilon = var_312_to_fp16, x = input_65_cast_fp16)[name = tensor<string, []>("op_364_cast_fp16")];
246
+ tensor<fp16, [1, 512, ?]> var_365_cast_fp16 = mul(x = var_363_cast_fp16, y = var_364_cast_fp16)[name = tensor<string, []>("op_365_cast_fp16")];
247
+ tensor<fp16, [1, 512, ?]> input_67_cast_fp16 = add(x = var_365_cast_fp16, y = var_360_cast_fp16_1)[name = tensor<string, []>("input_67_cast_fp16")];
248
+ tensor<fp16, [1, 512, ?]> input_69_cast_fp16 = leaky_relu(alpha = var_309, x = input_67_cast_fp16)[name = tensor<string, []>("input_69_cast_fp16")];
249
+ tensor<string, []> out_7_pad_type_0 = const()[name = tensor<string, []>("out_7_pad_type_0"), val = tensor<string, []>("custom")];
250
+ tensor<int32, [2]> out_7_pad_0 = const()[name = tensor<string, []>("out_7_pad_0"), val = tensor<int32, [2]>([1, 1])];
251
+ tensor<int32, [1]> out_7_strides_0 = const()[name = tensor<string, []>("out_7_strides_0"), val = tensor<int32, [1]>([1])];
252
+ tensor<int32, [1]> out_7_dilations_0 = const()[name = tensor<string, []>("out_7_dilations_0"), val = tensor<int32, [1]>([1])];
253
+ tensor<int32, []> out_7_groups_0 = const()[name = tensor<string, []>("out_7_groups_0"), val = tensor<int32, []>(1)];
254
+ tensor<fp16, [512, 512, 3]> weight_19_to_fp16 = const()[name = tensor<string, []>("weight_19_to_fp16"), val = tensor<fp16, [512, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(12351616)))];
255
+ tensor<fp16, [512]> N_0_conv2_bias_to_fp16 = const()[name = tensor<string, []>("N_0_conv2_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(13924544)))];
256
+ tensor<fp16, [1, 512, ?]> out_7_cast_fp16 = conv(bias = N_0_conv2_bias_to_fp16, dilations = out_7_dilations_0, groups = out_7_groups_0, pad = out_7_pad_0, pad_type = out_7_pad_type_0, strides = out_7_strides_0, weight = weight_19_to_fp16, x = input_69_cast_fp16)[name = tensor<string, []>("out_7_cast_fp16")];
257
+ tensor<fp16, [1, 512, ?]> var_376_cast_fp16 = add(x = out_7_cast_fp16, y = input_3_cast_fp16)[name = tensor<string, []>("op_376_cast_fp16")];
258
+ tensor<fp16, []> _inversed_input_73_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_73_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
259
+ tensor<fp16, [1, 512, ?]> _inversed_input_73_cast_fp16 = mul(x = var_376_cast_fp16, y = _inversed_input_73_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_73_cast_fp16")];
260
+ tensor<fp32, []> var_382 = const()[name = tensor<string, []>("op_382"), val = tensor<fp32, []>(0x1.99999ap-3)];
261
+ tensor<fp16, [1024, 128]> N_1_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_1_norm1_fc_weight_to_fp16"), val = tensor<fp16, [1024, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(13925632)))];
262
+ tensor<fp16, [1024]> N_1_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_1_norm1_fc_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14187840)))];
263
+ tensor<fp16, [1, 1024]> linear_8_cast_fp16 = linear(bias = N_1_norm1_fc_bias_to_fp16, weight = N_1_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_8_cast_fp16")];
264
+ tensor<int32, [3]> var_417 = const()[name = tensor<string, []>("op_417"), val = tensor<int32, [3]>([1, 1024, 1])];
265
+ tensor<fp16, [1, 1024, 1]> h_35_cast_fp16 = reshape(shape = var_417, x = linear_8_cast_fp16)[name = tensor<string, []>("h_35_cast_fp16")];
266
+ tensor<int32, [2]> var_419_split_sizes_0 = const()[name = tensor<string, []>("op_419_split_sizes_0"), val = tensor<int32, [2]>([512, 512])];
267
+ tensor<int32, []> var_419_axis_0 = const()[name = tensor<string, []>("op_419_axis_0"), val = tensor<int32, []>(1)];
268
+ tensor<fp16, [1, 512, 1]> var_419_cast_fp16_0, tensor<fp16, [1, 512, 1]> var_419_cast_fp16_1 = split(axis = var_419_axis_0, split_sizes = var_419_split_sizes_0, x = h_35_cast_fp16)[name = tensor<string, []>("op_419_cast_fp16")];
269
+ tensor<fp16, []> var_421_promoted_to_fp16 = const()[name = tensor<string, []>("op_421_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
270
+ tensor<fp16, [1, 512, 1]> var_422_cast_fp16 = add(x = var_419_cast_fp16_0, y = var_421_promoted_to_fp16)[name = tensor<string, []>("op_422_cast_fp16")];
271
+ tensor<fp16, []> var_386_to_fp16 = const()[name = tensor<string, []>("op_386_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
272
+ tensor<fp16, [1, 512, ?]> var_423_cast_fp16 = instance_norm(epsilon = var_386_to_fp16, x = _inversed_input_73_cast_fp16)[name = tensor<string, []>("op_423_cast_fp16")];
273
+ tensor<fp16, [1, 512, ?]> var_424_cast_fp16 = mul(x = var_422_cast_fp16, y = var_423_cast_fp16)[name = tensor<string, []>("op_424_cast_fp16")];
274
+ tensor<fp16, [1, 512, ?]> input_75_cast_fp16 = add(x = var_424_cast_fp16, y = var_419_cast_fp16_1)[name = tensor<string, []>("input_75_cast_fp16")];
275
+ tensor<fp16, [1, 512, ?]> input_77_cast_fp16 = leaky_relu(alpha = var_382, x = input_75_cast_fp16)[name = tensor<string, []>("input_77_cast_fp16")];
276
+ tensor<string, []> conv_transpose_1_pad_type_0 = const()[name = tensor<string, []>("conv_transpose_1_pad_type_0"), val = tensor<string, []>("custom")];
277
+ tensor<int32, [2]> conv_transpose_1_pad_0 = const()[name = tensor<string, []>("conv_transpose_1_pad_0"), val = tensor<int32, [2]>([0, 0])];
278
+ tensor<int32, [1]> conv_transpose_1_strides_0 = const()[name = tensor<string, []>("conv_transpose_1_strides_0"), val = tensor<int32, [1]>([2])];
279
+ tensor<int32, []> conv_transpose_1_groups_0 = const()[name = tensor<string, []>("conv_transpose_1_groups_0"), val = tensor<int32, []>(512)];
280
+ tensor<int32, [1]> conv_transpose_1_dilations_0 = const()[name = tensor<string, []>("conv_transpose_1_dilations_0"), val = tensor<int32, [1]>([1])];
281
+ tensor<fp16, [512, 1, 3]> var_427_to_fp16 = const()[name = tensor<string, []>("op_427_to_fp16"), val = tensor<fp16, [512, 1, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14189952)))];
282
+ tensor<fp16, [512]> N_1_pool_bias_to_fp16 = const()[name = tensor<string, []>("N_1_pool_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14193088)))];
283
+ tensor<fp16, [1, 512, ?]> conv_transpose_1_cast_fp16 = conv_transpose(bias = N_1_pool_bias_to_fp16, dilations = conv_transpose_1_dilations_0, groups = conv_transpose_1_groups_0, pad = conv_transpose_1_pad_0, pad_type = conv_transpose_1_pad_type_0, strides = conv_transpose_1_strides_0, weight = var_427_to_fp16, x = input_77_cast_fp16)[name = tensor<string, []>("conv_transpose_1_cast_fp16")];
284
+ tensor<int32, [3]> input_79_begin_0 = const()[name = tensor<string, []>("input_79_begin_0"), val = tensor<int32, [3]>([0, 0, 1])];
285
+ tensor<int32, [3]> input_79_end_0 = const()[name = tensor<string, []>("input_79_end_0"), val = tensor<int32, [3]>([0, 0, 0])];
286
+ tensor<bool, [3]> input_79_begin_mask_0 = const()[name = tensor<string, []>("input_79_begin_mask_0"), val = tensor<bool, [3]>([true, true, false])];
287
+ tensor<bool, [3]> input_79_end_mask_0 = const()[name = tensor<string, []>("input_79_end_mask_0"), val = tensor<bool, [3]>([true, true, true])];
288
+ tensor<fp16, [1, 512, ?]> input_79_cast_fp16 = slice_by_index(begin = input_79_begin_0, begin_mask = input_79_begin_mask_0, end = input_79_end_0, end_mask = input_79_end_mask_0, x = conv_transpose_1_cast_fp16)[name = tensor<string, []>("input_79_cast_fp16")];
289
+ tensor<string, []> input_83_pad_type_0 = const()[name = tensor<string, []>("input_83_pad_type_0"), val = tensor<string, []>("custom")];
290
+ tensor<int32, [2]> input_83_pad_0 = const()[name = tensor<string, []>("input_83_pad_0"), val = tensor<int32, [2]>([1, 1])];
291
+ tensor<int32, [1]> input_83_strides_0 = const()[name = tensor<string, []>("input_83_strides_0"), val = tensor<int32, [1]>([1])];
292
+ tensor<int32, [1]> input_83_dilations_0 = const()[name = tensor<string, []>("input_83_dilations_0"), val = tensor<int32, [1]>([1])];
293
+ tensor<int32, []> input_83_groups_0 = const()[name = tensor<string, []>("input_83_groups_0"), val = tensor<int32, []>(1)];
294
+ tensor<fp16, [256, 512, 3]> weight_21_to_fp16 = const()[name = tensor<string, []>("weight_21_to_fp16"), val = tensor<fp16, [256, 512, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14194176)))];
295
+ tensor<fp16, [256]> N_1_conv1_bias_to_fp16 = const()[name = tensor<string, []>("N_1_conv1_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14980672)))];
296
+ tensor<fp16, [1, 256, ?]> input_83_cast_fp16 = conv(bias = N_1_conv1_bias_to_fp16, dilations = input_83_dilations_0, groups = input_83_groups_0, pad = input_83_pad_0, pad_type = input_83_pad_type_0, strides = input_83_strides_0, weight = weight_21_to_fp16, x = input_79_cast_fp16)[name = tensor<string, []>("input_83_cast_fp16")];
297
+ tensor<fp16, [512, 128]> N_1_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_1_norm2_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14981248)))];
298
+ tensor<fp16, [512]> N_1_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_1_norm2_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15112384)))];
299
+ tensor<fp16, [1, 512]> linear_9_cast_fp16 = linear(bias = N_1_norm2_fc_bias_to_fp16, weight = N_1_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_9_cast_fp16")];
300
+ tensor<int32, [3]> var_448 = const()[name = tensor<string, []>("op_448"), val = tensor<int32, [3]>([1, 512, 1])];
301
+ tensor<fp16, [1, 512, 1]> h_39_cast_fp16 = reshape(shape = var_448, x = linear_9_cast_fp16)[name = tensor<string, []>("h_39_cast_fp16")];
302
+ tensor<int32, [2]> var_450_split_sizes_0 = const()[name = tensor<string, []>("op_450_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
303
+ tensor<int32, []> var_450_axis_0 = const()[name = tensor<string, []>("op_450_axis_0"), val = tensor<int32, []>(1)];
304
+ tensor<fp16, [1, 256, 1]> var_450_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_450_cast_fp16_1 = split(axis = var_450_axis_0, split_sizes = var_450_split_sizes_0, x = h_39_cast_fp16)[name = tensor<string, []>("op_450_cast_fp16")];
305
+ tensor<fp16, []> var_452_promoted_to_fp16 = const()[name = tensor<string, []>("op_452_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
306
+ tensor<fp16, [1, 256, 1]> var_453_cast_fp16 = add(x = var_450_cast_fp16_0, y = var_452_promoted_to_fp16)[name = tensor<string, []>("op_453_cast_fp16")];
307
+ tensor<fp16, [1, 256, ?]> var_454_cast_fp16 = instance_norm(epsilon = var_386_to_fp16, x = input_83_cast_fp16)[name = tensor<string, []>("op_454_cast_fp16")];
308
+ tensor<fp16, [1, 256, ?]> var_455_cast_fp16 = mul(x = var_453_cast_fp16, y = var_454_cast_fp16)[name = tensor<string, []>("op_455_cast_fp16")];
309
+ tensor<fp16, [1, 256, ?]> input_85_cast_fp16 = add(x = var_455_cast_fp16, y = var_450_cast_fp16_1)[name = tensor<string, []>("input_85_cast_fp16")];
310
+ tensor<fp16, [1, 256, ?]> input_87_cast_fp16 = leaky_relu(alpha = var_382, x = input_85_cast_fp16)[name = tensor<string, []>("input_87_cast_fp16")];
311
+ tensor<string, []> out_9_pad_type_0 = const()[name = tensor<string, []>("out_9_pad_type_0"), val = tensor<string, []>("custom")];
312
+ tensor<int32, [2]> out_9_pad_0 = const()[name = tensor<string, []>("out_9_pad_0"), val = tensor<int32, [2]>([1, 1])];
313
+ tensor<int32, [1]> out_9_strides_0 = const()[name = tensor<string, []>("out_9_strides_0"), val = tensor<int32, [1]>([1])];
314
+ tensor<int32, [1]> out_9_dilations_0 = const()[name = tensor<string, []>("out_9_dilations_0"), val = tensor<int32, [1]>([1])];
315
+ tensor<int32, []> out_9_groups_0 = const()[name = tensor<string, []>("out_9_groups_0"), val = tensor<int32, []>(1)];
316
+ tensor<fp16, [256, 256, 3]> weight_23_to_fp16 = const()[name = tensor<string, []>("weight_23_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15113472)))];
317
+ tensor<fp16, [256]> N_1_conv2_bias_to_fp16 = const()[name = tensor<string, []>("N_1_conv2_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15506752)))];
318
+ tensor<fp16, [1, 256, ?]> out_9_cast_fp16 = conv(bias = N_1_conv2_bias_to_fp16, dilations = out_9_dilations_0, groups = out_9_groups_0, pad = out_9_pad_0, pad_type = out_9_pad_type_0, strides = out_9_strides_0, weight = weight_23_to_fp16, x = input_87_cast_fp16)[name = tensor<string, []>("out_9_cast_fp16")];
319
+ tensor<int32, [1]> expand_dims_1_axes_0 = const()[name = tensor<string, []>("expand_dims_1_axes_0"), val = tensor<int32, [1]>([3])];
320
+ tensor<fp16, [1, 512, ?, 1]> expand_dims_1_cast_fp16 = expand_dims(axes = expand_dims_1_axes_0, x = _inversed_input_73_cast_fp16)[name = tensor<string, []>("expand_dims_1_cast_fp16")];
321
+ tensor<int32, []> upsample_nearest_neighbor_1_scale_factor_height_0 = const()[name = tensor<string, []>("upsample_nearest_neighbor_1_scale_factor_height_0"), val = tensor<int32, []>(2)];
322
+ tensor<int32, []> upsample_nearest_neighbor_1_scale_factor_width_0 = const()[name = tensor<string, []>("upsample_nearest_neighbor_1_scale_factor_width_0"), val = tensor<int32, []>(1)];
323
+ tensor<fp16, [1, 512, ?, 1]> upsample_nearest_neighbor_1_cast_fp16 = upsample_nearest_neighbor(scale_factor_height = upsample_nearest_neighbor_1_scale_factor_height_0, scale_factor_width = upsample_nearest_neighbor_1_scale_factor_width_0, x = expand_dims_1_cast_fp16)[name = tensor<string, []>("upsample_nearest_neighbor_1_cast_fp16")];
324
+ tensor<int32, [1]> input_91_axes_0 = const()[name = tensor<string, []>("input_91_axes_0"), val = tensor<int32, [1]>([3])];
325
+ tensor<fp16, [1, 512, ?]> input_91_cast_fp16 = squeeze(axes = input_91_axes_0, x = upsample_nearest_neighbor_1_cast_fp16)[name = tensor<string, []>("input_91_cast_fp16")];
326
+ tensor<string, []> var_473_pad_type_0 = const()[name = tensor<string, []>("op_473_pad_type_0"), val = tensor<string, []>("valid")];
327
+ tensor<int32, [1]> var_473_strides_0 = const()[name = tensor<string, []>("op_473_strides_0"), val = tensor<int32, [1]>([1])];
328
+ tensor<int32, [2]> var_473_pad_0 = const()[name = tensor<string, []>("op_473_pad_0"), val = tensor<int32, [2]>([0, 0])];
329
+ tensor<int32, [1]> var_473_dilations_0 = const()[name = tensor<string, []>("op_473_dilations_0"), val = tensor<int32, [1]>([1])];
330
+ tensor<int32, []> var_473_groups_0 = const()[name = tensor<string, []>("op_473_groups_0"), val = tensor<int32, []>(1)];
331
+ tensor<fp16, [256, 512, 1]> weight_25_to_fp16 = const()[name = tensor<string, []>("weight_25_to_fp16"), val = tensor<fp16, [256, 512, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15507328)))];
332
+ tensor<fp16, [1, 256, ?]> var_473_cast_fp16 = conv(dilations = var_473_dilations_0, groups = var_473_groups_0, pad = var_473_pad_0, pad_type = var_473_pad_type_0, strides = var_473_strides_0, weight = weight_25_to_fp16, x = input_91_cast_fp16)[name = tensor<string, []>("op_473_cast_fp16")];
333
+ tensor<fp16, [1, 256, ?]> var_474_cast_fp16 = add(x = out_9_cast_fp16, y = var_473_cast_fp16)[name = tensor<string, []>("op_474_cast_fp16")];
334
+ tensor<fp16, []> _inversed_input_93_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_93_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
335
+ tensor<fp16, [1, 256, ?]> _inversed_input_93_cast_fp16 = mul(x = var_474_cast_fp16, y = _inversed_input_93_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_93_cast_fp16")];
336
+ tensor<fp32, []> var_478 = const()[name = tensor<string, []>("op_478"), val = tensor<fp32, []>(0x1.99999ap-3)];
337
+ tensor<fp16, [512, 128]> N_2_norm1_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_2_norm1_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15769536)))];
338
+ tensor<fp16, [512]> N_2_norm1_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_2_norm1_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15900672)))];
339
+ tensor<fp16, [1, 512]> linear_10_cast_fp16 = linear(bias = N_2_norm1_fc_bias_to_fp16, weight = N_2_norm1_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_10_cast_fp16")];
340
+ tensor<int32, [3]> var_503 = const()[name = tensor<string, []>("op_503"), val = tensor<int32, [3]>([1, 512, 1])];
341
+ tensor<fp16, [1, 512, 1]> h_43_cast_fp16 = reshape(shape = var_503, x = linear_10_cast_fp16)[name = tensor<string, []>("h_43_cast_fp16")];
342
+ tensor<int32, [2]> var_505_split_sizes_0 = const()[name = tensor<string, []>("op_505_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
343
+ tensor<int32, []> var_505_axis_0 = const()[name = tensor<string, []>("op_505_axis_0"), val = tensor<int32, []>(1)];
344
+ tensor<fp16, [1, 256, 1]> var_505_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_505_cast_fp16_1 = split(axis = var_505_axis_0, split_sizes = var_505_split_sizes_0, x = h_43_cast_fp16)[name = tensor<string, []>("op_505_cast_fp16")];
345
+ tensor<fp16, []> var_507_promoted_to_fp16 = const()[name = tensor<string, []>("op_507_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
346
+ tensor<fp16, [1, 256, 1]> var_508_cast_fp16 = add(x = var_505_cast_fp16_0, y = var_507_promoted_to_fp16)[name = tensor<string, []>("op_508_cast_fp16")];
347
+ tensor<fp16, []> var_481_to_fp16 = const()[name = tensor<string, []>("op_481_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
348
+ tensor<fp16, [1, 256, ?]> var_509_cast_fp16 = instance_norm(epsilon = var_481_to_fp16, x = _inversed_input_93_cast_fp16)[name = tensor<string, []>("op_509_cast_fp16")];
349
+ tensor<fp16, [1, 256, ?]> var_510_cast_fp16 = mul(x = var_508_cast_fp16, y = var_509_cast_fp16)[name = tensor<string, []>("op_510_cast_fp16")];
350
+ tensor<fp16, [1, 256, ?]> input_95_cast_fp16 = add(x = var_510_cast_fp16, y = var_505_cast_fp16_1)[name = tensor<string, []>("input_95_cast_fp16")];
351
+ tensor<fp16, [1, 256, ?]> input_97_cast_fp16 = leaky_relu(alpha = var_478, x = input_95_cast_fp16)[name = tensor<string, []>("input_97_cast_fp16")];
352
+ tensor<string, []> input_101_pad_type_0 = const()[name = tensor<string, []>("input_101_pad_type_0"), val = tensor<string, []>("custom")];
353
+ tensor<int32, [2]> input_101_pad_0 = const()[name = tensor<string, []>("input_101_pad_0"), val = tensor<int32, [2]>([1, 1])];
354
+ tensor<int32, [1]> input_101_strides_0 = const()[name = tensor<string, []>("input_101_strides_0"), val = tensor<int32, [1]>([1])];
355
+ tensor<int32, [1]> input_101_dilations_0 = const()[name = tensor<string, []>("input_101_dilations_0"), val = tensor<int32, [1]>([1])];
356
+ tensor<int32, []> input_101_groups_0 = const()[name = tensor<string, []>("input_101_groups_0"), val = tensor<int32, []>(1)];
357
+ tensor<fp16, [256, 256, 3]> weight_27_to_fp16 = const()[name = tensor<string, []>("weight_27_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15901760)))];
358
+ tensor<fp16, [256]> N_2_conv1_bias_to_fp16 = const()[name = tensor<string, []>("N_2_conv1_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16295040)))];
359
+ tensor<fp16, [1, 256, ?]> input_101_cast_fp16 = conv(bias = N_2_conv1_bias_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = weight_27_to_fp16, x = input_97_cast_fp16)[name = tensor<string, []>("input_101_cast_fp16")];
360
+ tensor<fp16, [512, 128]> N_2_norm2_fc_weight_to_fp16 = const()[name = tensor<string, []>("N_2_norm2_fc_weight_to_fp16"), val = tensor<fp16, [512, 128]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16295616)))];
361
+ tensor<fp16, [512]> N_2_norm2_fc_bias_to_fp16 = const()[name = tensor<string, []>("N_2_norm2_fc_bias_to_fp16"), val = tensor<fp16, [512]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16426752)))];
362
+ tensor<fp16, [1, 512]> linear_11_cast_fp16 = linear(bias = N_2_norm2_fc_bias_to_fp16, weight = N_2_norm2_fc_weight_to_fp16, x = s_to_fp16)[name = tensor<string, []>("linear_11_cast_fp16")];
363
+ tensor<int32, [3]> var_527 = const()[name = tensor<string, []>("op_527"), val = tensor<int32, [3]>([1, 512, 1])];
364
+ tensor<fp16, [1, 512, 1]> h_cast_fp16 = reshape(shape = var_527, x = linear_11_cast_fp16)[name = tensor<string, []>("h_cast_fp16")];
365
+ tensor<int32, [2]> var_529_split_sizes_0 = const()[name = tensor<string, []>("op_529_split_sizes_0"), val = tensor<int32, [2]>([256, 256])];
366
+ tensor<int32, []> var_529_axis_0 = const()[name = tensor<string, []>("op_529_axis_0"), val = tensor<int32, []>(1)];
367
+ tensor<fp16, [1, 256, 1]> var_529_cast_fp16_0, tensor<fp16, [1, 256, 1]> var_529_cast_fp16_1 = split(axis = var_529_axis_0, split_sizes = var_529_split_sizes_0, x = h_cast_fp16)[name = tensor<string, []>("op_529_cast_fp16")];
368
+ tensor<fp16, []> var_531_promoted_to_fp16 = const()[name = tensor<string, []>("op_531_promoted_to_fp16"), val = tensor<fp16, []>(0x1p+0)];
369
+ tensor<fp16, [1, 256, 1]> var_532_cast_fp16 = add(x = var_529_cast_fp16_0, y = var_531_promoted_to_fp16)[name = tensor<string, []>("op_532_cast_fp16")];
370
+ tensor<fp16, [1, 256, ?]> var_533_cast_fp16 = instance_norm(epsilon = var_481_to_fp16, x = input_101_cast_fp16)[name = tensor<string, []>("op_533_cast_fp16")];
371
+ tensor<fp16, [1, 256, ?]> var_534_cast_fp16 = mul(x = var_532_cast_fp16, y = var_533_cast_fp16)[name = tensor<string, []>("op_534_cast_fp16")];
372
+ tensor<fp16, [1, 256, ?]> input_103_cast_fp16 = add(x = var_534_cast_fp16, y = var_529_cast_fp16_1)[name = tensor<string, []>("input_103_cast_fp16")];
373
+ tensor<fp16, [1, 256, ?]> input_105_cast_fp16 = leaky_relu(alpha = var_478, x = input_103_cast_fp16)[name = tensor<string, []>("input_105_cast_fp16")];
374
+ tensor<string, []> out_pad_type_0 = const()[name = tensor<string, []>("out_pad_type_0"), val = tensor<string, []>("custom")];
375
+ tensor<int32, [2]> out_pad_0 = const()[name = tensor<string, []>("out_pad_0"), val = tensor<int32, [2]>([1, 1])];
376
+ tensor<int32, [1]> out_strides_0 = const()[name = tensor<string, []>("out_strides_0"), val = tensor<int32, [1]>([1])];
377
+ tensor<int32, [1]> out_dilations_0 = const()[name = tensor<string, []>("out_dilations_0"), val = tensor<int32, [1]>([1])];
378
+ tensor<int32, []> out_groups_0 = const()[name = tensor<string, []>("out_groups_0"), val = tensor<int32, []>(1)];
379
+ tensor<fp16, [256, 256, 3]> weight_29_to_fp16 = const()[name = tensor<string, []>("weight_29_to_fp16"), val = tensor<fp16, [256, 256, 3]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16427840)))];
380
+ tensor<fp16, [256]> N_2_conv2_bias_to_fp16 = const()[name = tensor<string, []>("N_2_conv2_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16821120)))];
381
+ tensor<fp16, [1, 256, ?]> out_cast_fp16 = conv(bias = N_2_conv2_bias_to_fp16, dilations = out_dilations_0, groups = out_groups_0, pad = out_pad_0, pad_type = out_pad_type_0, strides = out_strides_0, weight = weight_29_to_fp16, x = input_105_cast_fp16)[name = tensor<string, []>("out_cast_fp16")];
382
+ tensor<fp16, [1, 256, ?]> var_545_cast_fp16 = add(x = out_cast_fp16, y = _inversed_input_93_cast_fp16)[name = tensor<string, []>("op_545_cast_fp16")];
383
+ tensor<fp16, []> _inversed_input_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_input_y_0_to_fp16"), val = tensor<fp16, []>(0x1.6ap-1)];
384
+ tensor<fp16, [1, 256, ?]> _inversed_input_cast_fp16 = mul(x = var_545_cast_fp16, y = _inversed_input_y_0_to_fp16)[name = tensor<string, []>("_inversed_input_cast_fp16")];
385
+ tensor<string, []> var_558_pad_type_0 = const()[name = tensor<string, []>("op_558_pad_type_0"), val = tensor<string, []>("valid")];
386
+ tensor<int32, [1]> var_558_strides_0 = const()[name = tensor<string, []>("op_558_strides_0"), val = tensor<int32, [1]>([1])];
387
+ tensor<int32, [2]> var_558_pad_0 = const()[name = tensor<string, []>("op_558_pad_0"), val = tensor<int32, [2]>([0, 0])];
388
+ tensor<int32, [1]> var_558_dilations_0 = const()[name = tensor<string, []>("op_558_dilations_0"), val = tensor<int32, [1]>([1])];
389
+ tensor<int32, []> var_558_groups_0 = const()[name = tensor<string, []>("op_558_groups_0"), val = tensor<int32, []>(1)];
390
+ tensor<fp16, [1, 256, 1]> N_proj_weight_to_fp16 = const()[name = tensor<string, []>("N_proj_weight_to_fp16"), val = tensor<fp16, [1, 256, 1]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16821696)))];
391
+ tensor<fp16, [1]> N_proj_bias_to_fp16 = const()[name = tensor<string, []>("N_proj_bias_to_fp16"), val = tensor<fp16, [1]>([0x1.6ep-4])];
392
+ tensor<fp16, [1, 1, ?]> var_558_cast_fp16 = conv(bias = N_proj_bias_to_fp16, dilations = var_558_dilations_0, groups = var_558_groups_0, pad = var_558_pad_0, pad_type = var_558_pad_type_0, strides = var_558_strides_0, weight = N_proj_weight_to_fp16, x = _inversed_input_cast_fp16)[name = tensor<string, []>("op_558_cast_fp16")];
393
+ tensor<int32, [1]> var_560_axes_0 = const()[name = tensor<string, []>("op_560_axes_0"), val = tensor<int32, [1]>([1])];
394
+ tensor<fp16, [1, ?]> N = squeeze(axes = var_560_axes_0, x = var_558_cast_fp16)[name = tensor<string, []>("op_560_cast_fp16")];
395
+ } -> (F0, N);
396
+ }
compiled/styletts2_f0n_energy.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b21f183d2ff876842ea2df14cdc033c8935a1805382b70b241c4f5a1bf32b3a8
3
+ size 16822272
compiled/styletts2_text_predictor_128.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e33f8046365fa207d90593ce420673cd6921bb7156c5d3340cd5c6290990e4f9
3
+ size 243
compiled/styletts2_text_predictor_128.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5b8da13d13caee8b6e7fa3bf35b2d98fe29a782deded2477aea61d92d8c283
3
+ size 528
compiled/styletts2_text_predictor_128.mlmodelc/metadata.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 text+predictor (LibriTTS HiFi-GAN) @ T_tok=128.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 512 × 128)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 512, 128]",
13
+ "name" : "t_en",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 512 × 128)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 512, 128]",
23
+ "name" : "d_en",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 128 × 640)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 128, 640]",
33
+ "name" : "d",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 128 × 50)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 128, 50]",
43
+ "name" : "pred_dur_log",
44
+ "type" : "MultiArray"
45
+ },
46
+ {
47
+ "hasShapeFlexibility" : "0",
48
+ "isOptional" : "0",
49
+ "dataType" : "Float16",
50
+ "formattedType" : "MultiArray (Float16 1 × 128 × 768)",
51
+ "shortDescription" : "",
52
+ "shape" : "[1, 128, 768]",
53
+ "name" : "fixed_embedding",
54
+ "type" : "MultiArray"
55
+ },
56
+ {
57
+ "hasShapeFlexibility" : "0",
58
+ "isOptional" : "0",
59
+ "dataType" : "Float16",
60
+ "formattedType" : "MultiArray (Float16 1 × 128 × 768)",
61
+ "shortDescription" : "",
62
+ "shape" : "[1, 128, 768]",
63
+ "name" : "bert_dur",
64
+ "type" : "MultiArray"
65
+ }
66
+ ],
67
+ "storagePrecision" : "Float16",
68
+ "modelParameters" : [
69
+
70
+ ],
71
+ "specificationVersion" : 8,
72
+ "mlProgramOperationTypeHistogram" : {
73
+ "Split" : 3,
74
+ "Ios17.mul" : 15,
75
+ "Ios17.linear" : 78,
76
+ "Ios17.transpose" : 76,
77
+ "Ios17.matmul" : 24,
78
+ "Ios17.conv" : 3,
79
+ "Ios17.leakyRelu" : 3,
80
+ "Ios17.gather" : 2,
81
+ "Ios17.add" : 32,
82
+ "Ios17.lstm" : 5,
83
+ "Ios17.concat" : 4,
84
+ "Ios16.softmax" : 12,
85
+ "Ios16.gelu" : 12,
86
+ "Ios17.expandDims" : 1,
87
+ "Tile" : 1,
88
+ "Ios17.cast" : 3,
89
+ "Ios17.layerNorm" : 31,
90
+ "Ios17.reshape" : 51
91
+ },
92
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
93
+ "isUpdatable" : "0",
94
+ "stateSchema" : [
95
+
96
+ ],
97
+ "availability" : {
98
+ "macOS" : "14.0",
99
+ "tvOS" : "17.0",
100
+ "visionOS" : "1.0",
101
+ "watchOS" : "10.0",
102
+ "iOS" : "17.0",
103
+ "macCatalyst" : "17.0"
104
+ },
105
+ "modelType" : {
106
+ "name" : "MLModelType_mlProgram"
107
+ },
108
+ "inputSchema" : [
109
+ {
110
+ "hasShapeFlexibility" : "0",
111
+ "isOptional" : "0",
112
+ "dataType" : "Int32",
113
+ "formattedType" : "MultiArray (Int32 1 × 128)",
114
+ "shortDescription" : "",
115
+ "shape" : "[1, 128]",
116
+ "name" : "tokens",
117
+ "type" : "MultiArray"
118
+ },
119
+ {
120
+ "hasShapeFlexibility" : "0",
121
+ "isOptional" : "0",
122
+ "dataType" : "Float32",
123
+ "formattedType" : "MultiArray (Float32 1 × 128)",
124
+ "shortDescription" : "",
125
+ "shape" : "[1, 128]",
126
+ "name" : "style",
127
+ "type" : "MultiArray"
128
+ }
129
+ ],
130
+ "userDefinedMetadata" : {
131
+ "com.github.apple.coremltools.source_dialect" : "TorchScript",
132
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
133
+ "com.github.apple.coremltools.version" : "8.3.0"
134
+ },
135
+ "generatedClassName" : "styletts2_text_predictor_128",
136
+ "method" : "predict"
137
+ }
138
+ ]
compiled/styletts2_text_predictor_128.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_text_predictor_128.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ce18107b3f5fe1e518342354f55dd071f42a30e18f4512029fd9b3635105ce
3
+ size 39073856
compiled/styletts2_text_predictor_256.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2a3e12ca039dc9909255950043c9bd39018dea7a926c0e6f9c596043b7c75d6
3
+ size 243
compiled/styletts2_text_predictor_256.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:990616fa15d673099e240788d898fed9aea4f27dfebe3e34a4ea44a5f61416da
3
+ size 528
compiled/styletts2_text_predictor_256.mlmodelc/metadata.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 text+predictor (LibriTTS HiFi-GAN) @ T_tok=256.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 512 × 256)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 512, 256]",
13
+ "name" : "t_en",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 512 × 256)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 512, 256]",
23
+ "name" : "d_en",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 256 × 640)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 256, 640]",
33
+ "name" : "d",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 256 × 50)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 256, 50]",
43
+ "name" : "pred_dur_log",
44
+ "type" : "MultiArray"
45
+ },
46
+ {
47
+ "hasShapeFlexibility" : "0",
48
+ "isOptional" : "0",
49
+ "dataType" : "Float16",
50
+ "formattedType" : "MultiArray (Float16 1 × 256 × 768)",
51
+ "shortDescription" : "",
52
+ "shape" : "[1, 256, 768]",
53
+ "name" : "fixed_embedding",
54
+ "type" : "MultiArray"
55
+ },
56
+ {
57
+ "hasShapeFlexibility" : "0",
58
+ "isOptional" : "0",
59
+ "dataType" : "Float16",
60
+ "formattedType" : "MultiArray (Float16 1 × 256 × 768)",
61
+ "shortDescription" : "",
62
+ "shape" : "[1, 256, 768]",
63
+ "name" : "bert_dur",
64
+ "type" : "MultiArray"
65
+ }
66
+ ],
67
+ "storagePrecision" : "Float16",
68
+ "modelParameters" : [
69
+
70
+ ],
71
+ "specificationVersion" : 8,
72
+ "mlProgramOperationTypeHistogram" : {
73
+ "Split" : 3,
74
+ "Ios17.mul" : 15,
75
+ "Ios17.linear" : 78,
76
+ "Ios17.transpose" : 76,
77
+ "Ios17.matmul" : 24,
78
+ "Ios17.conv" : 3,
79
+ "Ios17.leakyRelu" : 3,
80
+ "Ios17.gather" : 2,
81
+ "Ios17.add" : 32,
82
+ "Ios17.lstm" : 5,
83
+ "Ios17.concat" : 4,
84
+ "Ios16.softmax" : 12,
85
+ "Ios16.gelu" : 12,
86
+ "Ios17.expandDims" : 1,
87
+ "Tile" : 1,
88
+ "Ios17.cast" : 3,
89
+ "Ios17.layerNorm" : 31,
90
+ "Ios17.reshape" : 51
91
+ },
92
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
93
+ "isUpdatable" : "0",
94
+ "stateSchema" : [
95
+
96
+ ],
97
+ "availability" : {
98
+ "macOS" : "14.0",
99
+ "tvOS" : "17.0",
100
+ "visionOS" : "1.0",
101
+ "watchOS" : "10.0",
102
+ "iOS" : "17.0",
103
+ "macCatalyst" : "17.0"
104
+ },
105
+ "modelType" : {
106
+ "name" : "MLModelType_mlProgram"
107
+ },
108
+ "inputSchema" : [
109
+ {
110
+ "hasShapeFlexibility" : "0",
111
+ "isOptional" : "0",
112
+ "dataType" : "Int32",
113
+ "formattedType" : "MultiArray (Int32 1 × 256)",
114
+ "shortDescription" : "",
115
+ "shape" : "[1, 256]",
116
+ "name" : "tokens",
117
+ "type" : "MultiArray"
118
+ },
119
+ {
120
+ "hasShapeFlexibility" : "0",
121
+ "isOptional" : "0",
122
+ "dataType" : "Float32",
123
+ "formattedType" : "MultiArray (Float32 1 × 128)",
124
+ "shortDescription" : "",
125
+ "shape" : "[1, 128]",
126
+ "name" : "style",
127
+ "type" : "MultiArray"
128
+ }
129
+ ],
130
+ "userDefinedMetadata" : {
131
+ "com.github.apple.coremltools.version" : "8.3.0",
132
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
133
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
134
+ },
135
+ "generatedClassName" : "styletts2_text_predictor_256",
136
+ "method" : "predict"
137
+ }
138
+ ]
compiled/styletts2_text_predictor_256.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
compiled/styletts2_text_predictor_256.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4221120586ac22d722b35894d94b7a370d35721fd5ffc5ca93bca4c0da90a868
3
+ size 39336000
compiled/styletts2_text_predictor_32.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8c53623b57a193d6fd43afe30a1126d4fb5d8685d668cb7e5545766ee39f80
3
+ size 243
compiled/styletts2_text_predictor_32.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452c32fad506218658305781fea9f2b734a00c71dae78f0c4bd5ff46a1a68230
3
+ size 520
compiled/styletts2_text_predictor_32.mlmodelc/metadata.json ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "shortDescription" : "StyleTTS2 text+predictor (LibriTTS HiFi-GAN) @ T_tok=32.",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 512 × 32)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 512, 32]",
13
+ "name" : "t_en",
14
+ "type" : "MultiArray"
15
+ },
16
+ {
17
+ "hasShapeFlexibility" : "0",
18
+ "isOptional" : "0",
19
+ "dataType" : "Float16",
20
+ "formattedType" : "MultiArray (Float16 1 × 512 × 32)",
21
+ "shortDescription" : "",
22
+ "shape" : "[1, 512, 32]",
23
+ "name" : "d_en",
24
+ "type" : "MultiArray"
25
+ },
26
+ {
27
+ "hasShapeFlexibility" : "0",
28
+ "isOptional" : "0",
29
+ "dataType" : "Float16",
30
+ "formattedType" : "MultiArray (Float16 1 × 32 × 640)",
31
+ "shortDescription" : "",
32
+ "shape" : "[1, 32, 640]",
33
+ "name" : "d",
34
+ "type" : "MultiArray"
35
+ },
36
+ {
37
+ "hasShapeFlexibility" : "0",
38
+ "isOptional" : "0",
39
+ "dataType" : "Float16",
40
+ "formattedType" : "MultiArray (Float16 1 × 32 × 50)",
41
+ "shortDescription" : "",
42
+ "shape" : "[1, 32, 50]",
43
+ "name" : "pred_dur_log",
44
+ "type" : "MultiArray"
45
+ },
46
+ {
47
+ "hasShapeFlexibility" : "0",
48
+ "isOptional" : "0",
49
+ "dataType" : "Float16",
50
+ "formattedType" : "MultiArray (Float16 1 × 32 × 768)",
51
+ "shortDescription" : "",
52
+ "shape" : "[1, 32, 768]",
53
+ "name" : "fixed_embedding",
54
+ "type" : "MultiArray"
55
+ },
56
+ {
57
+ "hasShapeFlexibility" : "0",
58
+ "isOptional" : "0",
59
+ "dataType" : "Float16",
60
+ "formattedType" : "MultiArray (Float16 1 × 32 × 768)",
61
+ "shortDescription" : "",
62
+ "shape" : "[1, 32, 768]",
63
+ "name" : "bert_dur",
64
+ "type" : "MultiArray"
65
+ }
66
+ ],
67
+ "storagePrecision" : "Float16",
68
+ "modelParameters" : [
69
+
70
+ ],
71
+ "specificationVersion" : 8,
72
+ "mlProgramOperationTypeHistogram" : {
73
+ "Split" : 3,
74
+ "Ios17.mul" : 15,
75
+ "Ios17.linear" : 78,
76
+ "Ios17.transpose" : 76,
77
+ "Ios17.matmul" : 24,
78
+ "Ios17.conv" : 3,
79
+ "Ios17.leakyRelu" : 3,
80
+ "Ios17.gather" : 2,
81
+ "Ios17.add" : 32,
82
+ "Ios17.lstm" : 5,
83
+ "Ios17.concat" : 4,
84
+ "Ios16.softmax" : 12,
85
+ "Ios16.gelu" : 12,
86
+ "Ios17.expandDims" : 1,
87
+ "Tile" : 1,
88
+ "Ios17.cast" : 3,
89
+ "Ios17.layerNorm" : 31,
90
+ "Ios17.reshape" : 51
91
+ },
92
+ "computePrecision" : "Mixed (Float16, Float32, Int16, Int32, UInt16)",
93
+ "isUpdatable" : "0",
94
+ "stateSchema" : [
95
+
96
+ ],
97
+ "availability" : {
98
+ "macOS" : "14.0",
99
+ "tvOS" : "17.0",
100
+ "visionOS" : "1.0",
101
+ "watchOS" : "10.0",
102
+ "iOS" : "17.0",
103
+ "macCatalyst" : "17.0"
104
+ },
105
+ "modelType" : {
106
+ "name" : "MLModelType_mlProgram"
107
+ },
108
+ "inputSchema" : [
109
+ {
110
+ "hasShapeFlexibility" : "0",
111
+ "isOptional" : "0",
112
+ "dataType" : "Int32",
113
+ "formattedType" : "MultiArray (Int32 1 × 32)",
114
+ "shortDescription" : "",
115
+ "shape" : "[1, 32]",
116
+ "name" : "tokens",
117
+ "type" : "MultiArray"
118
+ },
119
+ {
120
+ "hasShapeFlexibility" : "0",
121
+ "isOptional" : "0",
122
+ "dataType" : "Float32",
123
+ "formattedType" : "MultiArray (Float32 1 × 128)",
124
+ "shortDescription" : "",
125
+ "shape" : "[1, 128]",
126
+ "name" : "style",
127
+ "type" : "MultiArray"
128
+ }
129
+ ],
130
+ "userDefinedMetadata" : {
131
+ "com.github.apple.coremltools.version" : "8.3.0",
132
+ "com.github.apple.coremltools.source" : "torch==2.11.0",
133
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
134
+ },
135
+ "generatedClassName" : "styletts2_text_predictor_32",
136
+ "method" : "predict"
137
+ }
138
+ ]