diff --git "a/SplitEinsum-Resources-6bit/TextEncoder2.mlmodelc/model.mil" "b/SplitEinsum-Resources-6bit/TextEncoder2.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/SplitEinsum-Resources-6bit/TextEncoder2.mlmodelc/model.mil" @@ -0,0 +1,2211 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3304.5.2"}, {"coremlc-version", "3304.6.2"}})] +{ + func main(tensor input_ids) { + tensor var_5 = const()[name = tensor("op_5"), val = tensor(-1)]; + tensor var_6 = const()[name = tensor("op_6"), val = tensor(false)]; + tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; + tensor inputs_embeds_axis_0 = const()[name = tensor("inputs_embeds_axis_0"), val = tensor(0)]; + tensor inputs_embeds_batch_dims_0 = const()[name = tensor("inputs_embeds_batch_dims_0"), val = tensor(0)]; + tensor text_encoder_text_model_embeddings_token_embedding_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_embeddings_token_embedding_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor cast_1 = cast(dtype = cast_1_dtype_0, x = input_ids)[name = tensor("cast_2")]; + tensor inputs_embeds_cast_fp16 = gather(axis = inputs_embeds_axis_0, batch_dims = inputs_embeds_batch_dims_0, indices = cast_1, x = text_encoder_text_model_embeddings_token_embedding_weight_to_fp16)[name = tensor("inputs_embeds_cast_fp16")]; + tensor position_embeddings_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126484608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126558592))), name = tensor("position_embeddings_to_fp16_palettized"), shape = tensor([1, 77, 1280])]; + tensor input_3_cast_fp16 = add(x = inputs_embeds_cast_fp16, y = position_embeddings_to_fp16_palettized)[name = tensor("input_3_cast_fp16")]; + tensor hidden_states_1_axes_0 = const()[name = tensor("hidden_states_1_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126558784)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126561408)))]; + tensor var_15_to_fp16 = const()[name = tensor("op_15_to_fp16"), val = tensor(0x1.5p-17)]; + tensor hidden_states_1_cast_fp16 = layer_norm(axes = hidden_states_1_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("hidden_states_1_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126564032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127792896))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127793088)))]; + tensor linear_0_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_149_to_fp16 = const()[name = tensor("op_149_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_5_cast_fp16 = mul(x = linear_0_cast_fp16, y = var_149_to_fp16)[name = tensor("tensor_5_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127795712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129024576))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129024768)))]; + tensor linear_1_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_154 = const()[name = tensor("op_154"), val = tensor([1, -1, 20, 64])]; + tensor var_155_cast_fp16 = reshape(shape = var_154, x = linear_1_cast_fp16)[name = tensor("op_155_cast_fp16")]; + tensor var_156_perm_0 = const()[name = tensor("op_156_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129027392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130256256))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130256448)))]; + tensor linear_2_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_161 = const()[name = tensor("op_161"), val = tensor([1, -1, 20, 64])]; + tensor var_162_cast_fp16 = reshape(shape = var_161, x = linear_2_cast_fp16)[name = tensor("op_162_cast_fp16")]; + tensor var_163_perm_0 = const()[name = tensor("op_163_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_170 = const()[name = tensor("op_170"), val = tensor([1, 77, 20, 64])]; + tensor var_171_cast_fp16 = reshape(shape = var_170, x = tensor_5_cast_fp16)[name = tensor("op_171_cast_fp16")]; + tensor var_172_perm_0 = const()[name = tensor("op_172_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_174 = const()[name = tensor("op_174"), val = tensor([20, -1, 64])]; + tensor var_172_cast_fp16 = transpose(perm = var_172_perm_0, x = var_171_cast_fp16)[name = tensor("transpose_127")]; + tensor query_states_1_cast_fp16 = reshape(shape = var_174, x = var_172_cast_fp16)[name = tensor("query_states_1_cast_fp16")]; + tensor var_176 = const()[name = tensor("op_176"), val = tensor([20, -1, 64])]; + tensor var_156_cast_fp16 = transpose(perm = var_156_perm_0, x = var_155_cast_fp16)[name = tensor("transpose_126")]; + tensor key_states_3_cast_fp16 = reshape(shape = var_176, x = var_156_cast_fp16)[name = tensor("key_states_3_cast_fp16")]; + tensor var_178 = const()[name = tensor("op_178"), val = tensor([20, -1, 64])]; + tensor var_163_cast_fp16 = transpose(perm = var_163_perm_0, x = var_162_cast_fp16)[name = tensor("transpose_125")]; + tensor value_states_3_cast_fp16 = reshape(shape = var_178, x = var_163_cast_fp16)[name = tensor("value_states_3_cast_fp16")]; + tensor attn_weights_1_transpose_x_1 = const()[name = tensor("attn_weights_1_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_1_transpose_y_1 = const()[name = tensor("attn_weights_1_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_1_cast_fp16 = matmul(transpose_x = attn_weights_1_transpose_x_1, transpose_y = attn_weights_1_transpose_y_1, x = query_states_1_cast_fp16, y = key_states_3_cast_fp16)[name = tensor("attn_weights_1_cast_fp16")]; + tensor var_183 = const()[name = tensor("op_183"), val = tensor([1, 20, 77, 77])]; + tensor var_184_cast_fp16 = reshape(shape = var_183, x = attn_weights_1_cast_fp16)[name = tensor("op_184_cast_fp16")]; + tensor op_58_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130259072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130263616))), name = tensor("op_58_to_fp16_palettized"), shape = tensor([1, 1, 77, 77])]; + tensor attn_weights_3_cast_fp16 = add(x = var_184_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_3_cast_fp16")]; + tensor var_189 = const()[name = tensor("op_189"), val = tensor([20, 77, 77])]; + tensor input_5_cast_fp16 = reshape(shape = var_189, x = attn_weights_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = softmax(axis = var_5, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor attn_output_1_transpose_x_0 = const()[name = tensor("attn_output_1_transpose_x_0"), val = tensor(false)]; + tensor attn_output_1_transpose_y_0 = const()[name = tensor("attn_output_1_transpose_y_0"), val = tensor(false)]; + tensor attn_output_1_cast_fp16 = matmul(transpose_x = attn_output_1_transpose_x_0, transpose_y = attn_output_1_transpose_y_0, x = input_7_cast_fp16, y = value_states_3_cast_fp16)[name = tensor("attn_output_1_cast_fp16")]; + tensor var_194 = const()[name = tensor("op_194"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_3_cast_fp16 = reshape(shape = var_194, x = attn_output_1_cast_fp16)[name = tensor("attn_output_3_cast_fp16")]; + tensor attn_output_5_perm_0 = const()[name = tensor("attn_output_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_197 = const()[name = tensor("op_197"), val = tensor([1, 77, 1280])]; + tensor attn_output_5_cast_fp16 = transpose(perm = attn_output_5_perm_0, x = attn_output_3_cast_fp16)[name = tensor("transpose_124")]; + tensor input_9_cast_fp16 = reshape(shape = var_197, x = attn_output_5_cast_fp16)[name = tensor("input_9_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130263808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131492672))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131492864)))]; + tensor linear_3_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized, x = input_9_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor input_11_cast_fp16 = add(x = input_3_cast_fp16, y = linear_3_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_axes_0 = const()[name = tensor("input_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131495488)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131498112)))]; + tensor input_13_cast_fp16 = layer_norm(axes = input_13_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131500736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136416000))), name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136416192)))]; + tensor linear_4_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized, x = input_13_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("EXACT")]; + tensor input_17_cast_fp16 = gelu(mode = input_17_mode_0, x = linear_4_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136426496))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141341760))), name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141341952)))]; + tensor linear_5_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized, x = input_17_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor input_19_cast_fp16 = add(x = input_11_cast_fp16, y = linear_5_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor hidden_states_7_axes_0 = const()[name = tensor("hidden_states_7_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141344576)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141347200)))]; + tensor hidden_states_7_cast_fp16 = layer_norm(axes = hidden_states_7_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("hidden_states_7_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141349824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142578688))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142578880)))]; + tensor linear_6_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_6_cast_fp16")]; + tensor var_236_to_fp16 = const()[name = tensor("op_236_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_11_cast_fp16 = mul(x = linear_6_cast_fp16, y = var_236_to_fp16)[name = tensor("tensor_11_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142581504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143810368))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143810560)))]; + tensor linear_7_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor var_241 = const()[name = tensor("op_241"), val = tensor([1, -1, 20, 64])]; + tensor var_242_cast_fp16 = reshape(shape = var_241, x = linear_7_cast_fp16)[name = tensor("op_242_cast_fp16")]; + tensor var_243_perm_0 = const()[name = tensor("op_243_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143813184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145042048))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145042240)))]; + tensor linear_8_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor var_248 = const()[name = tensor("op_248"), val = tensor([1, -1, 20, 64])]; + tensor var_249_cast_fp16 = reshape(shape = var_248, x = linear_8_cast_fp16)[name = tensor("op_249_cast_fp16")]; + tensor var_250_perm_0 = const()[name = tensor("op_250_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_257 = const()[name = tensor("op_257"), val = tensor([1, 77, 20, 64])]; + tensor var_258_cast_fp16 = reshape(shape = var_257, x = tensor_11_cast_fp16)[name = tensor("op_258_cast_fp16")]; + tensor var_259_perm_0 = const()[name = tensor("op_259_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_261 = const()[name = tensor("op_261"), val = tensor([20, -1, 64])]; + tensor var_259_cast_fp16 = transpose(perm = var_259_perm_0, x = var_258_cast_fp16)[name = tensor("transpose_123")]; + tensor query_states_3_cast_fp16 = reshape(shape = var_261, x = var_259_cast_fp16)[name = tensor("query_states_3_cast_fp16")]; + tensor var_263 = const()[name = tensor("op_263"), val = tensor([20, -1, 64])]; + tensor var_243_cast_fp16 = transpose(perm = var_243_perm_0, x = var_242_cast_fp16)[name = tensor("transpose_122")]; + tensor key_states_7_cast_fp16 = reshape(shape = var_263, x = var_243_cast_fp16)[name = tensor("key_states_7_cast_fp16")]; + tensor var_265 = const()[name = tensor("op_265"), val = tensor([20, -1, 64])]; + tensor var_250_cast_fp16 = transpose(perm = var_250_perm_0, x = var_249_cast_fp16)[name = tensor("transpose_121")]; + tensor value_states_7_cast_fp16 = reshape(shape = var_265, x = var_250_cast_fp16)[name = tensor("value_states_7_cast_fp16")]; + tensor attn_weights_7_transpose_x_1 = const()[name = tensor("attn_weights_7_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_7_transpose_y_1 = const()[name = tensor("attn_weights_7_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_7_cast_fp16 = matmul(transpose_x = attn_weights_7_transpose_x_1, transpose_y = attn_weights_7_transpose_y_1, x = query_states_3_cast_fp16, y = key_states_7_cast_fp16)[name = tensor("attn_weights_7_cast_fp16")]; + tensor var_270 = const()[name = tensor("op_270"), val = tensor([1, 20, 77, 77])]; + tensor var_271_cast_fp16 = reshape(shape = var_270, x = attn_weights_7_cast_fp16)[name = tensor("op_271_cast_fp16")]; + tensor attn_weights_9_cast_fp16 = add(x = var_271_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_9_cast_fp16")]; + tensor var_276 = const()[name = tensor("op_276"), val = tensor([20, 77, 77])]; + tensor input_21_cast_fp16 = reshape(shape = var_276, x = attn_weights_9_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor input_23_cast_fp16 = softmax(axis = var_5, x = input_21_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor attn_output_7_transpose_x_0 = const()[name = tensor("attn_output_7_transpose_x_0"), val = tensor(false)]; + tensor attn_output_7_transpose_y_0 = const()[name = tensor("attn_output_7_transpose_y_0"), val = tensor(false)]; + tensor attn_output_7_cast_fp16 = matmul(transpose_x = attn_output_7_transpose_x_0, transpose_y = attn_output_7_transpose_y_0, x = input_23_cast_fp16, y = value_states_7_cast_fp16)[name = tensor("attn_output_7_cast_fp16")]; + tensor var_281 = const()[name = tensor("op_281"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_9_cast_fp16 = reshape(shape = var_281, x = attn_output_7_cast_fp16)[name = tensor("attn_output_9_cast_fp16")]; + tensor attn_output_11_perm_0 = const()[name = tensor("attn_output_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_284 = const()[name = tensor("op_284"), val = tensor([1, 77, 1280])]; + tensor attn_output_11_cast_fp16 = transpose(perm = attn_output_11_perm_0, x = attn_output_9_cast_fp16)[name = tensor("transpose_120")]; + tensor input_25_cast_fp16 = reshape(shape = var_284, x = attn_output_11_cast_fp16)[name = tensor("input_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145044864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146273728))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146273920)))]; + tensor linear_9_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor input_27_cast_fp16 = add(x = input_19_cast_fp16, y = linear_9_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor input_29_axes_0 = const()[name = tensor("input_29_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146276544)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146279168)))]; + tensor input_29_cast_fp16 = layer_norm(axes = input_29_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("input_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146281792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151197056))), name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151197248)))]; + tensor linear_10_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized, x = input_29_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_33_mode_0 = const()[name = tensor("input_33_mode_0"), val = tensor("EXACT")]; + tensor input_33_cast_fp16 = gelu(mode = input_33_mode_0, x = linear_10_cast_fp16)[name = tensor("input_33_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151207552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156122816))), name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156123008)))]; + tensor linear_11_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized, x = input_33_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor input_35_cast_fp16 = add(x = input_27_cast_fp16, y = linear_11_cast_fp16)[name = tensor("input_35_cast_fp16")]; + tensor hidden_states_13_axes_0 = const()[name = tensor("hidden_states_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156125632)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156128256)))]; + tensor hidden_states_13_cast_fp16 = layer_norm(axes = hidden_states_13_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("hidden_states_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156130880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157359744))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157359936)))]; + tensor linear_12_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_323_to_fp16 = const()[name = tensor("op_323_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_17_cast_fp16 = mul(x = linear_12_cast_fp16, y = var_323_to_fp16)[name = tensor("tensor_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157362560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158591424))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158591616)))]; + tensor linear_13_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_328 = const()[name = tensor("op_328"), val = tensor([1, -1, 20, 64])]; + tensor var_329_cast_fp16 = reshape(shape = var_328, x = linear_13_cast_fp16)[name = tensor("op_329_cast_fp16")]; + tensor var_330_perm_0 = const()[name = tensor("op_330_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158594240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159823104))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159823296)))]; + tensor linear_14_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_335 = const()[name = tensor("op_335"), val = tensor([1, -1, 20, 64])]; + tensor var_336_cast_fp16 = reshape(shape = var_335, x = linear_14_cast_fp16)[name = tensor("op_336_cast_fp16")]; + tensor var_337_perm_0 = const()[name = tensor("op_337_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_344 = const()[name = tensor("op_344"), val = tensor([1, 77, 20, 64])]; + tensor var_345_cast_fp16 = reshape(shape = var_344, x = tensor_17_cast_fp16)[name = tensor("op_345_cast_fp16")]; + tensor var_346_perm_0 = const()[name = tensor("op_346_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_348 = const()[name = tensor("op_348"), val = tensor([20, -1, 64])]; + tensor var_346_cast_fp16 = transpose(perm = var_346_perm_0, x = var_345_cast_fp16)[name = tensor("transpose_119")]; + tensor query_states_5_cast_fp16 = reshape(shape = var_348, x = var_346_cast_fp16)[name = tensor("query_states_5_cast_fp16")]; + tensor var_350 = const()[name = tensor("op_350"), val = tensor([20, -1, 64])]; + tensor var_330_cast_fp16 = transpose(perm = var_330_perm_0, x = var_329_cast_fp16)[name = tensor("transpose_118")]; + tensor key_states_11_cast_fp16 = reshape(shape = var_350, x = var_330_cast_fp16)[name = tensor("key_states_11_cast_fp16")]; + tensor var_352 = const()[name = tensor("op_352"), val = tensor([20, -1, 64])]; + tensor var_337_cast_fp16 = transpose(perm = var_337_perm_0, x = var_336_cast_fp16)[name = tensor("transpose_117")]; + tensor value_states_11_cast_fp16 = reshape(shape = var_352, x = var_337_cast_fp16)[name = tensor("value_states_11_cast_fp16")]; + tensor attn_weights_13_transpose_x_1 = const()[name = tensor("attn_weights_13_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_13_transpose_y_1 = const()[name = tensor("attn_weights_13_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_13_cast_fp16 = matmul(transpose_x = attn_weights_13_transpose_x_1, transpose_y = attn_weights_13_transpose_y_1, x = query_states_5_cast_fp16, y = key_states_11_cast_fp16)[name = tensor("attn_weights_13_cast_fp16")]; + tensor var_357 = const()[name = tensor("op_357"), val = tensor([1, 20, 77, 77])]; + tensor var_358_cast_fp16 = reshape(shape = var_357, x = attn_weights_13_cast_fp16)[name = tensor("op_358_cast_fp16")]; + tensor attn_weights_15_cast_fp16 = add(x = var_358_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_15_cast_fp16")]; + tensor var_363 = const()[name = tensor("op_363"), val = tensor([20, 77, 77])]; + tensor input_37_cast_fp16 = reshape(shape = var_363, x = attn_weights_15_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor input_39_cast_fp16 = softmax(axis = var_5, x = input_37_cast_fp16)[name = tensor("input_39_cast_fp16")]; + tensor attn_output_13_transpose_x_0 = const()[name = tensor("attn_output_13_transpose_x_0"), val = tensor(false)]; + tensor attn_output_13_transpose_y_0 = const()[name = tensor("attn_output_13_transpose_y_0"), val = tensor(false)]; + tensor attn_output_13_cast_fp16 = matmul(transpose_x = attn_output_13_transpose_x_0, transpose_y = attn_output_13_transpose_y_0, x = input_39_cast_fp16, y = value_states_11_cast_fp16)[name = tensor("attn_output_13_cast_fp16")]; + tensor var_368 = const()[name = tensor("op_368"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_15_cast_fp16 = reshape(shape = var_368, x = attn_output_13_cast_fp16)[name = tensor("attn_output_15_cast_fp16")]; + tensor attn_output_17_perm_0 = const()[name = tensor("attn_output_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_371 = const()[name = tensor("op_371"), val = tensor([1, 77, 1280])]; + tensor attn_output_17_cast_fp16 = transpose(perm = attn_output_17_perm_0, x = attn_output_15_cast_fp16)[name = tensor("transpose_116")]; + tensor input_41_cast_fp16 = reshape(shape = var_371, x = attn_output_17_cast_fp16)[name = tensor("input_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159825920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161054784))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161054976)))]; + tensor linear_15_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized, x = input_41_cast_fp16)[name = tensor("linear_15_cast_fp16")]; + tensor input_43_cast_fp16 = add(x = input_35_cast_fp16, y = linear_15_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor input_45_axes_0 = const()[name = tensor("input_45_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161057600)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161060224)))]; + tensor input_45_cast_fp16 = layer_norm(axes = input_45_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("input_45_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161062848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165978112))), name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165978304)))]; + tensor linear_16_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized, x = input_45_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_49_mode_0 = const()[name = tensor("input_49_mode_0"), val = tensor("EXACT")]; + tensor input_49_cast_fp16 = gelu(mode = input_49_mode_0, x = linear_16_cast_fp16)[name = tensor("input_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165988608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170903872))), name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170904064)))]; + tensor linear_17_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized, x = input_49_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_51_cast_fp16 = add(x = input_43_cast_fp16, y = linear_17_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor hidden_states_19_axes_0 = const()[name = tensor("hidden_states_19_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170906688)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170909312)))]; + tensor hidden_states_19_cast_fp16 = layer_norm(axes = hidden_states_19_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16, x = input_51_cast_fp16)[name = tensor("hidden_states_19_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170911936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172140800))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172140992)))]; + tensor linear_18_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_410_to_fp16 = const()[name = tensor("op_410_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_23_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_410_to_fp16)[name = tensor("tensor_23_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172143616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173372480))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173372672)))]; + tensor linear_19_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor var_415 = const()[name = tensor("op_415"), val = tensor([1, -1, 20, 64])]; + tensor var_416_cast_fp16 = reshape(shape = var_415, x = linear_19_cast_fp16)[name = tensor("op_416_cast_fp16")]; + tensor var_417_perm_0 = const()[name = tensor("op_417_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173375296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174604160))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174604352)))]; + tensor linear_20_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_422 = const()[name = tensor("op_422"), val = tensor([1, -1, 20, 64])]; + tensor var_423_cast_fp16 = reshape(shape = var_422, x = linear_20_cast_fp16)[name = tensor("op_423_cast_fp16")]; + tensor var_424_perm_0 = const()[name = tensor("op_424_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_431 = const()[name = tensor("op_431"), val = tensor([1, 77, 20, 64])]; + tensor var_432_cast_fp16 = reshape(shape = var_431, x = tensor_23_cast_fp16)[name = tensor("op_432_cast_fp16")]; + tensor var_433_perm_0 = const()[name = tensor("op_433_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_435 = const()[name = tensor("op_435"), val = tensor([20, -1, 64])]; + tensor var_433_cast_fp16 = transpose(perm = var_433_perm_0, x = var_432_cast_fp16)[name = tensor("transpose_115")]; + tensor query_states_7_cast_fp16 = reshape(shape = var_435, x = var_433_cast_fp16)[name = tensor("query_states_7_cast_fp16")]; + tensor var_437 = const()[name = tensor("op_437"), val = tensor([20, -1, 64])]; + tensor var_417_cast_fp16 = transpose(perm = var_417_perm_0, x = var_416_cast_fp16)[name = tensor("transpose_114")]; + tensor key_states_15_cast_fp16 = reshape(shape = var_437, x = var_417_cast_fp16)[name = tensor("key_states_15_cast_fp16")]; + tensor var_439 = const()[name = tensor("op_439"), val = tensor([20, -1, 64])]; + tensor var_424_cast_fp16 = transpose(perm = var_424_perm_0, x = var_423_cast_fp16)[name = tensor("transpose_113")]; + tensor value_states_15_cast_fp16 = reshape(shape = var_439, x = var_424_cast_fp16)[name = tensor("value_states_15_cast_fp16")]; + tensor attn_weights_19_transpose_x_1 = const()[name = tensor("attn_weights_19_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_19_transpose_y_1 = const()[name = tensor("attn_weights_19_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_19_cast_fp16 = matmul(transpose_x = attn_weights_19_transpose_x_1, transpose_y = attn_weights_19_transpose_y_1, x = query_states_7_cast_fp16, y = key_states_15_cast_fp16)[name = tensor("attn_weights_19_cast_fp16")]; + tensor var_444 = const()[name = tensor("op_444"), val = tensor([1, 20, 77, 77])]; + tensor var_445_cast_fp16 = reshape(shape = var_444, x = attn_weights_19_cast_fp16)[name = tensor("op_445_cast_fp16")]; + tensor attn_weights_21_cast_fp16 = add(x = var_445_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_21_cast_fp16")]; + tensor var_450 = const()[name = tensor("op_450"), val = tensor([20, 77, 77])]; + tensor input_53_cast_fp16 = reshape(shape = var_450, x = attn_weights_21_cast_fp16)[name = tensor("input_53_cast_fp16")]; + tensor input_55_cast_fp16 = softmax(axis = var_5, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; + tensor attn_output_19_transpose_x_0 = const()[name = tensor("attn_output_19_transpose_x_0"), val = tensor(false)]; + tensor attn_output_19_transpose_y_0 = const()[name = tensor("attn_output_19_transpose_y_0"), val = tensor(false)]; + tensor attn_output_19_cast_fp16 = matmul(transpose_x = attn_output_19_transpose_x_0, transpose_y = attn_output_19_transpose_y_0, x = input_55_cast_fp16, y = value_states_15_cast_fp16)[name = tensor("attn_output_19_cast_fp16")]; + tensor var_455 = const()[name = tensor("op_455"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_21_cast_fp16 = reshape(shape = var_455, x = attn_output_19_cast_fp16)[name = tensor("attn_output_21_cast_fp16")]; + tensor attn_output_23_perm_0 = const()[name = tensor("attn_output_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_458 = const()[name = tensor("op_458"), val = tensor([1, 77, 1280])]; + tensor attn_output_23_cast_fp16 = transpose(perm = attn_output_23_perm_0, x = attn_output_21_cast_fp16)[name = tensor("transpose_112")]; + tensor input_57_cast_fp16 = reshape(shape = var_458, x = attn_output_23_cast_fp16)[name = tensor("input_57_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174606976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175835840))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175836032)))]; + tensor linear_21_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized, x = input_57_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor input_59_cast_fp16 = add(x = input_51_cast_fp16, y = linear_21_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor input_61_axes_0 = const()[name = tensor("input_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175838656)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175841280)))]; + tensor input_61_cast_fp16 = layer_norm(axes = input_61_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175843904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180759168))), name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180759360)))]; + tensor linear_22_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized, x = input_61_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor input_65_mode_0 = const()[name = tensor("input_65_mode_0"), val = tensor("EXACT")]; + tensor input_65_cast_fp16 = gelu(mode = input_65_mode_0, x = linear_22_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180769664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185684928))), name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185685120)))]; + tensor linear_23_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized, x = input_65_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor input_67_cast_fp16 = add(x = input_59_cast_fp16, y = linear_23_cast_fp16)[name = tensor("input_67_cast_fp16")]; + tensor hidden_states_25_axes_0 = const()[name = tensor("hidden_states_25_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185687744)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185690368)))]; + tensor hidden_states_25_cast_fp16 = layer_norm(axes = hidden_states_25_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("hidden_states_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185692992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186921856))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186922048)))]; + tensor linear_24_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_24_cast_fp16")]; + tensor var_497_to_fp16 = const()[name = tensor("op_497_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_29_cast_fp16 = mul(x = linear_24_cast_fp16, y = var_497_to_fp16)[name = tensor("tensor_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186924672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188153536))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188153728)))]; + tensor linear_25_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor var_502 = const()[name = tensor("op_502"), val = tensor([1, -1, 20, 64])]; + tensor var_503_cast_fp16 = reshape(shape = var_502, x = linear_25_cast_fp16)[name = tensor("op_503_cast_fp16")]; + tensor var_504_perm_0 = const()[name = tensor("op_504_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188156352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189385216))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189385408)))]; + tensor linear_26_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor var_509 = const()[name = tensor("op_509"), val = tensor([1, -1, 20, 64])]; + tensor var_510_cast_fp16 = reshape(shape = var_509, x = linear_26_cast_fp16)[name = tensor("op_510_cast_fp16")]; + tensor var_511_perm_0 = const()[name = tensor("op_511_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_518 = const()[name = tensor("op_518"), val = tensor([1, 77, 20, 64])]; + tensor var_519_cast_fp16 = reshape(shape = var_518, x = tensor_29_cast_fp16)[name = tensor("op_519_cast_fp16")]; + tensor var_520_perm_0 = const()[name = tensor("op_520_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_522 = const()[name = tensor("op_522"), val = tensor([20, -1, 64])]; + tensor var_520_cast_fp16 = transpose(perm = var_520_perm_0, x = var_519_cast_fp16)[name = tensor("transpose_111")]; + tensor query_states_9_cast_fp16 = reshape(shape = var_522, x = var_520_cast_fp16)[name = tensor("query_states_9_cast_fp16")]; + tensor var_524 = const()[name = tensor("op_524"), val = tensor([20, -1, 64])]; + tensor var_504_cast_fp16 = transpose(perm = var_504_perm_0, x = var_503_cast_fp16)[name = tensor("transpose_110")]; + tensor key_states_19_cast_fp16 = reshape(shape = var_524, x = var_504_cast_fp16)[name = tensor("key_states_19_cast_fp16")]; + tensor var_526 = const()[name = tensor("op_526"), val = tensor([20, -1, 64])]; + tensor var_511_cast_fp16 = transpose(perm = var_511_perm_0, x = var_510_cast_fp16)[name = tensor("transpose_109")]; + tensor value_states_19_cast_fp16 = reshape(shape = var_526, x = var_511_cast_fp16)[name = tensor("value_states_19_cast_fp16")]; + tensor attn_weights_25_transpose_x_1 = const()[name = tensor("attn_weights_25_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_25_transpose_y_1 = const()[name = tensor("attn_weights_25_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_25_cast_fp16 = matmul(transpose_x = attn_weights_25_transpose_x_1, transpose_y = attn_weights_25_transpose_y_1, x = query_states_9_cast_fp16, y = key_states_19_cast_fp16)[name = tensor("attn_weights_25_cast_fp16")]; + tensor var_531 = const()[name = tensor("op_531"), val = tensor([1, 20, 77, 77])]; + tensor var_532_cast_fp16 = reshape(shape = var_531, x = attn_weights_25_cast_fp16)[name = tensor("op_532_cast_fp16")]; + tensor attn_weights_27_cast_fp16 = add(x = var_532_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_27_cast_fp16")]; + tensor var_537 = const()[name = tensor("op_537"), val = tensor([20, 77, 77])]; + tensor input_69_cast_fp16 = reshape(shape = var_537, x = attn_weights_27_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor input_71_cast_fp16 = softmax(axis = var_5, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; + tensor attn_output_25_transpose_x_0 = const()[name = tensor("attn_output_25_transpose_x_0"), val = tensor(false)]; + tensor attn_output_25_transpose_y_0 = const()[name = tensor("attn_output_25_transpose_y_0"), val = tensor(false)]; + tensor attn_output_25_cast_fp16 = matmul(transpose_x = attn_output_25_transpose_x_0, transpose_y = attn_output_25_transpose_y_0, x = input_71_cast_fp16, y = value_states_19_cast_fp16)[name = tensor("attn_output_25_cast_fp16")]; + tensor var_542 = const()[name = tensor("op_542"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_27_cast_fp16 = reshape(shape = var_542, x = attn_output_25_cast_fp16)[name = tensor("attn_output_27_cast_fp16")]; + tensor attn_output_29_perm_0 = const()[name = tensor("attn_output_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_545 = const()[name = tensor("op_545"), val = tensor([1, 77, 1280])]; + tensor attn_output_29_cast_fp16 = transpose(perm = attn_output_29_perm_0, x = attn_output_27_cast_fp16)[name = tensor("transpose_108")]; + tensor input_73_cast_fp16 = reshape(shape = var_545, x = attn_output_29_cast_fp16)[name = tensor("input_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189388032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190616896))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190617088)))]; + tensor linear_27_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized, x = input_73_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_67_cast_fp16, y = linear_27_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190619712)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190622336)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190624960))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195540224))), name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195540416)))]; + tensor linear_28_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized, x = input_77_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_81_mode_0 = const()[name = tensor("input_81_mode_0"), val = tensor("EXACT")]; + tensor input_81_cast_fp16 = gelu(mode = input_81_mode_0, x = linear_28_cast_fp16)[name = tensor("input_81_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195550720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200465984))), name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200466176)))]; + tensor linear_29_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized, x = input_81_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor input_83_cast_fp16 = add(x = input_75_cast_fp16, y = linear_29_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor hidden_states_31_axes_0 = const()[name = tensor("hidden_states_31_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200468800)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200471424)))]; + tensor hidden_states_31_cast_fp16 = layer_norm(axes = hidden_states_31_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("hidden_states_31_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200474048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201702912))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201703104)))]; + tensor linear_30_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_584_to_fp16 = const()[name = tensor("op_584_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_35_cast_fp16 = mul(x = linear_30_cast_fp16, y = var_584_to_fp16)[name = tensor("tensor_35_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201705728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202934592))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202934784)))]; + tensor linear_31_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_589 = const()[name = tensor("op_589"), val = tensor([1, -1, 20, 64])]; + tensor var_590_cast_fp16 = reshape(shape = var_589, x = linear_31_cast_fp16)[name = tensor("op_590_cast_fp16")]; + tensor var_591_perm_0 = const()[name = tensor("op_591_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202937408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204166272))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204166464)))]; + tensor linear_32_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_596 = const()[name = tensor("op_596"), val = tensor([1, -1, 20, 64])]; + tensor var_597_cast_fp16 = reshape(shape = var_596, x = linear_32_cast_fp16)[name = tensor("op_597_cast_fp16")]; + tensor var_598_perm_0 = const()[name = tensor("op_598_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_605 = const()[name = tensor("op_605"), val = tensor([1, 77, 20, 64])]; + tensor var_606_cast_fp16 = reshape(shape = var_605, x = tensor_35_cast_fp16)[name = tensor("op_606_cast_fp16")]; + tensor var_607_perm_0 = const()[name = tensor("op_607_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_609 = const()[name = tensor("op_609"), val = tensor([20, -1, 64])]; + tensor var_607_cast_fp16 = transpose(perm = var_607_perm_0, x = var_606_cast_fp16)[name = tensor("transpose_107")]; + tensor query_states_11_cast_fp16 = reshape(shape = var_609, x = var_607_cast_fp16)[name = tensor("query_states_11_cast_fp16")]; + tensor var_611 = const()[name = tensor("op_611"), val = tensor([20, -1, 64])]; + tensor var_591_cast_fp16 = transpose(perm = var_591_perm_0, x = var_590_cast_fp16)[name = tensor("transpose_106")]; + tensor key_states_23_cast_fp16 = reshape(shape = var_611, x = var_591_cast_fp16)[name = tensor("key_states_23_cast_fp16")]; + tensor var_613 = const()[name = tensor("op_613"), val = tensor([20, -1, 64])]; + tensor var_598_cast_fp16 = transpose(perm = var_598_perm_0, x = var_597_cast_fp16)[name = tensor("transpose_105")]; + tensor value_states_23_cast_fp16 = reshape(shape = var_613, x = var_598_cast_fp16)[name = tensor("value_states_23_cast_fp16")]; + tensor attn_weights_31_transpose_x_1 = const()[name = tensor("attn_weights_31_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_31_transpose_y_1 = const()[name = tensor("attn_weights_31_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_31_cast_fp16 = matmul(transpose_x = attn_weights_31_transpose_x_1, transpose_y = attn_weights_31_transpose_y_1, x = query_states_11_cast_fp16, y = key_states_23_cast_fp16)[name = tensor("attn_weights_31_cast_fp16")]; + tensor var_618 = const()[name = tensor("op_618"), val = tensor([1, 20, 77, 77])]; + tensor var_619_cast_fp16 = reshape(shape = var_618, x = attn_weights_31_cast_fp16)[name = tensor("op_619_cast_fp16")]; + tensor attn_weights_33_cast_fp16 = add(x = var_619_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_33_cast_fp16")]; + tensor var_624 = const()[name = tensor("op_624"), val = tensor([20, 77, 77])]; + tensor input_85_cast_fp16 = reshape(shape = var_624, x = attn_weights_33_cast_fp16)[name = tensor("input_85_cast_fp16")]; + tensor input_87_cast_fp16 = softmax(axis = var_5, x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; + tensor attn_output_31_transpose_x_0 = const()[name = tensor("attn_output_31_transpose_x_0"), val = tensor(false)]; + tensor attn_output_31_transpose_y_0 = const()[name = tensor("attn_output_31_transpose_y_0"), val = tensor(false)]; + tensor attn_output_31_cast_fp16 = matmul(transpose_x = attn_output_31_transpose_x_0, transpose_y = attn_output_31_transpose_y_0, x = input_87_cast_fp16, y = value_states_23_cast_fp16)[name = tensor("attn_output_31_cast_fp16")]; + tensor var_629 = const()[name = tensor("op_629"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_33_cast_fp16 = reshape(shape = var_629, x = attn_output_31_cast_fp16)[name = tensor("attn_output_33_cast_fp16")]; + tensor attn_output_35_perm_0 = const()[name = tensor("attn_output_35_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_632 = const()[name = tensor("op_632"), val = tensor([1, 77, 1280])]; + tensor attn_output_35_cast_fp16 = transpose(perm = attn_output_35_perm_0, x = attn_output_33_cast_fp16)[name = tensor("transpose_104")]; + tensor input_89_cast_fp16 = reshape(shape = var_632, x = attn_output_35_cast_fp16)[name = tensor("input_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204169088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205397952))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205398144)))]; + tensor linear_33_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized, x = input_89_cast_fp16)[name = tensor("linear_33_cast_fp16")]; + tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_33_cast_fp16)[name = tensor("input_91_cast_fp16")]; + tensor input_93_axes_0 = const()[name = tensor("input_93_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205400768)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205403392)))]; + tensor input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205406016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210321280))), name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210321472)))]; + tensor linear_34_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized, x = input_93_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_97_mode_0 = const()[name = tensor("input_97_mode_0"), val = tensor("EXACT")]; + tensor input_97_cast_fp16 = gelu(mode = input_97_mode_0, x = linear_34_cast_fp16)[name = tensor("input_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210331776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215247040))), name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215247232)))]; + tensor linear_35_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized, x = input_97_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_35_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor hidden_states_37_axes_0 = const()[name = tensor("hidden_states_37_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215249856)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215252480)))]; + tensor hidden_states_37_cast_fp16 = layer_norm(axes = hidden_states_37_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("hidden_states_37_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215255104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216483968))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216484160)))]; + tensor linear_36_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_671_to_fp16 = const()[name = tensor("op_671_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_41_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_671_to_fp16)[name = tensor("tensor_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216486784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217715648))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217715840)))]; + tensor linear_37_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor var_676 = const()[name = tensor("op_676"), val = tensor([1, -1, 20, 64])]; + tensor var_677_cast_fp16 = reshape(shape = var_676, x = linear_37_cast_fp16)[name = tensor("op_677_cast_fp16")]; + tensor var_678_perm_0 = const()[name = tensor("op_678_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217718464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218947328))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218947520)))]; + tensor linear_38_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_683 = const()[name = tensor("op_683"), val = tensor([1, -1, 20, 64])]; + tensor var_684_cast_fp16 = reshape(shape = var_683, x = linear_38_cast_fp16)[name = tensor("op_684_cast_fp16")]; + tensor var_685_perm_0 = const()[name = tensor("op_685_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_692 = const()[name = tensor("op_692"), val = tensor([1, 77, 20, 64])]; + tensor var_693_cast_fp16 = reshape(shape = var_692, x = tensor_41_cast_fp16)[name = tensor("op_693_cast_fp16")]; + tensor var_694_perm_0 = const()[name = tensor("op_694_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_696 = const()[name = tensor("op_696"), val = tensor([20, -1, 64])]; + tensor var_694_cast_fp16 = transpose(perm = var_694_perm_0, x = var_693_cast_fp16)[name = tensor("transpose_103")]; + tensor query_states_13_cast_fp16 = reshape(shape = var_696, x = var_694_cast_fp16)[name = tensor("query_states_13_cast_fp16")]; + tensor var_698 = const()[name = tensor("op_698"), val = tensor([20, -1, 64])]; + tensor var_678_cast_fp16 = transpose(perm = var_678_perm_0, x = var_677_cast_fp16)[name = tensor("transpose_102")]; + tensor key_states_27_cast_fp16 = reshape(shape = var_698, x = var_678_cast_fp16)[name = tensor("key_states_27_cast_fp16")]; + tensor var_700 = const()[name = tensor("op_700"), val = tensor([20, -1, 64])]; + tensor var_685_cast_fp16 = transpose(perm = var_685_perm_0, x = var_684_cast_fp16)[name = tensor("transpose_101")]; + tensor value_states_27_cast_fp16 = reshape(shape = var_700, x = var_685_cast_fp16)[name = tensor("value_states_27_cast_fp16")]; + tensor attn_weights_37_transpose_x_1 = const()[name = tensor("attn_weights_37_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_37_transpose_y_1 = const()[name = tensor("attn_weights_37_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_37_cast_fp16 = matmul(transpose_x = attn_weights_37_transpose_x_1, transpose_y = attn_weights_37_transpose_y_1, x = query_states_13_cast_fp16, y = key_states_27_cast_fp16)[name = tensor("attn_weights_37_cast_fp16")]; + tensor var_705 = const()[name = tensor("op_705"), val = tensor([1, 20, 77, 77])]; + tensor var_706_cast_fp16 = reshape(shape = var_705, x = attn_weights_37_cast_fp16)[name = tensor("op_706_cast_fp16")]; + tensor attn_weights_39_cast_fp16 = add(x = var_706_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_39_cast_fp16")]; + tensor var_711 = const()[name = tensor("op_711"), val = tensor([20, 77, 77])]; + tensor input_101_cast_fp16 = reshape(shape = var_711, x = attn_weights_39_cast_fp16)[name = tensor("input_101_cast_fp16")]; + tensor input_103_cast_fp16 = softmax(axis = var_5, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor attn_output_37_transpose_x_0 = const()[name = tensor("attn_output_37_transpose_x_0"), val = tensor(false)]; + tensor attn_output_37_transpose_y_0 = const()[name = tensor("attn_output_37_transpose_y_0"), val = tensor(false)]; + tensor attn_output_37_cast_fp16 = matmul(transpose_x = attn_output_37_transpose_x_0, transpose_y = attn_output_37_transpose_y_0, x = input_103_cast_fp16, y = value_states_27_cast_fp16)[name = tensor("attn_output_37_cast_fp16")]; + tensor var_716 = const()[name = tensor("op_716"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_39_cast_fp16 = reshape(shape = var_716, x = attn_output_37_cast_fp16)[name = tensor("attn_output_39_cast_fp16")]; + tensor attn_output_41_perm_0 = const()[name = tensor("attn_output_41_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_719 = const()[name = tensor("op_719"), val = tensor([1, 77, 1280])]; + tensor attn_output_41_cast_fp16 = transpose(perm = attn_output_41_perm_0, x = attn_output_39_cast_fp16)[name = tensor("transpose_100")]; + tensor input_105_cast_fp16 = reshape(shape = var_719, x = attn_output_41_cast_fp16)[name = tensor("input_105_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218950144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220179008))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220179200)))]; + tensor linear_39_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized, x = input_105_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor input_107_cast_fp16 = add(x = input_99_cast_fp16, y = linear_39_cast_fp16)[name = tensor("input_107_cast_fp16")]; + tensor input_109_axes_0 = const()[name = tensor("input_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220181824)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220184448)))]; + tensor input_109_cast_fp16 = layer_norm(axes = input_109_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220187072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225102336))), name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225102528)))]; + tensor linear_40_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized, x = input_109_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor input_113_mode_0 = const()[name = tensor("input_113_mode_0"), val = tensor("EXACT")]; + tensor input_113_cast_fp16 = gelu(mode = input_113_mode_0, x = linear_40_cast_fp16)[name = tensor("input_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225112832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230028096))), name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230028288)))]; + tensor linear_41_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized, x = input_113_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor input_115_cast_fp16 = add(x = input_107_cast_fp16, y = linear_41_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor hidden_states_43_axes_0 = const()[name = tensor("hidden_states_43_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230030912)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230033536)))]; + tensor hidden_states_43_cast_fp16 = layer_norm(axes = hidden_states_43_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("hidden_states_43_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(230036160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231265024))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231265216)))]; + tensor linear_42_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_42_cast_fp16")]; + tensor var_758_to_fp16 = const()[name = tensor("op_758_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_47_cast_fp16 = mul(x = linear_42_cast_fp16, y = var_758_to_fp16)[name = tensor("tensor_47_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231267840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232496704))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232496896)))]; + tensor linear_43_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor var_763 = const()[name = tensor("op_763"), val = tensor([1, -1, 20, 64])]; + tensor var_764_cast_fp16 = reshape(shape = var_763, x = linear_43_cast_fp16)[name = tensor("op_764_cast_fp16")]; + tensor var_765_perm_0 = const()[name = tensor("op_765_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232499520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233728384))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233728576)))]; + tensor linear_44_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor var_770 = const()[name = tensor("op_770"), val = tensor([1, -1, 20, 64])]; + tensor var_771_cast_fp16 = reshape(shape = var_770, x = linear_44_cast_fp16)[name = tensor("op_771_cast_fp16")]; + tensor var_772_perm_0 = const()[name = tensor("op_772_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_779 = const()[name = tensor("op_779"), val = tensor([1, 77, 20, 64])]; + tensor var_780_cast_fp16 = reshape(shape = var_779, x = tensor_47_cast_fp16)[name = tensor("op_780_cast_fp16")]; + tensor var_781_perm_0 = const()[name = tensor("op_781_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_783 = const()[name = tensor("op_783"), val = tensor([20, -1, 64])]; + tensor var_781_cast_fp16 = transpose(perm = var_781_perm_0, x = var_780_cast_fp16)[name = tensor("transpose_99")]; + tensor query_states_15_cast_fp16 = reshape(shape = var_783, x = var_781_cast_fp16)[name = tensor("query_states_15_cast_fp16")]; + tensor var_785 = const()[name = tensor("op_785"), val = tensor([20, -1, 64])]; + tensor var_765_cast_fp16 = transpose(perm = var_765_perm_0, x = var_764_cast_fp16)[name = tensor("transpose_98")]; + tensor key_states_31_cast_fp16 = reshape(shape = var_785, x = var_765_cast_fp16)[name = tensor("key_states_31_cast_fp16")]; + tensor var_787 = const()[name = tensor("op_787"), val = tensor([20, -1, 64])]; + tensor var_772_cast_fp16 = transpose(perm = var_772_perm_0, x = var_771_cast_fp16)[name = tensor("transpose_97")]; + tensor value_states_31_cast_fp16 = reshape(shape = var_787, x = var_772_cast_fp16)[name = tensor("value_states_31_cast_fp16")]; + tensor attn_weights_43_transpose_x_1 = const()[name = tensor("attn_weights_43_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_43_transpose_y_1 = const()[name = tensor("attn_weights_43_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_43_cast_fp16 = matmul(transpose_x = attn_weights_43_transpose_x_1, transpose_y = attn_weights_43_transpose_y_1, x = query_states_15_cast_fp16, y = key_states_31_cast_fp16)[name = tensor("attn_weights_43_cast_fp16")]; + tensor var_792 = const()[name = tensor("op_792"), val = tensor([1, 20, 77, 77])]; + tensor var_793_cast_fp16 = reshape(shape = var_792, x = attn_weights_43_cast_fp16)[name = tensor("op_793_cast_fp16")]; + tensor attn_weights_45_cast_fp16 = add(x = var_793_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_45_cast_fp16")]; + tensor var_798 = const()[name = tensor("op_798"), val = tensor([20, 77, 77])]; + tensor input_117_cast_fp16 = reshape(shape = var_798, x = attn_weights_45_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor input_119_cast_fp16 = softmax(axis = var_5, x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; + tensor attn_output_43_transpose_x_0 = const()[name = tensor("attn_output_43_transpose_x_0"), val = tensor(false)]; + tensor attn_output_43_transpose_y_0 = const()[name = tensor("attn_output_43_transpose_y_0"), val = tensor(false)]; + tensor attn_output_43_cast_fp16 = matmul(transpose_x = attn_output_43_transpose_x_0, transpose_y = attn_output_43_transpose_y_0, x = input_119_cast_fp16, y = value_states_31_cast_fp16)[name = tensor("attn_output_43_cast_fp16")]; + tensor var_803 = const()[name = tensor("op_803"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_45_cast_fp16 = reshape(shape = var_803, x = attn_output_43_cast_fp16)[name = tensor("attn_output_45_cast_fp16")]; + tensor attn_output_47_perm_0 = const()[name = tensor("attn_output_47_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_806 = const()[name = tensor("op_806"), val = tensor([1, 77, 1280])]; + tensor attn_output_47_cast_fp16 = transpose(perm = attn_output_47_perm_0, x = attn_output_45_cast_fp16)[name = tensor("transpose_96")]; + tensor input_121_cast_fp16 = reshape(shape = var_806, x = attn_output_47_cast_fp16)[name = tensor("input_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233731200))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234960064))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234960256)))]; + tensor linear_45_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized, x = input_121_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor input_123_cast_fp16 = add(x = input_115_cast_fp16, y = linear_45_cast_fp16)[name = tensor("input_123_cast_fp16")]; + tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234962880)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234965504)))]; + tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234968128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239883392))), name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239883584)))]; + tensor linear_46_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized, x = input_125_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_129_mode_0 = const()[name = tensor("input_129_mode_0"), val = tensor("EXACT")]; + tensor input_129_cast_fp16 = gelu(mode = input_129_mode_0, x = linear_46_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239893888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244809152))), name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244809344)))]; + tensor linear_47_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized, x = input_129_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor input_131_cast_fp16 = add(x = input_123_cast_fp16, y = linear_47_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor hidden_states_49_axes_0 = const()[name = tensor("hidden_states_49_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244811968)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244814592)))]; + tensor hidden_states_49_cast_fp16 = layer_norm(axes = hidden_states_49_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("hidden_states_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244817216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(246046080))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(246046272)))]; + tensor linear_48_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_845_to_fp16 = const()[name = tensor("op_845_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_53_cast_fp16 = mul(x = linear_48_cast_fp16, y = var_845_to_fp16)[name = tensor("tensor_53_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(246048896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247277760))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247277952)))]; + tensor linear_49_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_850 = const()[name = tensor("op_850"), val = tensor([1, -1, 20, 64])]; + tensor var_851_cast_fp16 = reshape(shape = var_850, x = linear_49_cast_fp16)[name = tensor("op_851_cast_fp16")]; + tensor var_852_perm_0 = const()[name = tensor("op_852_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247280576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248509440))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248509632)))]; + tensor linear_50_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_857 = const()[name = tensor("op_857"), val = tensor([1, -1, 20, 64])]; + tensor var_858_cast_fp16 = reshape(shape = var_857, x = linear_50_cast_fp16)[name = tensor("op_858_cast_fp16")]; + tensor var_859_perm_0 = const()[name = tensor("op_859_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_866 = const()[name = tensor("op_866"), val = tensor([1, 77, 20, 64])]; + tensor var_867_cast_fp16 = reshape(shape = var_866, x = tensor_53_cast_fp16)[name = tensor("op_867_cast_fp16")]; + tensor var_868_perm_0 = const()[name = tensor("op_868_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_870 = const()[name = tensor("op_870"), val = tensor([20, -1, 64])]; + tensor var_868_cast_fp16 = transpose(perm = var_868_perm_0, x = var_867_cast_fp16)[name = tensor("transpose_95")]; + tensor query_states_17_cast_fp16 = reshape(shape = var_870, x = var_868_cast_fp16)[name = tensor("query_states_17_cast_fp16")]; + tensor var_872 = const()[name = tensor("op_872"), val = tensor([20, -1, 64])]; + tensor var_852_cast_fp16 = transpose(perm = var_852_perm_0, x = var_851_cast_fp16)[name = tensor("transpose_94")]; + tensor key_states_35_cast_fp16 = reshape(shape = var_872, x = var_852_cast_fp16)[name = tensor("key_states_35_cast_fp16")]; + tensor var_874 = const()[name = tensor("op_874"), val = tensor([20, -1, 64])]; + tensor var_859_cast_fp16 = transpose(perm = var_859_perm_0, x = var_858_cast_fp16)[name = tensor("transpose_93")]; + tensor value_states_35_cast_fp16 = reshape(shape = var_874, x = var_859_cast_fp16)[name = tensor("value_states_35_cast_fp16")]; + tensor attn_weights_49_transpose_x_1 = const()[name = tensor("attn_weights_49_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_49_transpose_y_1 = const()[name = tensor("attn_weights_49_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_49_cast_fp16 = matmul(transpose_x = attn_weights_49_transpose_x_1, transpose_y = attn_weights_49_transpose_y_1, x = query_states_17_cast_fp16, y = key_states_35_cast_fp16)[name = tensor("attn_weights_49_cast_fp16")]; + tensor var_879 = const()[name = tensor("op_879"), val = tensor([1, 20, 77, 77])]; + tensor var_880_cast_fp16 = reshape(shape = var_879, x = attn_weights_49_cast_fp16)[name = tensor("op_880_cast_fp16")]; + tensor attn_weights_51_cast_fp16 = add(x = var_880_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_51_cast_fp16")]; + tensor var_885 = const()[name = tensor("op_885"), val = tensor([20, 77, 77])]; + tensor input_133_cast_fp16 = reshape(shape = var_885, x = attn_weights_51_cast_fp16)[name = tensor("input_133_cast_fp16")]; + tensor input_135_cast_fp16 = softmax(axis = var_5, x = input_133_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor attn_output_49_transpose_x_0 = const()[name = tensor("attn_output_49_transpose_x_0"), val = tensor(false)]; + tensor attn_output_49_transpose_y_0 = const()[name = tensor("attn_output_49_transpose_y_0"), val = tensor(false)]; + tensor attn_output_49_cast_fp16 = matmul(transpose_x = attn_output_49_transpose_x_0, transpose_y = attn_output_49_transpose_y_0, x = input_135_cast_fp16, y = value_states_35_cast_fp16)[name = tensor("attn_output_49_cast_fp16")]; + tensor var_890 = const()[name = tensor("op_890"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_51_cast_fp16 = reshape(shape = var_890, x = attn_output_49_cast_fp16)[name = tensor("attn_output_51_cast_fp16")]; + tensor attn_output_53_perm_0 = const()[name = tensor("attn_output_53_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_893 = const()[name = tensor("op_893"), val = tensor([1, 77, 1280])]; + tensor attn_output_53_cast_fp16 = transpose(perm = attn_output_53_perm_0, x = attn_output_51_cast_fp16)[name = tensor("transpose_92")]; + tensor input_137_cast_fp16 = reshape(shape = var_893, x = attn_output_53_cast_fp16)[name = tensor("input_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248512256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249741120))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249741312)))]; + tensor linear_51_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized, x = input_137_cast_fp16)[name = tensor("linear_51_cast_fp16")]; + tensor input_139_cast_fp16 = add(x = input_131_cast_fp16, y = linear_51_cast_fp16)[name = tensor("input_139_cast_fp16")]; + tensor input_141_axes_0 = const()[name = tensor("input_141_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249743936)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249746560)))]; + tensor input_141_cast_fp16 = layer_norm(axes = input_141_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249749184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254664448))), name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254664640)))]; + tensor linear_52_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized, x = input_141_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_145_mode_0 = const()[name = tensor("input_145_mode_0"), val = tensor("EXACT")]; + tensor input_145_cast_fp16 = gelu(mode = input_145_mode_0, x = linear_52_cast_fp16)[name = tensor("input_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254674944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259590208))), name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259590400)))]; + tensor linear_53_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized, x = input_145_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_147_cast_fp16 = add(x = input_139_cast_fp16, y = linear_53_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor hidden_states_55_axes_0 = const()[name = tensor("hidden_states_55_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259593024)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259595648)))]; + tensor hidden_states_55_cast_fp16 = layer_norm(axes = hidden_states_55_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("hidden_states_55_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259598272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260827136))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260827328)))]; + tensor linear_54_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_932_to_fp16 = const()[name = tensor("op_932_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_59_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_932_to_fp16)[name = tensor("tensor_59_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260829952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262058816))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262059008)))]; + tensor linear_55_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor var_937 = const()[name = tensor("op_937"), val = tensor([1, -1, 20, 64])]; + tensor var_938_cast_fp16 = reshape(shape = var_937, x = linear_55_cast_fp16)[name = tensor("op_938_cast_fp16")]; + tensor var_939_perm_0 = const()[name = tensor("op_939_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262061632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263290496))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263290688)))]; + tensor linear_56_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_944 = const()[name = tensor("op_944"), val = tensor([1, -1, 20, 64])]; + tensor var_945_cast_fp16 = reshape(shape = var_944, x = linear_56_cast_fp16)[name = tensor("op_945_cast_fp16")]; + tensor var_946_perm_0 = const()[name = tensor("op_946_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_953 = const()[name = tensor("op_953"), val = tensor([1, 77, 20, 64])]; + tensor var_954_cast_fp16 = reshape(shape = var_953, x = tensor_59_cast_fp16)[name = tensor("op_954_cast_fp16")]; + tensor var_955_perm_0 = const()[name = tensor("op_955_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_957 = const()[name = tensor("op_957"), val = tensor([20, -1, 64])]; + tensor var_955_cast_fp16 = transpose(perm = var_955_perm_0, x = var_954_cast_fp16)[name = tensor("transpose_91")]; + tensor query_states_19_cast_fp16 = reshape(shape = var_957, x = var_955_cast_fp16)[name = tensor("query_states_19_cast_fp16")]; + tensor var_959 = const()[name = tensor("op_959"), val = tensor([20, -1, 64])]; + tensor var_939_cast_fp16 = transpose(perm = var_939_perm_0, x = var_938_cast_fp16)[name = tensor("transpose_90")]; + tensor key_states_39_cast_fp16 = reshape(shape = var_959, x = var_939_cast_fp16)[name = tensor("key_states_39_cast_fp16")]; + tensor var_961 = const()[name = tensor("op_961"), val = tensor([20, -1, 64])]; + tensor var_946_cast_fp16 = transpose(perm = var_946_perm_0, x = var_945_cast_fp16)[name = tensor("transpose_89")]; + tensor value_states_39_cast_fp16 = reshape(shape = var_961, x = var_946_cast_fp16)[name = tensor("value_states_39_cast_fp16")]; + tensor attn_weights_55_transpose_x_1 = const()[name = tensor("attn_weights_55_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_55_transpose_y_1 = const()[name = tensor("attn_weights_55_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_55_cast_fp16 = matmul(transpose_x = attn_weights_55_transpose_x_1, transpose_y = attn_weights_55_transpose_y_1, x = query_states_19_cast_fp16, y = key_states_39_cast_fp16)[name = tensor("attn_weights_55_cast_fp16")]; + tensor var_966 = const()[name = tensor("op_966"), val = tensor([1, 20, 77, 77])]; + tensor var_967_cast_fp16 = reshape(shape = var_966, x = attn_weights_55_cast_fp16)[name = tensor("op_967_cast_fp16")]; + tensor attn_weights_57_cast_fp16 = add(x = var_967_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_57_cast_fp16")]; + tensor var_972 = const()[name = tensor("op_972"), val = tensor([20, 77, 77])]; + tensor input_149_cast_fp16 = reshape(shape = var_972, x = attn_weights_57_cast_fp16)[name = tensor("input_149_cast_fp16")]; + tensor input_151_cast_fp16 = softmax(axis = var_5, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor attn_output_55_transpose_x_0 = const()[name = tensor("attn_output_55_transpose_x_0"), val = tensor(false)]; + tensor attn_output_55_transpose_y_0 = const()[name = tensor("attn_output_55_transpose_y_0"), val = tensor(false)]; + tensor attn_output_55_cast_fp16 = matmul(transpose_x = attn_output_55_transpose_x_0, transpose_y = attn_output_55_transpose_y_0, x = input_151_cast_fp16, y = value_states_39_cast_fp16)[name = tensor("attn_output_55_cast_fp16")]; + tensor var_977 = const()[name = tensor("op_977"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_57_cast_fp16 = reshape(shape = var_977, x = attn_output_55_cast_fp16)[name = tensor("attn_output_57_cast_fp16")]; + tensor attn_output_59_perm_0 = const()[name = tensor("attn_output_59_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_980 = const()[name = tensor("op_980"), val = tensor([1, 77, 1280])]; + tensor attn_output_59_cast_fp16 = transpose(perm = attn_output_59_perm_0, x = attn_output_57_cast_fp16)[name = tensor("transpose_88")]; + tensor input_153_cast_fp16 = reshape(shape = var_980, x = attn_output_59_cast_fp16)[name = tensor("input_153_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263293312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264522176))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264522368)))]; + tensor linear_57_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized, x = input_153_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor input_155_cast_fp16 = add(x = input_147_cast_fp16, y = linear_57_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor input_157_axes_0 = const()[name = tensor("input_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264524992)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264527616)))]; + tensor input_157_cast_fp16 = layer_norm(axes = input_157_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16, x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264530240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269445504))), name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269445696)))]; + tensor linear_58_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized, x = input_157_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor input_161_mode_0 = const()[name = tensor("input_161_mode_0"), val = tensor("EXACT")]; + tensor input_161_cast_fp16 = gelu(mode = input_161_mode_0, x = linear_58_cast_fp16)[name = tensor("input_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269456000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274371264))), name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274371456)))]; + tensor linear_59_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized, x = input_161_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor input_163_cast_fp16 = add(x = input_155_cast_fp16, y = linear_59_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor hidden_states_61_axes_0 = const()[name = tensor("hidden_states_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274374080)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274376704)))]; + tensor hidden_states_61_cast_fp16 = layer_norm(axes = hidden_states_61_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("hidden_states_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274379328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275608192))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275608384)))]; + tensor linear_60_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_60_cast_fp16")]; + tensor var_1019_to_fp16 = const()[name = tensor("op_1019_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_65_cast_fp16 = mul(x = linear_60_cast_fp16, y = var_1019_to_fp16)[name = tensor("tensor_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275611008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276839872))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276840064)))]; + tensor linear_61_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor var_1024 = const()[name = tensor("op_1024"), val = tensor([1, -1, 20, 64])]; + tensor var_1025_cast_fp16 = reshape(shape = var_1024, x = linear_61_cast_fp16)[name = tensor("op_1025_cast_fp16")]; + tensor var_1026_perm_0 = const()[name = tensor("op_1026_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276842688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278071552))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278071744)))]; + tensor linear_62_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor var_1031 = const()[name = tensor("op_1031"), val = tensor([1, -1, 20, 64])]; + tensor var_1032_cast_fp16 = reshape(shape = var_1031, x = linear_62_cast_fp16)[name = tensor("op_1032_cast_fp16")]; + tensor var_1033_perm_0 = const()[name = tensor("op_1033_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1040 = const()[name = tensor("op_1040"), val = tensor([1, 77, 20, 64])]; + tensor var_1041_cast_fp16 = reshape(shape = var_1040, x = tensor_65_cast_fp16)[name = tensor("op_1041_cast_fp16")]; + tensor var_1042_perm_0 = const()[name = tensor("op_1042_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1044 = const()[name = tensor("op_1044"), val = tensor([20, -1, 64])]; + tensor var_1042_cast_fp16 = transpose(perm = var_1042_perm_0, x = var_1041_cast_fp16)[name = tensor("transpose_87")]; + tensor query_states_21_cast_fp16 = reshape(shape = var_1044, x = var_1042_cast_fp16)[name = tensor("query_states_21_cast_fp16")]; + tensor var_1046 = const()[name = tensor("op_1046"), val = tensor([20, -1, 64])]; + tensor var_1026_cast_fp16 = transpose(perm = var_1026_perm_0, x = var_1025_cast_fp16)[name = tensor("transpose_86")]; + tensor key_states_43_cast_fp16 = reshape(shape = var_1046, x = var_1026_cast_fp16)[name = tensor("key_states_43_cast_fp16")]; + tensor var_1048 = const()[name = tensor("op_1048"), val = tensor([20, -1, 64])]; + tensor var_1033_cast_fp16 = transpose(perm = var_1033_perm_0, x = var_1032_cast_fp16)[name = tensor("transpose_85")]; + tensor value_states_43_cast_fp16 = reshape(shape = var_1048, x = var_1033_cast_fp16)[name = tensor("value_states_43_cast_fp16")]; + tensor attn_weights_61_transpose_x_1 = const()[name = tensor("attn_weights_61_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_61_transpose_y_1 = const()[name = tensor("attn_weights_61_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_61_cast_fp16 = matmul(transpose_x = attn_weights_61_transpose_x_1, transpose_y = attn_weights_61_transpose_y_1, x = query_states_21_cast_fp16, y = key_states_43_cast_fp16)[name = tensor("attn_weights_61_cast_fp16")]; + tensor var_1053 = const()[name = tensor("op_1053"), val = tensor([1, 20, 77, 77])]; + tensor var_1054_cast_fp16 = reshape(shape = var_1053, x = attn_weights_61_cast_fp16)[name = tensor("op_1054_cast_fp16")]; + tensor attn_weights_63_cast_fp16 = add(x = var_1054_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_63_cast_fp16")]; + tensor var_1059 = const()[name = tensor("op_1059"), val = tensor([20, 77, 77])]; + tensor input_165_cast_fp16 = reshape(shape = var_1059, x = attn_weights_63_cast_fp16)[name = tensor("input_165_cast_fp16")]; + tensor input_167_cast_fp16 = softmax(axis = var_5, x = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor attn_output_61_transpose_x_0 = const()[name = tensor("attn_output_61_transpose_x_0"), val = tensor(false)]; + tensor attn_output_61_transpose_y_0 = const()[name = tensor("attn_output_61_transpose_y_0"), val = tensor(false)]; + tensor attn_output_61_cast_fp16 = matmul(transpose_x = attn_output_61_transpose_x_0, transpose_y = attn_output_61_transpose_y_0, x = input_167_cast_fp16, y = value_states_43_cast_fp16)[name = tensor("attn_output_61_cast_fp16")]; + tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_63_cast_fp16 = reshape(shape = var_1064, x = attn_output_61_cast_fp16)[name = tensor("attn_output_63_cast_fp16")]; + tensor attn_output_65_perm_0 = const()[name = tensor("attn_output_65_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1067 = const()[name = tensor("op_1067"), val = tensor([1, 77, 1280])]; + tensor attn_output_65_cast_fp16 = transpose(perm = attn_output_65_perm_0, x = attn_output_63_cast_fp16)[name = tensor("transpose_84")]; + tensor input_169_cast_fp16 = reshape(shape = var_1067, x = attn_output_65_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278074368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279303232))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279303424)))]; + tensor linear_63_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized, x = input_169_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor input_171_cast_fp16 = add(x = input_163_cast_fp16, y = linear_63_cast_fp16)[name = tensor("input_171_cast_fp16")]; + tensor input_173_axes_0 = const()[name = tensor("input_173_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279306048)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279308672)))]; + tensor input_173_cast_fp16 = layer_norm(axes = input_173_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279311296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284226560))), name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284226752)))]; + tensor linear_64_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized, x = input_173_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_177_mode_0 = const()[name = tensor("input_177_mode_0"), val = tensor("EXACT")]; + tensor input_177_cast_fp16 = gelu(mode = input_177_mode_0, x = linear_64_cast_fp16)[name = tensor("input_177_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284237056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289152320))), name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289152512)))]; + tensor linear_65_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized, x = input_177_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_171_cast_fp16, y = linear_65_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor hidden_states_67_axes_0 = const()[name = tensor("hidden_states_67_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289155136)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289157760)))]; + tensor hidden_states_67_cast_fp16 = layer_norm(axes = hidden_states_67_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("hidden_states_67_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289160384))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290389248))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290389440)))]; + tensor linear_66_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1106_to_fp16 = const()[name = tensor("op_1106_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_71_cast_fp16 = mul(x = linear_66_cast_fp16, y = var_1106_to_fp16)[name = tensor("tensor_71_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290392064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291620928))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291621120)))]; + tensor linear_67_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1111 = const()[name = tensor("op_1111"), val = tensor([1, -1, 20, 64])]; + tensor var_1112_cast_fp16 = reshape(shape = var_1111, x = linear_67_cast_fp16)[name = tensor("op_1112_cast_fp16")]; + tensor var_1113_perm_0 = const()[name = tensor("op_1113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291623744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292852608))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292852800)))]; + tensor linear_68_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1118 = const()[name = tensor("op_1118"), val = tensor([1, -1, 20, 64])]; + tensor var_1119_cast_fp16 = reshape(shape = var_1118, x = linear_68_cast_fp16)[name = tensor("op_1119_cast_fp16")]; + tensor var_1120_perm_0 = const()[name = tensor("op_1120_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1127 = const()[name = tensor("op_1127"), val = tensor([1, 77, 20, 64])]; + tensor var_1128_cast_fp16 = reshape(shape = var_1127, x = tensor_71_cast_fp16)[name = tensor("op_1128_cast_fp16")]; + tensor var_1129_perm_0 = const()[name = tensor("op_1129_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1131 = const()[name = tensor("op_1131"), val = tensor([20, -1, 64])]; + tensor var_1129_cast_fp16 = transpose(perm = var_1129_perm_0, x = var_1128_cast_fp16)[name = tensor("transpose_83")]; + tensor query_states_23_cast_fp16 = reshape(shape = var_1131, x = var_1129_cast_fp16)[name = tensor("query_states_23_cast_fp16")]; + tensor var_1133 = const()[name = tensor("op_1133"), val = tensor([20, -1, 64])]; + tensor var_1113_cast_fp16 = transpose(perm = var_1113_perm_0, x = var_1112_cast_fp16)[name = tensor("transpose_82")]; + tensor key_states_47_cast_fp16 = reshape(shape = var_1133, x = var_1113_cast_fp16)[name = tensor("key_states_47_cast_fp16")]; + tensor var_1135 = const()[name = tensor("op_1135"), val = tensor([20, -1, 64])]; + tensor var_1120_cast_fp16 = transpose(perm = var_1120_perm_0, x = var_1119_cast_fp16)[name = tensor("transpose_81")]; + tensor value_states_47_cast_fp16 = reshape(shape = var_1135, x = var_1120_cast_fp16)[name = tensor("value_states_47_cast_fp16")]; + tensor attn_weights_67_transpose_x_1 = const()[name = tensor("attn_weights_67_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_67_transpose_y_1 = const()[name = tensor("attn_weights_67_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_67_cast_fp16 = matmul(transpose_x = attn_weights_67_transpose_x_1, transpose_y = attn_weights_67_transpose_y_1, x = query_states_23_cast_fp16, y = key_states_47_cast_fp16)[name = tensor("attn_weights_67_cast_fp16")]; + tensor var_1140 = const()[name = tensor("op_1140"), val = tensor([1, 20, 77, 77])]; + tensor var_1141_cast_fp16 = reshape(shape = var_1140, x = attn_weights_67_cast_fp16)[name = tensor("op_1141_cast_fp16")]; + tensor attn_weights_69_cast_fp16 = add(x = var_1141_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_69_cast_fp16")]; + tensor var_1146 = const()[name = tensor("op_1146"), val = tensor([20, 77, 77])]; + tensor input_181_cast_fp16 = reshape(shape = var_1146, x = attn_weights_69_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor input_183_cast_fp16 = softmax(axis = var_5, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor attn_output_67_transpose_x_0 = const()[name = tensor("attn_output_67_transpose_x_0"), val = tensor(false)]; + tensor attn_output_67_transpose_y_0 = const()[name = tensor("attn_output_67_transpose_y_0"), val = tensor(false)]; + tensor attn_output_67_cast_fp16 = matmul(transpose_x = attn_output_67_transpose_x_0, transpose_y = attn_output_67_transpose_y_0, x = input_183_cast_fp16, y = value_states_47_cast_fp16)[name = tensor("attn_output_67_cast_fp16")]; + tensor var_1151 = const()[name = tensor("op_1151"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_69_cast_fp16 = reshape(shape = var_1151, x = attn_output_67_cast_fp16)[name = tensor("attn_output_69_cast_fp16")]; + tensor attn_output_71_perm_0 = const()[name = tensor("attn_output_71_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1154 = const()[name = tensor("op_1154"), val = tensor([1, 77, 1280])]; + tensor attn_output_71_cast_fp16 = transpose(perm = attn_output_71_perm_0, x = attn_output_69_cast_fp16)[name = tensor("transpose_80")]; + tensor input_185_cast_fp16 = reshape(shape = var_1154, x = attn_output_71_cast_fp16)[name = tensor("input_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292855424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294084288))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294084480)))]; + tensor linear_69_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized, x = input_185_cast_fp16)[name = tensor("linear_69_cast_fp16")]; + tensor input_187_cast_fp16 = add(x = input_179_cast_fp16, y = linear_69_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor input_189_axes_0 = const()[name = tensor("input_189_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294087104)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294089728)))]; + tensor input_189_cast_fp16 = layer_norm(axes = input_189_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("input_189_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294092352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299007616))), name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299007808)))]; + tensor linear_70_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized, x = input_189_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_193_mode_0 = const()[name = tensor("input_193_mode_0"), val = tensor("EXACT")]; + tensor input_193_cast_fp16 = gelu(mode = input_193_mode_0, x = linear_70_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(299018112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303933376))), name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303933568)))]; + tensor linear_71_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized, x = input_193_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_71_cast_fp16)[name = tensor("input_195_cast_fp16")]; + tensor hidden_states_73_axes_0 = const()[name = tensor("hidden_states_73_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303936192)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303938816)))]; + tensor hidden_states_73_cast_fp16 = layer_norm(axes = hidden_states_73_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("hidden_states_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303941440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305170304))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305170496)))]; + tensor linear_72_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1193_to_fp16 = const()[name = tensor("op_1193_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_77_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1193_to_fp16)[name = tensor("tensor_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305173120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306401984))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306402176)))]; + tensor linear_73_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor var_1198 = const()[name = tensor("op_1198"), val = tensor([1, -1, 20, 64])]; + tensor var_1199_cast_fp16 = reshape(shape = var_1198, x = linear_73_cast_fp16)[name = tensor("op_1199_cast_fp16")]; + tensor var_1200_perm_0 = const()[name = tensor("op_1200_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306404800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307633664))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307633856)))]; + tensor linear_74_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1205 = const()[name = tensor("op_1205"), val = tensor([1, -1, 20, 64])]; + tensor var_1206_cast_fp16 = reshape(shape = var_1205, x = linear_74_cast_fp16)[name = tensor("op_1206_cast_fp16")]; + tensor var_1207_perm_0 = const()[name = tensor("op_1207_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1214 = const()[name = tensor("op_1214"), val = tensor([1, 77, 20, 64])]; + tensor var_1215_cast_fp16 = reshape(shape = var_1214, x = tensor_77_cast_fp16)[name = tensor("op_1215_cast_fp16")]; + tensor var_1216_perm_0 = const()[name = tensor("op_1216_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1218 = const()[name = tensor("op_1218"), val = tensor([20, -1, 64])]; + tensor var_1216_cast_fp16 = transpose(perm = var_1216_perm_0, x = var_1215_cast_fp16)[name = tensor("transpose_79")]; + tensor query_states_25_cast_fp16 = reshape(shape = var_1218, x = var_1216_cast_fp16)[name = tensor("query_states_25_cast_fp16")]; + tensor var_1220 = const()[name = tensor("op_1220"), val = tensor([20, -1, 64])]; + tensor var_1200_cast_fp16 = transpose(perm = var_1200_perm_0, x = var_1199_cast_fp16)[name = tensor("transpose_78")]; + tensor key_states_51_cast_fp16 = reshape(shape = var_1220, x = var_1200_cast_fp16)[name = tensor("key_states_51_cast_fp16")]; + tensor var_1222 = const()[name = tensor("op_1222"), val = tensor([20, -1, 64])]; + tensor var_1207_cast_fp16 = transpose(perm = var_1207_perm_0, x = var_1206_cast_fp16)[name = tensor("transpose_77")]; + tensor value_states_51_cast_fp16 = reshape(shape = var_1222, x = var_1207_cast_fp16)[name = tensor("value_states_51_cast_fp16")]; + tensor attn_weights_73_transpose_x_1 = const()[name = tensor("attn_weights_73_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_73_transpose_y_1 = const()[name = tensor("attn_weights_73_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_73_cast_fp16 = matmul(transpose_x = attn_weights_73_transpose_x_1, transpose_y = attn_weights_73_transpose_y_1, x = query_states_25_cast_fp16, y = key_states_51_cast_fp16)[name = tensor("attn_weights_73_cast_fp16")]; + tensor var_1227 = const()[name = tensor("op_1227"), val = tensor([1, 20, 77, 77])]; + tensor var_1228_cast_fp16 = reshape(shape = var_1227, x = attn_weights_73_cast_fp16)[name = tensor("op_1228_cast_fp16")]; + tensor attn_weights_75_cast_fp16 = add(x = var_1228_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_75_cast_fp16")]; + tensor var_1233 = const()[name = tensor("op_1233"), val = tensor([20, 77, 77])]; + tensor input_197_cast_fp16 = reshape(shape = var_1233, x = attn_weights_75_cast_fp16)[name = tensor("input_197_cast_fp16")]; + tensor input_199_cast_fp16 = softmax(axis = var_5, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor attn_output_73_transpose_x_0 = const()[name = tensor("attn_output_73_transpose_x_0"), val = tensor(false)]; + tensor attn_output_73_transpose_y_0 = const()[name = tensor("attn_output_73_transpose_y_0"), val = tensor(false)]; + tensor attn_output_73_cast_fp16 = matmul(transpose_x = attn_output_73_transpose_x_0, transpose_y = attn_output_73_transpose_y_0, x = input_199_cast_fp16, y = value_states_51_cast_fp16)[name = tensor("attn_output_73_cast_fp16")]; + tensor var_1238 = const()[name = tensor("op_1238"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_75_cast_fp16 = reshape(shape = var_1238, x = attn_output_73_cast_fp16)[name = tensor("attn_output_75_cast_fp16")]; + tensor attn_output_77_perm_0 = const()[name = tensor("attn_output_77_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1241 = const()[name = tensor("op_1241"), val = tensor([1, 77, 1280])]; + tensor attn_output_77_cast_fp16 = transpose(perm = attn_output_77_perm_0, x = attn_output_75_cast_fp16)[name = tensor("transpose_76")]; + tensor input_201_cast_fp16 = reshape(shape = var_1241, x = attn_output_77_cast_fp16)[name = tensor("input_201_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307636480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308865344))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308865536)))]; + tensor linear_75_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized, x = input_201_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_195_cast_fp16, y = linear_75_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor input_205_axes_0 = const()[name = tensor("input_205_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308868160)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308870784)))]; + tensor input_205_cast_fp16 = layer_norm(axes = input_205_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("input_205_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308873408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313788672))), name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313788864)))]; + tensor linear_76_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized, x = input_205_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor input_209_mode_0 = const()[name = tensor("input_209_mode_0"), val = tensor("EXACT")]; + tensor input_209_cast_fp16 = gelu(mode = input_209_mode_0, x = linear_76_cast_fp16)[name = tensor("input_209_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313799168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318714432))), name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318714624)))]; + tensor linear_77_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized, x = input_209_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor input_211_cast_fp16 = add(x = input_203_cast_fp16, y = linear_77_cast_fp16)[name = tensor("input_211_cast_fp16")]; + tensor hidden_states_79_axes_0 = const()[name = tensor("hidden_states_79_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318717248)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318719872)))]; + tensor hidden_states_79_cast_fp16 = layer_norm(axes = hidden_states_79_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("hidden_states_79_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318722496))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319951360))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319951552)))]; + tensor linear_78_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_78_cast_fp16")]; + tensor var_1280_to_fp16 = const()[name = tensor("op_1280_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_83_cast_fp16 = mul(x = linear_78_cast_fp16, y = var_1280_to_fp16)[name = tensor("tensor_83_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319954176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321183040))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321183232)))]; + tensor linear_79_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor var_1285 = const()[name = tensor("op_1285"), val = tensor([1, -1, 20, 64])]; + tensor var_1286_cast_fp16 = reshape(shape = var_1285, x = linear_79_cast_fp16)[name = tensor("op_1286_cast_fp16")]; + tensor var_1287_perm_0 = const()[name = tensor("op_1287_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321185856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322414720))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322414912)))]; + tensor linear_80_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor var_1292 = const()[name = tensor("op_1292"), val = tensor([1, -1, 20, 64])]; + tensor var_1293_cast_fp16 = reshape(shape = var_1292, x = linear_80_cast_fp16)[name = tensor("op_1293_cast_fp16")]; + tensor var_1294_perm_0 = const()[name = tensor("op_1294_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1301 = const()[name = tensor("op_1301"), val = tensor([1, 77, 20, 64])]; + tensor var_1302_cast_fp16 = reshape(shape = var_1301, x = tensor_83_cast_fp16)[name = tensor("op_1302_cast_fp16")]; + tensor var_1303_perm_0 = const()[name = tensor("op_1303_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([20, -1, 64])]; + tensor var_1303_cast_fp16 = transpose(perm = var_1303_perm_0, x = var_1302_cast_fp16)[name = tensor("transpose_75")]; + tensor query_states_27_cast_fp16 = reshape(shape = var_1305, x = var_1303_cast_fp16)[name = tensor("query_states_27_cast_fp16")]; + tensor var_1307 = const()[name = tensor("op_1307"), val = tensor([20, -1, 64])]; + tensor var_1287_cast_fp16 = transpose(perm = var_1287_perm_0, x = var_1286_cast_fp16)[name = tensor("transpose_74")]; + tensor key_states_55_cast_fp16 = reshape(shape = var_1307, x = var_1287_cast_fp16)[name = tensor("key_states_55_cast_fp16")]; + tensor var_1309 = const()[name = tensor("op_1309"), val = tensor([20, -1, 64])]; + tensor var_1294_cast_fp16 = transpose(perm = var_1294_perm_0, x = var_1293_cast_fp16)[name = tensor("transpose_73")]; + tensor value_states_55_cast_fp16 = reshape(shape = var_1309, x = var_1294_cast_fp16)[name = tensor("value_states_55_cast_fp16")]; + tensor attn_weights_79_transpose_x_1 = const()[name = tensor("attn_weights_79_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_79_transpose_y_1 = const()[name = tensor("attn_weights_79_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_79_cast_fp16 = matmul(transpose_x = attn_weights_79_transpose_x_1, transpose_y = attn_weights_79_transpose_y_1, x = query_states_27_cast_fp16, y = key_states_55_cast_fp16)[name = tensor("attn_weights_79_cast_fp16")]; + tensor var_1314 = const()[name = tensor("op_1314"), val = tensor([1, 20, 77, 77])]; + tensor var_1315_cast_fp16 = reshape(shape = var_1314, x = attn_weights_79_cast_fp16)[name = tensor("op_1315_cast_fp16")]; + tensor attn_weights_81_cast_fp16 = add(x = var_1315_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_81_cast_fp16")]; + tensor var_1320 = const()[name = tensor("op_1320"), val = tensor([20, 77, 77])]; + tensor input_213_cast_fp16 = reshape(shape = var_1320, x = attn_weights_81_cast_fp16)[name = tensor("input_213_cast_fp16")]; + tensor input_215_cast_fp16 = softmax(axis = var_5, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor attn_output_79_transpose_x_0 = const()[name = tensor("attn_output_79_transpose_x_0"), val = tensor(false)]; + tensor attn_output_79_transpose_y_0 = const()[name = tensor("attn_output_79_transpose_y_0"), val = tensor(false)]; + tensor attn_output_79_cast_fp16 = matmul(transpose_x = attn_output_79_transpose_x_0, transpose_y = attn_output_79_transpose_y_0, x = input_215_cast_fp16, y = value_states_55_cast_fp16)[name = tensor("attn_output_79_cast_fp16")]; + tensor var_1325 = const()[name = tensor("op_1325"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_81_cast_fp16 = reshape(shape = var_1325, x = attn_output_79_cast_fp16)[name = tensor("attn_output_81_cast_fp16")]; + tensor attn_output_83_perm_0 = const()[name = tensor("attn_output_83_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1328 = const()[name = tensor("op_1328"), val = tensor([1, 77, 1280])]; + tensor attn_output_83_cast_fp16 = transpose(perm = attn_output_83_perm_0, x = attn_output_81_cast_fp16)[name = tensor("transpose_72")]; + tensor input_217_cast_fp16 = reshape(shape = var_1328, x = attn_output_83_cast_fp16)[name = tensor("input_217_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322417536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323646400))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323646592)))]; + tensor linear_81_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized, x = input_217_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor input_219_cast_fp16 = add(x = input_211_cast_fp16, y = linear_81_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323649216)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323651840)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323654464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328569728))), name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328569920)))]; + tensor linear_82_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized, x = input_221_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_225_mode_0 = const()[name = tensor("input_225_mode_0"), val = tensor("EXACT")]; + tensor input_225_cast_fp16 = gelu(mode = input_225_mode_0, x = linear_82_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328580224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333495488))), name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333495680)))]; + tensor linear_83_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized, x = input_225_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor input_227_cast_fp16 = add(x = input_219_cast_fp16, y = linear_83_cast_fp16)[name = tensor("input_227_cast_fp16")]; + tensor hidden_states_85_axes_0 = const()[name = tensor("hidden_states_85_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333498304)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333500928)))]; + tensor hidden_states_85_cast_fp16 = layer_norm(axes = hidden_states_85_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("hidden_states_85_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333503552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334732416))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334732608)))]; + tensor linear_84_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_1367_to_fp16 = const()[name = tensor("op_1367_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_89_cast_fp16 = mul(x = linear_84_cast_fp16, y = var_1367_to_fp16)[name = tensor("tensor_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334735232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335964096))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335964288)))]; + tensor linear_85_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_1372 = const()[name = tensor("op_1372"), val = tensor([1, -1, 20, 64])]; + tensor var_1373_cast_fp16 = reshape(shape = var_1372, x = linear_85_cast_fp16)[name = tensor("op_1373_cast_fp16")]; + tensor var_1374_perm_0 = const()[name = tensor("op_1374_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335966912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337195776))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337195968)))]; + tensor linear_86_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_1379 = const()[name = tensor("op_1379"), val = tensor([1, -1, 20, 64])]; + tensor var_1380_cast_fp16 = reshape(shape = var_1379, x = linear_86_cast_fp16)[name = tensor("op_1380_cast_fp16")]; + tensor var_1381_perm_0 = const()[name = tensor("op_1381_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1388 = const()[name = tensor("op_1388"), val = tensor([1, 77, 20, 64])]; + tensor var_1389_cast_fp16 = reshape(shape = var_1388, x = tensor_89_cast_fp16)[name = tensor("op_1389_cast_fp16")]; + tensor var_1390_perm_0 = const()[name = tensor("op_1390_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1392 = const()[name = tensor("op_1392"), val = tensor([20, -1, 64])]; + tensor var_1390_cast_fp16 = transpose(perm = var_1390_perm_0, x = var_1389_cast_fp16)[name = tensor("transpose_71")]; + tensor query_states_29_cast_fp16 = reshape(shape = var_1392, x = var_1390_cast_fp16)[name = tensor("query_states_29_cast_fp16")]; + tensor var_1394 = const()[name = tensor("op_1394"), val = tensor([20, -1, 64])]; + tensor var_1374_cast_fp16 = transpose(perm = var_1374_perm_0, x = var_1373_cast_fp16)[name = tensor("transpose_70")]; + tensor key_states_59_cast_fp16 = reshape(shape = var_1394, x = var_1374_cast_fp16)[name = tensor("key_states_59_cast_fp16")]; + tensor var_1396 = const()[name = tensor("op_1396"), val = tensor([20, -1, 64])]; + tensor var_1381_cast_fp16 = transpose(perm = var_1381_perm_0, x = var_1380_cast_fp16)[name = tensor("transpose_69")]; + tensor value_states_59_cast_fp16 = reshape(shape = var_1396, x = var_1381_cast_fp16)[name = tensor("value_states_59_cast_fp16")]; + tensor attn_weights_85_transpose_x_1 = const()[name = tensor("attn_weights_85_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_85_transpose_y_1 = const()[name = tensor("attn_weights_85_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_85_cast_fp16 = matmul(transpose_x = attn_weights_85_transpose_x_1, transpose_y = attn_weights_85_transpose_y_1, x = query_states_29_cast_fp16, y = key_states_59_cast_fp16)[name = tensor("attn_weights_85_cast_fp16")]; + tensor var_1401 = const()[name = tensor("op_1401"), val = tensor([1, 20, 77, 77])]; + tensor var_1402_cast_fp16 = reshape(shape = var_1401, x = attn_weights_85_cast_fp16)[name = tensor("op_1402_cast_fp16")]; + tensor attn_weights_87_cast_fp16 = add(x = var_1402_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_87_cast_fp16")]; + tensor var_1407 = const()[name = tensor("op_1407"), val = tensor([20, 77, 77])]; + tensor input_229_cast_fp16 = reshape(shape = var_1407, x = attn_weights_87_cast_fp16)[name = tensor("input_229_cast_fp16")]; + tensor input_231_cast_fp16 = softmax(axis = var_5, x = input_229_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor attn_output_85_transpose_x_0 = const()[name = tensor("attn_output_85_transpose_x_0"), val = tensor(false)]; + tensor attn_output_85_transpose_y_0 = const()[name = tensor("attn_output_85_transpose_y_0"), val = tensor(false)]; + tensor attn_output_85_cast_fp16 = matmul(transpose_x = attn_output_85_transpose_x_0, transpose_y = attn_output_85_transpose_y_0, x = input_231_cast_fp16, y = value_states_59_cast_fp16)[name = tensor("attn_output_85_cast_fp16")]; + tensor var_1412 = const()[name = tensor("op_1412"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_87_cast_fp16 = reshape(shape = var_1412, x = attn_output_85_cast_fp16)[name = tensor("attn_output_87_cast_fp16")]; + tensor attn_output_89_perm_0 = const()[name = tensor("attn_output_89_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1415 = const()[name = tensor("op_1415"), val = tensor([1, 77, 1280])]; + tensor attn_output_89_cast_fp16 = transpose(perm = attn_output_89_perm_0, x = attn_output_87_cast_fp16)[name = tensor("transpose_68")]; + tensor input_233_cast_fp16 = reshape(shape = var_1415, x = attn_output_89_cast_fp16)[name = tensor("input_233_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337198592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338427456))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338427648)))]; + tensor linear_87_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized, x = input_233_cast_fp16)[name = tensor("linear_87_cast_fp16")]; + tensor input_235_cast_fp16 = add(x = input_227_cast_fp16, y = linear_87_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor input_237_axes_0 = const()[name = tensor("input_237_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338430272)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338432896)))]; + tensor input_237_cast_fp16 = layer_norm(axes = input_237_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("input_237_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338435520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343350784))), name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343350976)))]; + tensor linear_88_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized, x = input_237_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_241_mode_0 = const()[name = tensor("input_241_mode_0"), val = tensor("EXACT")]; + tensor input_241_cast_fp16 = gelu(mode = input_241_mode_0, x = linear_88_cast_fp16)[name = tensor("input_241_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343361280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348276544))), name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348276736)))]; + tensor linear_89_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized, x = input_241_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_243_cast_fp16 = add(x = input_235_cast_fp16, y = linear_89_cast_fp16)[name = tensor("input_243_cast_fp16")]; + tensor hidden_states_91_axes_0 = const()[name = tensor("hidden_states_91_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348279360)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348281984)))]; + tensor hidden_states_91_cast_fp16 = layer_norm(axes = hidden_states_91_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("hidden_states_91_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348284608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349513472))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349513664)))]; + tensor linear_90_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_1454_to_fp16 = const()[name = tensor("op_1454_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_95_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1454_to_fp16)[name = tensor("tensor_95_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349516288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350745152))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350745344)))]; + tensor linear_91_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor var_1459 = const()[name = tensor("op_1459"), val = tensor([1, -1, 20, 64])]; + tensor var_1460_cast_fp16 = reshape(shape = var_1459, x = linear_91_cast_fp16)[name = tensor("op_1460_cast_fp16")]; + tensor var_1461_perm_0 = const()[name = tensor("op_1461_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350747968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351976832))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351977024)))]; + tensor linear_92_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_1466 = const()[name = tensor("op_1466"), val = tensor([1, -1, 20, 64])]; + tensor var_1467_cast_fp16 = reshape(shape = var_1466, x = linear_92_cast_fp16)[name = tensor("op_1467_cast_fp16")]; + tensor var_1468_perm_0 = const()[name = tensor("op_1468_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1475 = const()[name = tensor("op_1475"), val = tensor([1, 77, 20, 64])]; + tensor var_1476_cast_fp16 = reshape(shape = var_1475, x = tensor_95_cast_fp16)[name = tensor("op_1476_cast_fp16")]; + tensor var_1477_perm_0 = const()[name = tensor("op_1477_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1479 = const()[name = tensor("op_1479"), val = tensor([20, -1, 64])]; + tensor var_1477_cast_fp16 = transpose(perm = var_1477_perm_0, x = var_1476_cast_fp16)[name = tensor("transpose_67")]; + tensor query_states_31_cast_fp16 = reshape(shape = var_1479, x = var_1477_cast_fp16)[name = tensor("query_states_31_cast_fp16")]; + tensor var_1481 = const()[name = tensor("op_1481"), val = tensor([20, -1, 64])]; + tensor var_1461_cast_fp16 = transpose(perm = var_1461_perm_0, x = var_1460_cast_fp16)[name = tensor("transpose_66")]; + tensor key_states_63_cast_fp16 = reshape(shape = var_1481, x = var_1461_cast_fp16)[name = tensor("key_states_63_cast_fp16")]; + tensor var_1483 = const()[name = tensor("op_1483"), val = tensor([20, -1, 64])]; + tensor var_1468_cast_fp16 = transpose(perm = var_1468_perm_0, x = var_1467_cast_fp16)[name = tensor("transpose_65")]; + tensor value_states_63_cast_fp16 = reshape(shape = var_1483, x = var_1468_cast_fp16)[name = tensor("value_states_63_cast_fp16")]; + tensor attn_weights_91_transpose_x_1 = const()[name = tensor("attn_weights_91_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_91_transpose_y_1 = const()[name = tensor("attn_weights_91_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_91_cast_fp16 = matmul(transpose_x = attn_weights_91_transpose_x_1, transpose_y = attn_weights_91_transpose_y_1, x = query_states_31_cast_fp16, y = key_states_63_cast_fp16)[name = tensor("attn_weights_91_cast_fp16")]; + tensor var_1488 = const()[name = tensor("op_1488"), val = tensor([1, 20, 77, 77])]; + tensor var_1489_cast_fp16 = reshape(shape = var_1488, x = attn_weights_91_cast_fp16)[name = tensor("op_1489_cast_fp16")]; + tensor attn_weights_93_cast_fp16 = add(x = var_1489_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_93_cast_fp16")]; + tensor var_1494 = const()[name = tensor("op_1494"), val = tensor([20, 77, 77])]; + tensor input_245_cast_fp16 = reshape(shape = var_1494, x = attn_weights_93_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor input_247_cast_fp16 = softmax(axis = var_5, x = input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; + tensor attn_output_91_transpose_x_0 = const()[name = tensor("attn_output_91_transpose_x_0"), val = tensor(false)]; + tensor attn_output_91_transpose_y_0 = const()[name = tensor("attn_output_91_transpose_y_0"), val = tensor(false)]; + tensor attn_output_91_cast_fp16 = matmul(transpose_x = attn_output_91_transpose_x_0, transpose_y = attn_output_91_transpose_y_0, x = input_247_cast_fp16, y = value_states_63_cast_fp16)[name = tensor("attn_output_91_cast_fp16")]; + tensor var_1499 = const()[name = tensor("op_1499"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_93_cast_fp16 = reshape(shape = var_1499, x = attn_output_91_cast_fp16)[name = tensor("attn_output_93_cast_fp16")]; + tensor attn_output_95_perm_0 = const()[name = tensor("attn_output_95_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1502 = const()[name = tensor("op_1502"), val = tensor([1, 77, 1280])]; + tensor attn_output_95_cast_fp16 = transpose(perm = attn_output_95_perm_0, x = attn_output_93_cast_fp16)[name = tensor("transpose_64")]; + tensor input_249_cast_fp16 = reshape(shape = var_1502, x = attn_output_95_cast_fp16)[name = tensor("input_249_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351979648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353208512))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353208704)))]; + tensor linear_93_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized, x = input_249_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor input_251_cast_fp16 = add(x = input_243_cast_fp16, y = linear_93_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor input_253_axes_0 = const()[name = tensor("input_253_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353211328)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353213952)))]; + tensor input_253_cast_fp16 = layer_norm(axes = input_253_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("input_253_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353216576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358131840))), name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358132032)))]; + tensor linear_94_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized, x = input_253_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor input_257_mode_0 = const()[name = tensor("input_257_mode_0"), val = tensor("EXACT")]; + tensor input_257_cast_fp16 = gelu(mode = input_257_mode_0, x = linear_94_cast_fp16)[name = tensor("input_257_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358142336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363057600))), name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363057792)))]; + tensor linear_95_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized, x = input_257_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor input_259_cast_fp16 = add(x = input_251_cast_fp16, y = linear_95_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor hidden_states_97_axes_0 = const()[name = tensor("hidden_states_97_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363060416)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363063040)))]; + tensor hidden_states_97_cast_fp16 = layer_norm(axes = hidden_states_97_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16, x = input_259_cast_fp16)[name = tensor("hidden_states_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(363065664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364294528))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364294720)))]; + tensor linear_96_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_96_cast_fp16")]; + tensor var_1541_to_fp16 = const()[name = tensor("op_1541_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_101_cast_fp16 = mul(x = linear_96_cast_fp16, y = var_1541_to_fp16)[name = tensor("tensor_101_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364297344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365526208))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365526400)))]; + tensor linear_97_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor var_1546 = const()[name = tensor("op_1546"), val = tensor([1, -1, 20, 64])]; + tensor var_1547_cast_fp16 = reshape(shape = var_1546, x = linear_97_cast_fp16)[name = tensor("op_1547_cast_fp16")]; + tensor var_1548_perm_0 = const()[name = tensor("op_1548_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365529024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366757888))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366758080)))]; + tensor linear_98_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor var_1553 = const()[name = tensor("op_1553"), val = tensor([1, -1, 20, 64])]; + tensor var_1554_cast_fp16 = reshape(shape = var_1553, x = linear_98_cast_fp16)[name = tensor("op_1554_cast_fp16")]; + tensor var_1555_perm_0 = const()[name = tensor("op_1555_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1562 = const()[name = tensor("op_1562"), val = tensor([1, 77, 20, 64])]; + tensor var_1563_cast_fp16 = reshape(shape = var_1562, x = tensor_101_cast_fp16)[name = tensor("op_1563_cast_fp16")]; + tensor var_1564_perm_0 = const()[name = tensor("op_1564_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1566 = const()[name = tensor("op_1566"), val = tensor([20, -1, 64])]; + tensor var_1564_cast_fp16 = transpose(perm = var_1564_perm_0, x = var_1563_cast_fp16)[name = tensor("transpose_63")]; + tensor query_states_33_cast_fp16 = reshape(shape = var_1566, x = var_1564_cast_fp16)[name = tensor("query_states_33_cast_fp16")]; + tensor var_1568 = const()[name = tensor("op_1568"), val = tensor([20, -1, 64])]; + tensor var_1548_cast_fp16 = transpose(perm = var_1548_perm_0, x = var_1547_cast_fp16)[name = tensor("transpose_62")]; + tensor key_states_67_cast_fp16 = reshape(shape = var_1568, x = var_1548_cast_fp16)[name = tensor("key_states_67_cast_fp16")]; + tensor var_1570 = const()[name = tensor("op_1570"), val = tensor([20, -1, 64])]; + tensor var_1555_cast_fp16 = transpose(perm = var_1555_perm_0, x = var_1554_cast_fp16)[name = tensor("transpose_61")]; + tensor value_states_67_cast_fp16 = reshape(shape = var_1570, x = var_1555_cast_fp16)[name = tensor("value_states_67_cast_fp16")]; + tensor attn_weights_97_transpose_x_1 = const()[name = tensor("attn_weights_97_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_97_transpose_y_1 = const()[name = tensor("attn_weights_97_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_97_cast_fp16 = matmul(transpose_x = attn_weights_97_transpose_x_1, transpose_y = attn_weights_97_transpose_y_1, x = query_states_33_cast_fp16, y = key_states_67_cast_fp16)[name = tensor("attn_weights_97_cast_fp16")]; + tensor var_1575 = const()[name = tensor("op_1575"), val = tensor([1, 20, 77, 77])]; + tensor var_1576_cast_fp16 = reshape(shape = var_1575, x = attn_weights_97_cast_fp16)[name = tensor("op_1576_cast_fp16")]; + tensor attn_weights_99_cast_fp16 = add(x = var_1576_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_99_cast_fp16")]; + tensor var_1581 = const()[name = tensor("op_1581"), val = tensor([20, 77, 77])]; + tensor input_261_cast_fp16 = reshape(shape = var_1581, x = attn_weights_99_cast_fp16)[name = tensor("input_261_cast_fp16")]; + tensor input_263_cast_fp16 = softmax(axis = var_5, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; + tensor attn_output_97_transpose_x_0 = const()[name = tensor("attn_output_97_transpose_x_0"), val = tensor(false)]; + tensor attn_output_97_transpose_y_0 = const()[name = tensor("attn_output_97_transpose_y_0"), val = tensor(false)]; + tensor attn_output_97_cast_fp16 = matmul(transpose_x = attn_output_97_transpose_x_0, transpose_y = attn_output_97_transpose_y_0, x = input_263_cast_fp16, y = value_states_67_cast_fp16)[name = tensor("attn_output_97_cast_fp16")]; + tensor var_1586 = const()[name = tensor("op_1586"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_99_cast_fp16 = reshape(shape = var_1586, x = attn_output_97_cast_fp16)[name = tensor("attn_output_99_cast_fp16")]; + tensor attn_output_101_perm_0 = const()[name = tensor("attn_output_101_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1589 = const()[name = tensor("op_1589"), val = tensor([1, 77, 1280])]; + tensor attn_output_101_cast_fp16 = transpose(perm = attn_output_101_perm_0, x = attn_output_99_cast_fp16)[name = tensor("transpose_60")]; + tensor input_265_cast_fp16 = reshape(shape = var_1589, x = attn_output_101_cast_fp16)[name = tensor("input_265_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366760704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367989568))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367989760)))]; + tensor linear_99_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized, x = input_265_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor input_267_cast_fp16 = add(x = input_259_cast_fp16, y = linear_99_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor input_269_axes_0 = const()[name = tensor("input_269_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367992384)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367995008)))]; + tensor input_269_cast_fp16 = layer_norm(axes = input_269_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367997632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372912896))), name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372913088)))]; + tensor linear_100_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized, x = input_269_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_273_mode_0 = const()[name = tensor("input_273_mode_0"), val = tensor("EXACT")]; + tensor input_273_cast_fp16 = gelu(mode = input_273_mode_0, x = linear_100_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372923392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377838656))), name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377838848)))]; + tensor linear_101_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized, x = input_273_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor input_275_cast_fp16 = add(x = input_267_cast_fp16, y = linear_101_cast_fp16)[name = tensor("input_275_cast_fp16")]; + tensor hidden_states_103_axes_0 = const()[name = tensor("hidden_states_103_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377841472)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377844096)))]; + tensor hidden_states_103_cast_fp16 = layer_norm(axes = hidden_states_103_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("hidden_states_103_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377846720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379075584))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379075776)))]; + tensor linear_102_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_1628_to_fp16 = const()[name = tensor("op_1628_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_107_cast_fp16 = mul(x = linear_102_cast_fp16, y = var_1628_to_fp16)[name = tensor("tensor_107_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379078400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380307264))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380307456)))]; + tensor linear_103_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_1633 = const()[name = tensor("op_1633"), val = tensor([1, -1, 20, 64])]; + tensor var_1634_cast_fp16 = reshape(shape = var_1633, x = linear_103_cast_fp16)[name = tensor("op_1634_cast_fp16")]; + tensor var_1635_perm_0 = const()[name = tensor("op_1635_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380310080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381538944))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381539136)))]; + tensor linear_104_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_1640 = const()[name = tensor("op_1640"), val = tensor([1, -1, 20, 64])]; + tensor var_1641_cast_fp16 = reshape(shape = var_1640, x = linear_104_cast_fp16)[name = tensor("op_1641_cast_fp16")]; + tensor var_1642_perm_0 = const()[name = tensor("op_1642_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1649 = const()[name = tensor("op_1649"), val = tensor([1, 77, 20, 64])]; + tensor var_1650_cast_fp16 = reshape(shape = var_1649, x = tensor_107_cast_fp16)[name = tensor("op_1650_cast_fp16")]; + tensor var_1651_perm_0 = const()[name = tensor("op_1651_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1653 = const()[name = tensor("op_1653"), val = tensor([20, -1, 64])]; + tensor var_1651_cast_fp16 = transpose(perm = var_1651_perm_0, x = var_1650_cast_fp16)[name = tensor("transpose_59")]; + tensor query_states_35_cast_fp16 = reshape(shape = var_1653, x = var_1651_cast_fp16)[name = tensor("query_states_35_cast_fp16")]; + tensor var_1655 = const()[name = tensor("op_1655"), val = tensor([20, -1, 64])]; + tensor var_1635_cast_fp16 = transpose(perm = var_1635_perm_0, x = var_1634_cast_fp16)[name = tensor("transpose_58")]; + tensor key_states_71_cast_fp16 = reshape(shape = var_1655, x = var_1635_cast_fp16)[name = tensor("key_states_71_cast_fp16")]; + tensor var_1657 = const()[name = tensor("op_1657"), val = tensor([20, -1, 64])]; + tensor var_1642_cast_fp16 = transpose(perm = var_1642_perm_0, x = var_1641_cast_fp16)[name = tensor("transpose_57")]; + tensor value_states_71_cast_fp16 = reshape(shape = var_1657, x = var_1642_cast_fp16)[name = tensor("value_states_71_cast_fp16")]; + tensor attn_weights_103_transpose_x_1 = const()[name = tensor("attn_weights_103_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_103_transpose_y_1 = const()[name = tensor("attn_weights_103_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_103_cast_fp16 = matmul(transpose_x = attn_weights_103_transpose_x_1, transpose_y = attn_weights_103_transpose_y_1, x = query_states_35_cast_fp16, y = key_states_71_cast_fp16)[name = tensor("attn_weights_103_cast_fp16")]; + tensor var_1662 = const()[name = tensor("op_1662"), val = tensor([1, 20, 77, 77])]; + tensor var_1663_cast_fp16 = reshape(shape = var_1662, x = attn_weights_103_cast_fp16)[name = tensor("op_1663_cast_fp16")]; + tensor attn_weights_105_cast_fp16 = add(x = var_1663_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_105_cast_fp16")]; + tensor var_1668 = const()[name = tensor("op_1668"), val = tensor([20, 77, 77])]; + tensor input_277_cast_fp16 = reshape(shape = var_1668, x = attn_weights_105_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor input_279_cast_fp16 = softmax(axis = var_5, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; + tensor attn_output_103_transpose_x_0 = const()[name = tensor("attn_output_103_transpose_x_0"), val = tensor(false)]; + tensor attn_output_103_transpose_y_0 = const()[name = tensor("attn_output_103_transpose_y_0"), val = tensor(false)]; + tensor attn_output_103_cast_fp16 = matmul(transpose_x = attn_output_103_transpose_x_0, transpose_y = attn_output_103_transpose_y_0, x = input_279_cast_fp16, y = value_states_71_cast_fp16)[name = tensor("attn_output_103_cast_fp16")]; + tensor var_1673 = const()[name = tensor("op_1673"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_105_cast_fp16 = reshape(shape = var_1673, x = attn_output_103_cast_fp16)[name = tensor("attn_output_105_cast_fp16")]; + tensor attn_output_107_perm_0 = const()[name = tensor("attn_output_107_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1676 = const()[name = tensor("op_1676"), val = tensor([1, 77, 1280])]; + tensor attn_output_107_cast_fp16 = transpose(perm = attn_output_107_perm_0, x = attn_output_105_cast_fp16)[name = tensor("transpose_56")]; + tensor input_281_cast_fp16 = reshape(shape = var_1676, x = attn_output_107_cast_fp16)[name = tensor("input_281_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381541760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382770624))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382770816)))]; + tensor linear_105_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized, x = input_281_cast_fp16)[name = tensor("linear_105_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_275_cast_fp16, y = linear_105_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382773440)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382776064)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382778688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387693952))), name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387694144)))]; + tensor linear_106_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized, x = input_285_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_289_mode_0 = const()[name = tensor("input_289_mode_0"), val = tensor("EXACT")]; + tensor input_289_cast_fp16 = gelu(mode = input_289_mode_0, x = linear_106_cast_fp16)[name = tensor("input_289_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387704448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392619712))), name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392619904)))]; + tensor linear_107_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized, x = input_289_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_291_cast_fp16 = add(x = input_283_cast_fp16, y = linear_107_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor hidden_states_109_axes_0 = const()[name = tensor("hidden_states_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392622528)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392625152)))]; + tensor hidden_states_109_cast_fp16 = layer_norm(axes = hidden_states_109_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("hidden_states_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392627776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393856640))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393856832)))]; + tensor linear_108_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_1715_to_fp16 = const()[name = tensor("op_1715_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_113_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_1715_to_fp16)[name = tensor("tensor_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393859456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(395088320))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(395088512)))]; + tensor linear_109_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([1, -1, 20, 64])]; + tensor var_1721_cast_fp16 = reshape(shape = var_1720, x = linear_109_cast_fp16)[name = tensor("op_1721_cast_fp16")]; + tensor var_1722_perm_0 = const()[name = tensor("op_1722_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(395091136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396320000))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396320192)))]; + tensor linear_110_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_1727 = const()[name = tensor("op_1727"), val = tensor([1, -1, 20, 64])]; + tensor var_1728_cast_fp16 = reshape(shape = var_1727, x = linear_110_cast_fp16)[name = tensor("op_1728_cast_fp16")]; + tensor var_1729_perm_0 = const()[name = tensor("op_1729_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1736 = const()[name = tensor("op_1736"), val = tensor([1, 77, 20, 64])]; + tensor var_1737_cast_fp16 = reshape(shape = var_1736, x = tensor_113_cast_fp16)[name = tensor("op_1737_cast_fp16")]; + tensor var_1738_perm_0 = const()[name = tensor("op_1738_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1740 = const()[name = tensor("op_1740"), val = tensor([20, -1, 64])]; + tensor var_1738_cast_fp16 = transpose(perm = var_1738_perm_0, x = var_1737_cast_fp16)[name = tensor("transpose_55")]; + tensor query_states_37_cast_fp16 = reshape(shape = var_1740, x = var_1738_cast_fp16)[name = tensor("query_states_37_cast_fp16")]; + tensor var_1742 = const()[name = tensor("op_1742"), val = tensor([20, -1, 64])]; + tensor var_1722_cast_fp16 = transpose(perm = var_1722_perm_0, x = var_1721_cast_fp16)[name = tensor("transpose_54")]; + tensor key_states_75_cast_fp16 = reshape(shape = var_1742, x = var_1722_cast_fp16)[name = tensor("key_states_75_cast_fp16")]; + tensor var_1744 = const()[name = tensor("op_1744"), val = tensor([20, -1, 64])]; + tensor var_1729_cast_fp16 = transpose(perm = var_1729_perm_0, x = var_1728_cast_fp16)[name = tensor("transpose_53")]; + tensor value_states_75_cast_fp16 = reshape(shape = var_1744, x = var_1729_cast_fp16)[name = tensor("value_states_75_cast_fp16")]; + tensor attn_weights_109_transpose_x_1 = const()[name = tensor("attn_weights_109_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_109_transpose_y_1 = const()[name = tensor("attn_weights_109_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_109_cast_fp16 = matmul(transpose_x = attn_weights_109_transpose_x_1, transpose_y = attn_weights_109_transpose_y_1, x = query_states_37_cast_fp16, y = key_states_75_cast_fp16)[name = tensor("attn_weights_109_cast_fp16")]; + tensor var_1749 = const()[name = tensor("op_1749"), val = tensor([1, 20, 77, 77])]; + tensor var_1750_cast_fp16 = reshape(shape = var_1749, x = attn_weights_109_cast_fp16)[name = tensor("op_1750_cast_fp16")]; + tensor attn_weights_111_cast_fp16 = add(x = var_1750_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_111_cast_fp16")]; + tensor var_1755 = const()[name = tensor("op_1755"), val = tensor([20, 77, 77])]; + tensor input_293_cast_fp16 = reshape(shape = var_1755, x = attn_weights_111_cast_fp16)[name = tensor("input_293_cast_fp16")]; + tensor input_295_cast_fp16 = softmax(axis = var_5, x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; + tensor attn_output_109_transpose_x_0 = const()[name = tensor("attn_output_109_transpose_x_0"), val = tensor(false)]; + tensor attn_output_109_transpose_y_0 = const()[name = tensor("attn_output_109_transpose_y_0"), val = tensor(false)]; + tensor attn_output_109_cast_fp16 = matmul(transpose_x = attn_output_109_transpose_x_0, transpose_y = attn_output_109_transpose_y_0, x = input_295_cast_fp16, y = value_states_75_cast_fp16)[name = tensor("attn_output_109_cast_fp16")]; + tensor var_1760 = const()[name = tensor("op_1760"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_111_cast_fp16 = reshape(shape = var_1760, x = attn_output_109_cast_fp16)[name = tensor("attn_output_111_cast_fp16")]; + tensor attn_output_113_perm_0 = const()[name = tensor("attn_output_113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1763 = const()[name = tensor("op_1763"), val = tensor([1, 77, 1280])]; + tensor attn_output_113_cast_fp16 = transpose(perm = attn_output_113_perm_0, x = attn_output_111_cast_fp16)[name = tensor("transpose_52")]; + tensor input_297_cast_fp16 = reshape(shape = var_1763, x = attn_output_113_cast_fp16)[name = tensor("input_297_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396322816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397551680))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397551872)))]; + tensor linear_111_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized, x = input_297_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_111_cast_fp16)[name = tensor("input_299_cast_fp16")]; + tensor input_301_axes_0 = const()[name = tensor("input_301_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397554496)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397557120)))]; + tensor input_301_cast_fp16 = layer_norm(axes = input_301_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397559744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402475008))), name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402475200)))]; + tensor linear_112_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized, x = input_301_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor input_305_mode_0 = const()[name = tensor("input_305_mode_0"), val = tensor("EXACT")]; + tensor input_305_cast_fp16 = gelu(mode = input_305_mode_0, x = linear_112_cast_fp16)[name = tensor("input_305_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402485504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407400768))), name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407400960)))]; + tensor linear_113_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized, x = input_305_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_299_cast_fp16, y = linear_113_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor hidden_states_115_axes_0 = const()[name = tensor("hidden_states_115_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407403584)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407406208)))]; + tensor hidden_states_115_cast_fp16 = layer_norm(axes = hidden_states_115_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("hidden_states_115_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407408832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408637696))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408637888)))]; + tensor linear_114_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_114_cast_fp16")]; + tensor var_1802_to_fp16 = const()[name = tensor("op_1802_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_119_cast_fp16 = mul(x = linear_114_cast_fp16, y = var_1802_to_fp16)[name = tensor("tensor_119_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408640512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409869376))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409869568)))]; + tensor linear_115_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor var_1807 = const()[name = tensor("op_1807"), val = tensor([1, -1, 20, 64])]; + tensor var_1808_cast_fp16 = reshape(shape = var_1807, x = linear_115_cast_fp16)[name = tensor("op_1808_cast_fp16")]; + tensor var_1809_perm_0 = const()[name = tensor("op_1809_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409872192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411101056))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411101248)))]; + tensor linear_116_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor var_1814 = const()[name = tensor("op_1814"), val = tensor([1, -1, 20, 64])]; + tensor var_1815_cast_fp16 = reshape(shape = var_1814, x = linear_116_cast_fp16)[name = tensor("op_1815_cast_fp16")]; + tensor var_1816_perm_0 = const()[name = tensor("op_1816_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1823 = const()[name = tensor("op_1823"), val = tensor([1, 77, 20, 64])]; + tensor var_1824_cast_fp16 = reshape(shape = var_1823, x = tensor_119_cast_fp16)[name = tensor("op_1824_cast_fp16")]; + tensor var_1825_perm_0 = const()[name = tensor("op_1825_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1827 = const()[name = tensor("op_1827"), val = tensor([20, -1, 64])]; + tensor var_1825_cast_fp16 = transpose(perm = var_1825_perm_0, x = var_1824_cast_fp16)[name = tensor("transpose_51")]; + tensor query_states_39_cast_fp16 = reshape(shape = var_1827, x = var_1825_cast_fp16)[name = tensor("query_states_39_cast_fp16")]; + tensor var_1829 = const()[name = tensor("op_1829"), val = tensor([20, -1, 64])]; + tensor var_1809_cast_fp16 = transpose(perm = var_1809_perm_0, x = var_1808_cast_fp16)[name = tensor("transpose_50")]; + tensor key_states_79_cast_fp16 = reshape(shape = var_1829, x = var_1809_cast_fp16)[name = tensor("key_states_79_cast_fp16")]; + tensor var_1831 = const()[name = tensor("op_1831"), val = tensor([20, -1, 64])]; + tensor var_1816_cast_fp16 = transpose(perm = var_1816_perm_0, x = var_1815_cast_fp16)[name = tensor("transpose_49")]; + tensor value_states_79_cast_fp16 = reshape(shape = var_1831, x = var_1816_cast_fp16)[name = tensor("value_states_79_cast_fp16")]; + tensor attn_weights_115_transpose_x_1 = const()[name = tensor("attn_weights_115_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_115_transpose_y_1 = const()[name = tensor("attn_weights_115_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_115_cast_fp16 = matmul(transpose_x = attn_weights_115_transpose_x_1, transpose_y = attn_weights_115_transpose_y_1, x = query_states_39_cast_fp16, y = key_states_79_cast_fp16)[name = tensor("attn_weights_115_cast_fp16")]; + tensor var_1836 = const()[name = tensor("op_1836"), val = tensor([1, 20, 77, 77])]; + tensor var_1837_cast_fp16 = reshape(shape = var_1836, x = attn_weights_115_cast_fp16)[name = tensor("op_1837_cast_fp16")]; + tensor attn_weights_117_cast_fp16 = add(x = var_1837_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_117_cast_fp16")]; + tensor var_1842 = const()[name = tensor("op_1842"), val = tensor([20, 77, 77])]; + tensor input_309_cast_fp16 = reshape(shape = var_1842, x = attn_weights_117_cast_fp16)[name = tensor("input_309_cast_fp16")]; + tensor input_311_cast_fp16 = softmax(axis = var_5, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor attn_output_115_transpose_x_0 = const()[name = tensor("attn_output_115_transpose_x_0"), val = tensor(false)]; + tensor attn_output_115_transpose_y_0 = const()[name = tensor("attn_output_115_transpose_y_0"), val = tensor(false)]; + tensor attn_output_115_cast_fp16 = matmul(transpose_x = attn_output_115_transpose_x_0, transpose_y = attn_output_115_transpose_y_0, x = input_311_cast_fp16, y = value_states_79_cast_fp16)[name = tensor("attn_output_115_cast_fp16")]; + tensor var_1847 = const()[name = tensor("op_1847"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_117_cast_fp16 = reshape(shape = var_1847, x = attn_output_115_cast_fp16)[name = tensor("attn_output_117_cast_fp16")]; + tensor attn_output_119_perm_0 = const()[name = tensor("attn_output_119_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1850 = const()[name = tensor("op_1850"), val = tensor([1, 77, 1280])]; + tensor attn_output_119_cast_fp16 = transpose(perm = attn_output_119_perm_0, x = attn_output_117_cast_fp16)[name = tensor("transpose_48")]; + tensor input_313_cast_fp16 = reshape(shape = var_1850, x = attn_output_119_cast_fp16)[name = tensor("input_313_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(411103872))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412332736))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412332928)))]; + tensor linear_117_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized, x = input_313_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor input_315_cast_fp16 = add(x = input_307_cast_fp16, y = linear_117_cast_fp16)[name = tensor("input_315_cast_fp16")]; + tensor input_317_axes_0 = const()[name = tensor("input_317_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412335552)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412338176)))]; + tensor input_317_cast_fp16 = layer_norm(axes = input_317_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412340800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417256064))), name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417256256)))]; + tensor linear_118_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized, x = input_317_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_321_mode_0 = const()[name = tensor("input_321_mode_0"), val = tensor("EXACT")]; + tensor input_321_cast_fp16 = gelu(mode = input_321_mode_0, x = linear_118_cast_fp16)[name = tensor("input_321_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417266560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422181824))), name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422182016)))]; + tensor linear_119_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized, x = input_321_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor input_323_cast_fp16 = add(x = input_315_cast_fp16, y = linear_119_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor hidden_states_121_axes_0 = const()[name = tensor("hidden_states_121_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422184640)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422187264)))]; + tensor hidden_states_121_cast_fp16 = layer_norm(axes = hidden_states_121_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("hidden_states_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422189888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423418752))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423418944)))]; + tensor linear_120_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_1889_to_fp16 = const()[name = tensor("op_1889_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_125_cast_fp16 = mul(x = linear_120_cast_fp16, y = var_1889_to_fp16)[name = tensor("tensor_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423421568))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424650432))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424650624)))]; + tensor linear_121_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_1894 = const()[name = tensor("op_1894"), val = tensor([1, -1, 20, 64])]; + tensor var_1895_cast_fp16 = reshape(shape = var_1894, x = linear_121_cast_fp16)[name = tensor("op_1895_cast_fp16")]; + tensor var_1896_perm_0 = const()[name = tensor("op_1896_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424653248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425882112))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425882304)))]; + tensor linear_122_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_1901 = const()[name = tensor("op_1901"), val = tensor([1, -1, 20, 64])]; + tensor var_1902_cast_fp16 = reshape(shape = var_1901, x = linear_122_cast_fp16)[name = tensor("op_1902_cast_fp16")]; + tensor var_1903_perm_0 = const()[name = tensor("op_1903_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1910 = const()[name = tensor("op_1910"), val = tensor([1, 77, 20, 64])]; + tensor var_1911_cast_fp16 = reshape(shape = var_1910, x = tensor_125_cast_fp16)[name = tensor("op_1911_cast_fp16")]; + tensor var_1912_perm_0 = const()[name = tensor("op_1912_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1914 = const()[name = tensor("op_1914"), val = tensor([20, -1, 64])]; + tensor var_1912_cast_fp16 = transpose(perm = var_1912_perm_0, x = var_1911_cast_fp16)[name = tensor("transpose_47")]; + tensor query_states_41_cast_fp16 = reshape(shape = var_1914, x = var_1912_cast_fp16)[name = tensor("query_states_41_cast_fp16")]; + tensor var_1916 = const()[name = tensor("op_1916"), val = tensor([20, -1, 64])]; + tensor var_1896_cast_fp16 = transpose(perm = var_1896_perm_0, x = var_1895_cast_fp16)[name = tensor("transpose_46")]; + tensor key_states_83_cast_fp16 = reshape(shape = var_1916, x = var_1896_cast_fp16)[name = tensor("key_states_83_cast_fp16")]; + tensor var_1918 = const()[name = tensor("op_1918"), val = tensor([20, -1, 64])]; + tensor var_1903_cast_fp16 = transpose(perm = var_1903_perm_0, x = var_1902_cast_fp16)[name = tensor("transpose_45")]; + tensor value_states_83_cast_fp16 = reshape(shape = var_1918, x = var_1903_cast_fp16)[name = tensor("value_states_83_cast_fp16")]; + tensor attn_weights_121_transpose_x_1 = const()[name = tensor("attn_weights_121_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_121_transpose_y_1 = const()[name = tensor("attn_weights_121_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_121_cast_fp16 = matmul(transpose_x = attn_weights_121_transpose_x_1, transpose_y = attn_weights_121_transpose_y_1, x = query_states_41_cast_fp16, y = key_states_83_cast_fp16)[name = tensor("attn_weights_121_cast_fp16")]; + tensor var_1923 = const()[name = tensor("op_1923"), val = tensor([1, 20, 77, 77])]; + tensor var_1924_cast_fp16 = reshape(shape = var_1923, x = attn_weights_121_cast_fp16)[name = tensor("op_1924_cast_fp16")]; + tensor attn_weights_123_cast_fp16 = add(x = var_1924_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_123_cast_fp16")]; + tensor var_1929 = const()[name = tensor("op_1929"), val = tensor([20, 77, 77])]; + tensor input_325_cast_fp16 = reshape(shape = var_1929, x = attn_weights_123_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor input_327_cast_fp16 = softmax(axis = var_5, x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; + tensor attn_output_121_transpose_x_0 = const()[name = tensor("attn_output_121_transpose_x_0"), val = tensor(false)]; + tensor attn_output_121_transpose_y_0 = const()[name = tensor("attn_output_121_transpose_y_0"), val = tensor(false)]; + tensor attn_output_121_cast_fp16 = matmul(transpose_x = attn_output_121_transpose_x_0, transpose_y = attn_output_121_transpose_y_0, x = input_327_cast_fp16, y = value_states_83_cast_fp16)[name = tensor("attn_output_121_cast_fp16")]; + tensor var_1934 = const()[name = tensor("op_1934"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_123_cast_fp16 = reshape(shape = var_1934, x = attn_output_121_cast_fp16)[name = tensor("attn_output_123_cast_fp16")]; + tensor attn_output_125_perm_0 = const()[name = tensor("attn_output_125_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1937 = const()[name = tensor("op_1937"), val = tensor([1, 77, 1280])]; + tensor attn_output_125_cast_fp16 = transpose(perm = attn_output_125_perm_0, x = attn_output_123_cast_fp16)[name = tensor("transpose_44")]; + tensor input_329_cast_fp16 = reshape(shape = var_1937, x = attn_output_125_cast_fp16)[name = tensor("input_329_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425884928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427113792))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427113984)))]; + tensor linear_123_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized, x = input_329_cast_fp16)[name = tensor("linear_123_cast_fp16")]; + tensor input_331_cast_fp16 = add(x = input_323_cast_fp16, y = linear_123_cast_fp16)[name = tensor("input_331_cast_fp16")]; + tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427116608)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427119232)))]; + tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427121856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432037120))), name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432037312)))]; + tensor linear_124_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized, x = input_333_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_337_mode_0 = const()[name = tensor("input_337_mode_0"), val = tensor("EXACT")]; + tensor input_337_cast_fp16 = gelu(mode = input_337_mode_0, x = linear_124_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432047616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436962880))), name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436963072)))]; + tensor linear_125_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized, x = input_337_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_339_cast_fp16 = add(x = input_331_cast_fp16, y = linear_125_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor hidden_states_127_axes_0 = const()[name = tensor("hidden_states_127_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436965696)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436968320)))]; + tensor hidden_states_127_cast_fp16 = layer_norm(axes = hidden_states_127_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("hidden_states_127_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436970944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438199808))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438200000)))]; + tensor linear_126_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_1976_to_fp16 = const()[name = tensor("op_1976_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_131_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_1976_to_fp16)[name = tensor("tensor_131_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438202624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439431488))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439431680)))]; + tensor linear_127_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor var_1981 = const()[name = tensor("op_1981"), val = tensor([1, -1, 20, 64])]; + tensor var_1982_cast_fp16 = reshape(shape = var_1981, x = linear_127_cast_fp16)[name = tensor("op_1982_cast_fp16")]; + tensor var_1983_perm_0 = const()[name = tensor("op_1983_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439434304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440663168))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440663360)))]; + tensor linear_128_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_1988 = const()[name = tensor("op_1988"), val = tensor([1, -1, 20, 64])]; + tensor var_1989_cast_fp16 = reshape(shape = var_1988, x = linear_128_cast_fp16)[name = tensor("op_1989_cast_fp16")]; + tensor var_1990_perm_0 = const()[name = tensor("op_1990_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1997 = const()[name = tensor("op_1997"), val = tensor([1, 77, 20, 64])]; + tensor var_1998_cast_fp16 = reshape(shape = var_1997, x = tensor_131_cast_fp16)[name = tensor("op_1998_cast_fp16")]; + tensor var_1999_perm_0 = const()[name = tensor("op_1999_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2001 = const()[name = tensor("op_2001"), val = tensor([20, -1, 64])]; + tensor var_1999_cast_fp16 = transpose(perm = var_1999_perm_0, x = var_1998_cast_fp16)[name = tensor("transpose_43")]; + tensor query_states_43_cast_fp16 = reshape(shape = var_2001, x = var_1999_cast_fp16)[name = tensor("query_states_43_cast_fp16")]; + tensor var_2003 = const()[name = tensor("op_2003"), val = tensor([20, -1, 64])]; + tensor var_1983_cast_fp16 = transpose(perm = var_1983_perm_0, x = var_1982_cast_fp16)[name = tensor("transpose_42")]; + tensor key_states_87_cast_fp16 = reshape(shape = var_2003, x = var_1983_cast_fp16)[name = tensor("key_states_87_cast_fp16")]; + tensor var_2005 = const()[name = tensor("op_2005"), val = tensor([20, -1, 64])]; + tensor var_1990_cast_fp16 = transpose(perm = var_1990_perm_0, x = var_1989_cast_fp16)[name = tensor("transpose_41")]; + tensor value_states_87_cast_fp16 = reshape(shape = var_2005, x = var_1990_cast_fp16)[name = tensor("value_states_87_cast_fp16")]; + tensor attn_weights_127_transpose_x_1 = const()[name = tensor("attn_weights_127_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_127_transpose_y_1 = const()[name = tensor("attn_weights_127_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_127_cast_fp16 = matmul(transpose_x = attn_weights_127_transpose_x_1, transpose_y = attn_weights_127_transpose_y_1, x = query_states_43_cast_fp16, y = key_states_87_cast_fp16)[name = tensor("attn_weights_127_cast_fp16")]; + tensor var_2010 = const()[name = tensor("op_2010"), val = tensor([1, 20, 77, 77])]; + tensor var_2011_cast_fp16 = reshape(shape = var_2010, x = attn_weights_127_cast_fp16)[name = tensor("op_2011_cast_fp16")]; + tensor attn_weights_129_cast_fp16 = add(x = var_2011_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_129_cast_fp16")]; + tensor var_2016 = const()[name = tensor("op_2016"), val = tensor([20, 77, 77])]; + tensor input_341_cast_fp16 = reshape(shape = var_2016, x = attn_weights_129_cast_fp16)[name = tensor("input_341_cast_fp16")]; + tensor input_343_cast_fp16 = softmax(axis = var_5, x = input_341_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor attn_output_127_transpose_x_0 = const()[name = tensor("attn_output_127_transpose_x_0"), val = tensor(false)]; + tensor attn_output_127_transpose_y_0 = const()[name = tensor("attn_output_127_transpose_y_0"), val = tensor(false)]; + tensor attn_output_127_cast_fp16 = matmul(transpose_x = attn_output_127_transpose_x_0, transpose_y = attn_output_127_transpose_y_0, x = input_343_cast_fp16, y = value_states_87_cast_fp16)[name = tensor("attn_output_127_cast_fp16")]; + tensor var_2021 = const()[name = tensor("op_2021"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_129_cast_fp16 = reshape(shape = var_2021, x = attn_output_127_cast_fp16)[name = tensor("attn_output_129_cast_fp16")]; + tensor attn_output_131_perm_0 = const()[name = tensor("attn_output_131_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2024 = const()[name = tensor("op_2024"), val = tensor([1, 77, 1280])]; + tensor attn_output_131_cast_fp16 = transpose(perm = attn_output_131_perm_0, x = attn_output_129_cast_fp16)[name = tensor("transpose_40")]; + tensor input_345_cast_fp16 = reshape(shape = var_2024, x = attn_output_131_cast_fp16)[name = tensor("input_345_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440665984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441894848))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441895040)))]; + tensor linear_129_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized, x = input_345_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor input_347_cast_fp16 = add(x = input_339_cast_fp16, y = linear_129_cast_fp16)[name = tensor("input_347_cast_fp16")]; + tensor input_349_axes_0 = const()[name = tensor("input_349_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441897664)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441900288)))]; + tensor input_349_cast_fp16 = layer_norm(axes = input_349_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441902912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446818176))), name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446818368)))]; + tensor linear_130_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized, x = input_349_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor input_353_mode_0 = const()[name = tensor("input_353_mode_0"), val = tensor("EXACT")]; + tensor input_353_cast_fp16 = gelu(mode = input_353_mode_0, x = linear_130_cast_fp16)[name = tensor("input_353_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446828672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451743936))), name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451744128)))]; + tensor linear_131_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized, x = input_353_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor input_355_cast_fp16 = add(x = input_347_cast_fp16, y = linear_131_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor hidden_states_133_axes_0 = const()[name = tensor("hidden_states_133_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451746752)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451749376)))]; + tensor hidden_states_133_cast_fp16 = layer_norm(axes = hidden_states_133_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("hidden_states_133_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451752000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452980864))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452981056)))]; + tensor linear_132_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_132_cast_fp16")]; + tensor var_2063_to_fp16 = const()[name = tensor("op_2063_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_137_cast_fp16 = mul(x = linear_132_cast_fp16, y = var_2063_to_fp16)[name = tensor("tensor_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452983680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454212544))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454212736)))]; + tensor linear_133_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor var_2068 = const()[name = tensor("op_2068"), val = tensor([1, -1, 20, 64])]; + tensor var_2069_cast_fp16 = reshape(shape = var_2068, x = linear_133_cast_fp16)[name = tensor("op_2069_cast_fp16")]; + tensor var_2070_perm_0 = const()[name = tensor("op_2070_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454215360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455444224))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455444416)))]; + tensor linear_134_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor var_2075 = const()[name = tensor("op_2075"), val = tensor([1, -1, 20, 64])]; + tensor var_2076_cast_fp16 = reshape(shape = var_2075, x = linear_134_cast_fp16)[name = tensor("op_2076_cast_fp16")]; + tensor var_2077_perm_0 = const()[name = tensor("op_2077_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2084 = const()[name = tensor("op_2084"), val = tensor([1, 77, 20, 64])]; + tensor var_2085_cast_fp16 = reshape(shape = var_2084, x = tensor_137_cast_fp16)[name = tensor("op_2085_cast_fp16")]; + tensor var_2086_perm_0 = const()[name = tensor("op_2086_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2088 = const()[name = tensor("op_2088"), val = tensor([20, -1, 64])]; + tensor var_2086_cast_fp16 = transpose(perm = var_2086_perm_0, x = var_2085_cast_fp16)[name = tensor("transpose_39")]; + tensor query_states_45_cast_fp16 = reshape(shape = var_2088, x = var_2086_cast_fp16)[name = tensor("query_states_45_cast_fp16")]; + tensor var_2090 = const()[name = tensor("op_2090"), val = tensor([20, -1, 64])]; + tensor var_2070_cast_fp16 = transpose(perm = var_2070_perm_0, x = var_2069_cast_fp16)[name = tensor("transpose_38")]; + tensor key_states_91_cast_fp16 = reshape(shape = var_2090, x = var_2070_cast_fp16)[name = tensor("key_states_91_cast_fp16")]; + tensor var_2092 = const()[name = tensor("op_2092"), val = tensor([20, -1, 64])]; + tensor var_2077_cast_fp16 = transpose(perm = var_2077_perm_0, x = var_2076_cast_fp16)[name = tensor("transpose_37")]; + tensor value_states_91_cast_fp16 = reshape(shape = var_2092, x = var_2077_cast_fp16)[name = tensor("value_states_91_cast_fp16")]; + tensor attn_weights_133_transpose_x_1 = const()[name = tensor("attn_weights_133_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_133_transpose_y_1 = const()[name = tensor("attn_weights_133_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_133_cast_fp16 = matmul(transpose_x = attn_weights_133_transpose_x_1, transpose_y = attn_weights_133_transpose_y_1, x = query_states_45_cast_fp16, y = key_states_91_cast_fp16)[name = tensor("attn_weights_133_cast_fp16")]; + tensor var_2097 = const()[name = tensor("op_2097"), val = tensor([1, 20, 77, 77])]; + tensor var_2098_cast_fp16 = reshape(shape = var_2097, x = attn_weights_133_cast_fp16)[name = tensor("op_2098_cast_fp16")]; + tensor attn_weights_135_cast_fp16 = add(x = var_2098_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_135_cast_fp16")]; + tensor var_2103 = const()[name = tensor("op_2103"), val = tensor([20, 77, 77])]; + tensor input_357_cast_fp16 = reshape(shape = var_2103, x = attn_weights_135_cast_fp16)[name = tensor("input_357_cast_fp16")]; + tensor input_359_cast_fp16 = softmax(axis = var_5, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor attn_output_133_transpose_x_0 = const()[name = tensor("attn_output_133_transpose_x_0"), val = tensor(false)]; + tensor attn_output_133_transpose_y_0 = const()[name = tensor("attn_output_133_transpose_y_0"), val = tensor(false)]; + tensor attn_output_133_cast_fp16 = matmul(transpose_x = attn_output_133_transpose_x_0, transpose_y = attn_output_133_transpose_y_0, x = input_359_cast_fp16, y = value_states_91_cast_fp16)[name = tensor("attn_output_133_cast_fp16")]; + tensor var_2108 = const()[name = tensor("op_2108"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_135_cast_fp16 = reshape(shape = var_2108, x = attn_output_133_cast_fp16)[name = tensor("attn_output_135_cast_fp16")]; + tensor attn_output_137_perm_0 = const()[name = tensor("attn_output_137_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2111 = const()[name = tensor("op_2111"), val = tensor([1, 77, 1280])]; + tensor attn_output_137_cast_fp16 = transpose(perm = attn_output_137_perm_0, x = attn_output_135_cast_fp16)[name = tensor("transpose_36")]; + tensor input_361_cast_fp16 = reshape(shape = var_2111, x = attn_output_137_cast_fp16)[name = tensor("input_361_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455447040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456675904))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456676096)))]; + tensor linear_135_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized, x = input_361_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor input_363_cast_fp16 = add(x = input_355_cast_fp16, y = linear_135_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor input_365_axes_0 = const()[name = tensor("input_365_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456678720)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456681344)))]; + tensor input_365_cast_fp16 = layer_norm(axes = input_365_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16, x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456683968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461599232))), name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461599424)))]; + tensor linear_136_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized, x = input_365_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_369_mode_0 = const()[name = tensor("input_369_mode_0"), val = tensor("EXACT")]; + tensor input_369_cast_fp16 = gelu(mode = input_369_mode_0, x = linear_136_cast_fp16)[name = tensor("input_369_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461609728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466524992))), name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466525184)))]; + tensor linear_137_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized, x = input_369_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor input_371_cast_fp16 = add(x = input_363_cast_fp16, y = linear_137_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor hidden_states_139_axes_0 = const()[name = tensor("hidden_states_139_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466527808)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466530432)))]; + tensor hidden_states_139_cast_fp16 = layer_norm(axes = hidden_states_139_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("hidden_states_139_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466533056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467761920))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467762112)))]; + tensor linear_138_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_2150_to_fp16 = const()[name = tensor("op_2150_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_143_cast_fp16 = mul(x = linear_138_cast_fp16, y = var_2150_to_fp16)[name = tensor("tensor_143_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467764736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468993600))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468993792)))]; + tensor linear_139_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_2155 = const()[name = tensor("op_2155"), val = tensor([1, -1, 20, 64])]; + tensor var_2156_cast_fp16 = reshape(shape = var_2155, x = linear_139_cast_fp16)[name = tensor("op_2156_cast_fp16")]; + tensor var_2157_perm_0 = const()[name = tensor("op_2157_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468996416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470225280))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470225472)))]; + tensor linear_140_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_2162 = const()[name = tensor("op_2162"), val = tensor([1, -1, 20, 64])]; + tensor var_2163_cast_fp16 = reshape(shape = var_2162, x = linear_140_cast_fp16)[name = tensor("op_2163_cast_fp16")]; + tensor var_2164_perm_0 = const()[name = tensor("op_2164_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2171 = const()[name = tensor("op_2171"), val = tensor([1, 77, 20, 64])]; + tensor var_2172_cast_fp16 = reshape(shape = var_2171, x = tensor_143_cast_fp16)[name = tensor("op_2172_cast_fp16")]; + tensor var_2173_perm_0 = const()[name = tensor("op_2173_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2175 = const()[name = tensor("op_2175"), val = tensor([20, -1, 64])]; + tensor var_2173_cast_fp16 = transpose(perm = var_2173_perm_0, x = var_2172_cast_fp16)[name = tensor("transpose_35")]; + tensor query_states_47_cast_fp16 = reshape(shape = var_2175, x = var_2173_cast_fp16)[name = tensor("query_states_47_cast_fp16")]; + tensor var_2177 = const()[name = tensor("op_2177"), val = tensor([20, -1, 64])]; + tensor var_2157_cast_fp16 = transpose(perm = var_2157_perm_0, x = var_2156_cast_fp16)[name = tensor("transpose_34")]; + tensor key_states_95_cast_fp16 = reshape(shape = var_2177, x = var_2157_cast_fp16)[name = tensor("key_states_95_cast_fp16")]; + tensor var_2179 = const()[name = tensor("op_2179"), val = tensor([20, -1, 64])]; + tensor var_2164_cast_fp16 = transpose(perm = var_2164_perm_0, x = var_2163_cast_fp16)[name = tensor("transpose_33")]; + tensor value_states_95_cast_fp16 = reshape(shape = var_2179, x = var_2164_cast_fp16)[name = tensor("value_states_95_cast_fp16")]; + tensor attn_weights_139_transpose_x_1 = const()[name = tensor("attn_weights_139_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_139_transpose_y_1 = const()[name = tensor("attn_weights_139_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_139_cast_fp16 = matmul(transpose_x = attn_weights_139_transpose_x_1, transpose_y = attn_weights_139_transpose_y_1, x = query_states_47_cast_fp16, y = key_states_95_cast_fp16)[name = tensor("attn_weights_139_cast_fp16")]; + tensor var_2184 = const()[name = tensor("op_2184"), val = tensor([1, 20, 77, 77])]; + tensor var_2185_cast_fp16 = reshape(shape = var_2184, x = attn_weights_139_cast_fp16)[name = tensor("op_2185_cast_fp16")]; + tensor attn_weights_141_cast_fp16 = add(x = var_2185_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_141_cast_fp16")]; + tensor var_2190 = const()[name = tensor("op_2190"), val = tensor([20, 77, 77])]; + tensor input_373_cast_fp16 = reshape(shape = var_2190, x = attn_weights_141_cast_fp16)[name = tensor("input_373_cast_fp16")]; + tensor input_375_cast_fp16 = softmax(axis = var_5, x = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor attn_output_139_transpose_x_0 = const()[name = tensor("attn_output_139_transpose_x_0"), val = tensor(false)]; + tensor attn_output_139_transpose_y_0 = const()[name = tensor("attn_output_139_transpose_y_0"), val = tensor(false)]; + tensor attn_output_139_cast_fp16 = matmul(transpose_x = attn_output_139_transpose_x_0, transpose_y = attn_output_139_transpose_y_0, x = input_375_cast_fp16, y = value_states_95_cast_fp16)[name = tensor("attn_output_139_cast_fp16")]; + tensor var_2195 = const()[name = tensor("op_2195"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_141_cast_fp16 = reshape(shape = var_2195, x = attn_output_139_cast_fp16)[name = tensor("attn_output_141_cast_fp16")]; + tensor attn_output_143_perm_0 = const()[name = tensor("attn_output_143_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2198 = const()[name = tensor("op_2198"), val = tensor([1, 77, 1280])]; + tensor attn_output_143_cast_fp16 = transpose(perm = attn_output_143_perm_0, x = attn_output_141_cast_fp16)[name = tensor("transpose_32")]; + tensor input_377_cast_fp16 = reshape(shape = var_2198, x = attn_output_143_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470228096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471456960))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471457152)))]; + tensor linear_141_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized, x = input_377_cast_fp16)[name = tensor("linear_141_cast_fp16")]; + tensor input_379_cast_fp16 = add(x = input_371_cast_fp16, y = linear_141_cast_fp16)[name = tensor("input_379_cast_fp16")]; + tensor input_381_axes_0 = const()[name = tensor("input_381_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471459776)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471462400)))]; + tensor input_381_cast_fp16 = layer_norm(axes = input_381_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471465024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476380288))), name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476380480)))]; + tensor linear_142_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized, x = input_381_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_385_mode_0 = const()[name = tensor("input_385_mode_0"), val = tensor("EXACT")]; + tensor input_385_cast_fp16 = gelu(mode = input_385_mode_0, x = linear_142_cast_fp16)[name = tensor("input_385_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476390784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481306048))), name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481306240)))]; + tensor linear_143_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized, x = input_385_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_379_cast_fp16, y = linear_143_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor hidden_states_145_axes_0 = const()[name = tensor("hidden_states_145_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481308864)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481311488)))]; + tensor hidden_states_145_cast_fp16 = layer_norm(axes = hidden_states_145_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("hidden_states_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481314112))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482542976))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482543168)))]; + tensor linear_144_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_2237_to_fp16 = const()[name = tensor("op_2237_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_149_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2237_to_fp16)[name = tensor("tensor_149_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482545792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483774656))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483774848)))]; + tensor linear_145_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor var_2242 = const()[name = tensor("op_2242"), val = tensor([1, -1, 20, 64])]; + tensor var_2243_cast_fp16 = reshape(shape = var_2242, x = linear_145_cast_fp16)[name = tensor("op_2243_cast_fp16")]; + tensor var_2244_perm_0 = const()[name = tensor("op_2244_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483777472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485006336))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485006528)))]; + tensor linear_146_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_2249 = const()[name = tensor("op_2249"), val = tensor([1, -1, 20, 64])]; + tensor var_2250_cast_fp16 = reshape(shape = var_2249, x = linear_146_cast_fp16)[name = tensor("op_2250_cast_fp16")]; + tensor var_2251_perm_0 = const()[name = tensor("op_2251_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2258 = const()[name = tensor("op_2258"), val = tensor([1, 77, 20, 64])]; + tensor var_2259_cast_fp16 = reshape(shape = var_2258, x = tensor_149_cast_fp16)[name = tensor("op_2259_cast_fp16")]; + tensor var_2260_perm_0 = const()[name = tensor("op_2260_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([20, -1, 64])]; + tensor var_2260_cast_fp16 = transpose(perm = var_2260_perm_0, x = var_2259_cast_fp16)[name = tensor("transpose_31")]; + tensor query_states_49_cast_fp16 = reshape(shape = var_2262, x = var_2260_cast_fp16)[name = tensor("query_states_49_cast_fp16")]; + tensor var_2264 = const()[name = tensor("op_2264"), val = tensor([20, -1, 64])]; + tensor var_2244_cast_fp16 = transpose(perm = var_2244_perm_0, x = var_2243_cast_fp16)[name = tensor("transpose_30")]; + tensor key_states_99_cast_fp16 = reshape(shape = var_2264, x = var_2244_cast_fp16)[name = tensor("key_states_99_cast_fp16")]; + tensor var_2266 = const()[name = tensor("op_2266"), val = tensor([20, -1, 64])]; + tensor var_2251_cast_fp16 = transpose(perm = var_2251_perm_0, x = var_2250_cast_fp16)[name = tensor("transpose_29")]; + tensor value_states_99_cast_fp16 = reshape(shape = var_2266, x = var_2251_cast_fp16)[name = tensor("value_states_99_cast_fp16")]; + tensor attn_weights_145_transpose_x_1 = const()[name = tensor("attn_weights_145_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_145_transpose_y_1 = const()[name = tensor("attn_weights_145_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_145_cast_fp16 = matmul(transpose_x = attn_weights_145_transpose_x_1, transpose_y = attn_weights_145_transpose_y_1, x = query_states_49_cast_fp16, y = key_states_99_cast_fp16)[name = tensor("attn_weights_145_cast_fp16")]; + tensor var_2271 = const()[name = tensor("op_2271"), val = tensor([1, 20, 77, 77])]; + tensor var_2272_cast_fp16 = reshape(shape = var_2271, x = attn_weights_145_cast_fp16)[name = tensor("op_2272_cast_fp16")]; + tensor attn_weights_147_cast_fp16 = add(x = var_2272_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_147_cast_fp16")]; + tensor var_2277 = const()[name = tensor("op_2277"), val = tensor([20, 77, 77])]; + tensor input_389_cast_fp16 = reshape(shape = var_2277, x = attn_weights_147_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor input_391_cast_fp16 = softmax(axis = var_5, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor attn_output_145_transpose_x_0 = const()[name = tensor("attn_output_145_transpose_x_0"), val = tensor(false)]; + tensor attn_output_145_transpose_y_0 = const()[name = tensor("attn_output_145_transpose_y_0"), val = tensor(false)]; + tensor attn_output_145_cast_fp16 = matmul(transpose_x = attn_output_145_transpose_x_0, transpose_y = attn_output_145_transpose_y_0, x = input_391_cast_fp16, y = value_states_99_cast_fp16)[name = tensor("attn_output_145_cast_fp16")]; + tensor var_2282 = const()[name = tensor("op_2282"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_147_cast_fp16 = reshape(shape = var_2282, x = attn_output_145_cast_fp16)[name = tensor("attn_output_147_cast_fp16")]; + tensor attn_output_149_perm_0 = const()[name = tensor("attn_output_149_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2285 = const()[name = tensor("op_2285"), val = tensor([1, 77, 1280])]; + tensor attn_output_149_cast_fp16 = transpose(perm = attn_output_149_perm_0, x = attn_output_147_cast_fp16)[name = tensor("transpose_28")]; + tensor input_393_cast_fp16 = reshape(shape = var_2285, x = attn_output_149_cast_fp16)[name = tensor("input_393_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485009152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486238016))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486238208)))]; + tensor linear_147_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized, x = input_393_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor input_395_cast_fp16 = add(x = input_387_cast_fp16, y = linear_147_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor input_397_axes_0 = const()[name = tensor("input_397_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486240832)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486243456)))]; + tensor input_397_cast_fp16 = layer_norm(axes = input_397_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("input_397_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486246080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491161344))), name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491161536)))]; + tensor linear_148_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized, x = input_397_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor input_401_mode_0 = const()[name = tensor("input_401_mode_0"), val = tensor("EXACT")]; + tensor input_401_cast_fp16 = gelu(mode = input_401_mode_0, x = linear_148_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491171840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496087104))), name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496087296)))]; + tensor linear_149_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized, x = input_401_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_149_cast_fp16)[name = tensor("input_403_cast_fp16")]; + tensor hidden_states_151_axes_0 = const()[name = tensor("hidden_states_151_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496089920)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496092544)))]; + tensor hidden_states_151_cast_fp16 = layer_norm(axes = hidden_states_151_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("hidden_states_151_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(496095168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497324032))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497324224)))]; + tensor linear_150_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_150_cast_fp16")]; + tensor var_2324_to_fp16 = const()[name = tensor("op_2324_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_155_cast_fp16 = mul(x = linear_150_cast_fp16, y = var_2324_to_fp16)[name = tensor("tensor_155_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497326848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498555712))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498555904)))]; + tensor linear_151_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor var_2329 = const()[name = tensor("op_2329"), val = tensor([1, -1, 20, 64])]; + tensor var_2330_cast_fp16 = reshape(shape = var_2329, x = linear_151_cast_fp16)[name = tensor("op_2330_cast_fp16")]; + tensor var_2331_perm_0 = const()[name = tensor("op_2331_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498558528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499787392))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499787584)))]; + tensor linear_152_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor var_2336 = const()[name = tensor("op_2336"), val = tensor([1, -1, 20, 64])]; + tensor var_2337_cast_fp16 = reshape(shape = var_2336, x = linear_152_cast_fp16)[name = tensor("op_2337_cast_fp16")]; + tensor var_2338_perm_0 = const()[name = tensor("op_2338_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2345 = const()[name = tensor("op_2345"), val = tensor([1, 77, 20, 64])]; + tensor var_2346_cast_fp16 = reshape(shape = var_2345, x = tensor_155_cast_fp16)[name = tensor("op_2346_cast_fp16")]; + tensor var_2347_perm_0 = const()[name = tensor("op_2347_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2349 = const()[name = tensor("op_2349"), val = tensor([20, -1, 64])]; + tensor var_2347_cast_fp16 = transpose(perm = var_2347_perm_0, x = var_2346_cast_fp16)[name = tensor("transpose_27")]; + tensor query_states_51_cast_fp16 = reshape(shape = var_2349, x = var_2347_cast_fp16)[name = tensor("query_states_51_cast_fp16")]; + tensor var_2351 = const()[name = tensor("op_2351"), val = tensor([20, -1, 64])]; + tensor var_2331_cast_fp16 = transpose(perm = var_2331_perm_0, x = var_2330_cast_fp16)[name = tensor("transpose_26")]; + tensor key_states_103_cast_fp16 = reshape(shape = var_2351, x = var_2331_cast_fp16)[name = tensor("key_states_103_cast_fp16")]; + tensor var_2353 = const()[name = tensor("op_2353"), val = tensor([20, -1, 64])]; + tensor var_2338_cast_fp16 = transpose(perm = var_2338_perm_0, x = var_2337_cast_fp16)[name = tensor("transpose_25")]; + tensor value_states_103_cast_fp16 = reshape(shape = var_2353, x = var_2338_cast_fp16)[name = tensor("value_states_103_cast_fp16")]; + tensor attn_weights_151_transpose_x_1 = const()[name = tensor("attn_weights_151_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_151_transpose_y_1 = const()[name = tensor("attn_weights_151_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_151_cast_fp16 = matmul(transpose_x = attn_weights_151_transpose_x_1, transpose_y = attn_weights_151_transpose_y_1, x = query_states_51_cast_fp16, y = key_states_103_cast_fp16)[name = tensor("attn_weights_151_cast_fp16")]; + tensor var_2358 = const()[name = tensor("op_2358"), val = tensor([1, 20, 77, 77])]; + tensor var_2359_cast_fp16 = reshape(shape = var_2358, x = attn_weights_151_cast_fp16)[name = tensor("op_2359_cast_fp16")]; + tensor attn_weights_153_cast_fp16 = add(x = var_2359_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_153_cast_fp16")]; + tensor var_2364 = const()[name = tensor("op_2364"), val = tensor([20, 77, 77])]; + tensor input_405_cast_fp16 = reshape(shape = var_2364, x = attn_weights_153_cast_fp16)[name = tensor("input_405_cast_fp16")]; + tensor input_407_cast_fp16 = softmax(axis = var_5, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor attn_output_151_transpose_x_0 = const()[name = tensor("attn_output_151_transpose_x_0"), val = tensor(false)]; + tensor attn_output_151_transpose_y_0 = const()[name = tensor("attn_output_151_transpose_y_0"), val = tensor(false)]; + tensor attn_output_151_cast_fp16 = matmul(transpose_x = attn_output_151_transpose_x_0, transpose_y = attn_output_151_transpose_y_0, x = input_407_cast_fp16, y = value_states_103_cast_fp16)[name = tensor("attn_output_151_cast_fp16")]; + tensor var_2369 = const()[name = tensor("op_2369"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_153_cast_fp16 = reshape(shape = var_2369, x = attn_output_151_cast_fp16)[name = tensor("attn_output_153_cast_fp16")]; + tensor attn_output_155_perm_0 = const()[name = tensor("attn_output_155_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2372 = const()[name = tensor("op_2372"), val = tensor([1, 77, 1280])]; + tensor attn_output_155_cast_fp16 = transpose(perm = attn_output_155_perm_0, x = attn_output_153_cast_fp16)[name = tensor("transpose_24")]; + tensor input_409_cast_fp16 = reshape(shape = var_2372, x = attn_output_155_cast_fp16)[name = tensor("input_409_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499790208))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501019072))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501019264)))]; + tensor linear_153_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized, x = input_409_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_403_cast_fp16, y = linear_153_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor input_413_axes_0 = const()[name = tensor("input_413_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501021888)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501024512)))]; + tensor input_413_cast_fp16 = layer_norm(axes = input_413_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("input_413_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501027136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505942400))), name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505942592)))]; + tensor linear_154_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized, x = input_413_cast_fp16)[name = tensor("linear_154_cast_fp16")]; + tensor input_417_mode_0 = const()[name = tensor("input_417_mode_0"), val = tensor("EXACT")]; + tensor input_417_cast_fp16 = gelu(mode = input_417_mode_0, x = linear_154_cast_fp16)[name = tensor("input_417_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505952896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510868160))), name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510868352)))]; + tensor linear_155_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized, x = input_417_cast_fp16)[name = tensor("linear_155_cast_fp16")]; + tensor input_419_cast_fp16 = add(x = input_411_cast_fp16, y = linear_155_cast_fp16)[name = tensor("input_419_cast_fp16")]; + tensor hidden_states_157_axes_0 = const()[name = tensor("hidden_states_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510870976)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510873600)))]; + tensor hidden_states_157_cast_fp16 = layer_norm(axes = hidden_states_157_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("hidden_states_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510876224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512105088))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512105280)))]; + tensor linear_156_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_156_cast_fp16")]; + tensor var_2411_to_fp16 = const()[name = tensor("op_2411_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_161_cast_fp16 = mul(x = linear_156_cast_fp16, y = var_2411_to_fp16)[name = tensor("tensor_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(512107904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513336768))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513336960)))]; + tensor linear_157_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_157_cast_fp16")]; + tensor var_2416 = const()[name = tensor("op_2416"), val = tensor([1, -1, 20, 64])]; + tensor var_2417_cast_fp16 = reshape(shape = var_2416, x = linear_157_cast_fp16)[name = tensor("op_2417_cast_fp16")]; + tensor var_2418_perm_0 = const()[name = tensor("op_2418_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513339584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514568448))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514568640)))]; + tensor linear_158_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_158_cast_fp16")]; + tensor var_2423 = const()[name = tensor("op_2423"), val = tensor([1, -1, 20, 64])]; + tensor var_2424_cast_fp16 = reshape(shape = var_2423, x = linear_158_cast_fp16)[name = tensor("op_2424_cast_fp16")]; + tensor var_2425_perm_0 = const()[name = tensor("op_2425_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2432 = const()[name = tensor("op_2432"), val = tensor([1, 77, 20, 64])]; + tensor var_2433_cast_fp16 = reshape(shape = var_2432, x = tensor_161_cast_fp16)[name = tensor("op_2433_cast_fp16")]; + tensor var_2434_perm_0 = const()[name = tensor("op_2434_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2436 = const()[name = tensor("op_2436"), val = tensor([20, -1, 64])]; + tensor var_2434_cast_fp16 = transpose(perm = var_2434_perm_0, x = var_2433_cast_fp16)[name = tensor("transpose_23")]; + tensor query_states_53_cast_fp16 = reshape(shape = var_2436, x = var_2434_cast_fp16)[name = tensor("query_states_53_cast_fp16")]; + tensor var_2438 = const()[name = tensor("op_2438"), val = tensor([20, -1, 64])]; + tensor var_2418_cast_fp16 = transpose(perm = var_2418_perm_0, x = var_2417_cast_fp16)[name = tensor("transpose_22")]; + tensor key_states_107_cast_fp16 = reshape(shape = var_2438, x = var_2418_cast_fp16)[name = tensor("key_states_107_cast_fp16")]; + tensor var_2440 = const()[name = tensor("op_2440"), val = tensor([20, -1, 64])]; + tensor var_2425_cast_fp16 = transpose(perm = var_2425_perm_0, x = var_2424_cast_fp16)[name = tensor("transpose_21")]; + tensor value_states_107_cast_fp16 = reshape(shape = var_2440, x = var_2425_cast_fp16)[name = tensor("value_states_107_cast_fp16")]; + tensor attn_weights_157_transpose_x_1 = const()[name = tensor("attn_weights_157_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_157_transpose_y_1 = const()[name = tensor("attn_weights_157_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_157_cast_fp16 = matmul(transpose_x = attn_weights_157_transpose_x_1, transpose_y = attn_weights_157_transpose_y_1, x = query_states_53_cast_fp16, y = key_states_107_cast_fp16)[name = tensor("attn_weights_157_cast_fp16")]; + tensor var_2445 = const()[name = tensor("op_2445"), val = tensor([1, 20, 77, 77])]; + tensor var_2446_cast_fp16 = reshape(shape = var_2445, x = attn_weights_157_cast_fp16)[name = tensor("op_2446_cast_fp16")]; + tensor attn_weights_159_cast_fp16 = add(x = var_2446_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_159_cast_fp16")]; + tensor var_2451 = const()[name = tensor("op_2451"), val = tensor([20, 77, 77])]; + tensor input_421_cast_fp16 = reshape(shape = var_2451, x = attn_weights_159_cast_fp16)[name = tensor("input_421_cast_fp16")]; + tensor input_423_cast_fp16 = softmax(axis = var_5, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor attn_output_157_transpose_x_0 = const()[name = tensor("attn_output_157_transpose_x_0"), val = tensor(false)]; + tensor attn_output_157_transpose_y_0 = const()[name = tensor("attn_output_157_transpose_y_0"), val = tensor(false)]; + tensor attn_output_157_cast_fp16 = matmul(transpose_x = attn_output_157_transpose_x_0, transpose_y = attn_output_157_transpose_y_0, x = input_423_cast_fp16, y = value_states_107_cast_fp16)[name = tensor("attn_output_157_cast_fp16")]; + tensor var_2456 = const()[name = tensor("op_2456"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_159_cast_fp16 = reshape(shape = var_2456, x = attn_output_157_cast_fp16)[name = tensor("attn_output_159_cast_fp16")]; + tensor attn_output_161_perm_0 = const()[name = tensor("attn_output_161_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2459 = const()[name = tensor("op_2459"), val = tensor([1, 77, 1280])]; + tensor attn_output_161_cast_fp16 = transpose(perm = attn_output_161_perm_0, x = attn_output_159_cast_fp16)[name = tensor("transpose_20")]; + tensor input_425_cast_fp16 = reshape(shape = var_2459, x = attn_output_161_cast_fp16)[name = tensor("input_425_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514571264))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515800128))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515800320)))]; + tensor linear_159_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized, x = input_425_cast_fp16)[name = tensor("linear_159_cast_fp16")]; + tensor input_427_cast_fp16 = add(x = input_419_cast_fp16, y = linear_159_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515802944)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515805568)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515808192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520723456))), name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520723648)))]; + tensor linear_160_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized, x = input_429_cast_fp16)[name = tensor("linear_160_cast_fp16")]; + tensor input_433_mode_0 = const()[name = tensor("input_433_mode_0"), val = tensor("EXACT")]; + tensor input_433_cast_fp16 = gelu(mode = input_433_mode_0, x = linear_160_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520733952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525649216))), name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525649408)))]; + tensor linear_161_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized, x = input_433_cast_fp16)[name = tensor("linear_161_cast_fp16")]; + tensor input_435_cast_fp16 = add(x = input_427_cast_fp16, y = linear_161_cast_fp16)[name = tensor("input_435_cast_fp16")]; + tensor hidden_states_163_axes_0 = const()[name = tensor("hidden_states_163_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525652032)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525654656)))]; + tensor hidden_states_163_cast_fp16 = layer_norm(axes = hidden_states_163_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("hidden_states_163_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525657280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526886144))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526886336)))]; + tensor linear_162_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_162_cast_fp16")]; + tensor var_2498_to_fp16 = const()[name = tensor("op_2498_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_167_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_2498_to_fp16)[name = tensor("tensor_167_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526888960))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528117824))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528118016)))]; + tensor linear_163_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_163_cast_fp16")]; + tensor var_2503 = const()[name = tensor("op_2503"), val = tensor([1, -1, 20, 64])]; + tensor var_2504_cast_fp16 = reshape(shape = var_2503, x = linear_163_cast_fp16)[name = tensor("op_2504_cast_fp16")]; + tensor var_2505_perm_0 = const()[name = tensor("op_2505_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(528120640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529349504))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529349696)))]; + tensor linear_164_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_164_cast_fp16")]; + tensor var_2510 = const()[name = tensor("op_2510"), val = tensor([1, -1, 20, 64])]; + tensor var_2511_cast_fp16 = reshape(shape = var_2510, x = linear_164_cast_fp16)[name = tensor("op_2511_cast_fp16")]; + tensor var_2512_perm_0 = const()[name = tensor("op_2512_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2519 = const()[name = tensor("op_2519"), val = tensor([1, 77, 20, 64])]; + tensor var_2520_cast_fp16 = reshape(shape = var_2519, x = tensor_167_cast_fp16)[name = tensor("op_2520_cast_fp16")]; + tensor var_2521_perm_0 = const()[name = tensor("op_2521_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2523 = const()[name = tensor("op_2523"), val = tensor([20, -1, 64])]; + tensor var_2521_cast_fp16 = transpose(perm = var_2521_perm_0, x = var_2520_cast_fp16)[name = tensor("transpose_19")]; + tensor query_states_55_cast_fp16 = reshape(shape = var_2523, x = var_2521_cast_fp16)[name = tensor("query_states_55_cast_fp16")]; + tensor var_2525 = const()[name = tensor("op_2525"), val = tensor([20, -1, 64])]; + tensor var_2505_cast_fp16 = transpose(perm = var_2505_perm_0, x = var_2504_cast_fp16)[name = tensor("transpose_18")]; + tensor key_states_111_cast_fp16 = reshape(shape = var_2525, x = var_2505_cast_fp16)[name = tensor("key_states_111_cast_fp16")]; + tensor var_2527 = const()[name = tensor("op_2527"), val = tensor([20, -1, 64])]; + tensor var_2512_cast_fp16 = transpose(perm = var_2512_perm_0, x = var_2511_cast_fp16)[name = tensor("transpose_17")]; + tensor value_states_111_cast_fp16 = reshape(shape = var_2527, x = var_2512_cast_fp16)[name = tensor("value_states_111_cast_fp16")]; + tensor attn_weights_163_transpose_x_1 = const()[name = tensor("attn_weights_163_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_163_transpose_y_1 = const()[name = tensor("attn_weights_163_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_163_cast_fp16 = matmul(transpose_x = attn_weights_163_transpose_x_1, transpose_y = attn_weights_163_transpose_y_1, x = query_states_55_cast_fp16, y = key_states_111_cast_fp16)[name = tensor("attn_weights_163_cast_fp16")]; + tensor var_2532 = const()[name = tensor("op_2532"), val = tensor([1, 20, 77, 77])]; + tensor var_2533_cast_fp16 = reshape(shape = var_2532, x = attn_weights_163_cast_fp16)[name = tensor("op_2533_cast_fp16")]; + tensor attn_weights_165_cast_fp16 = add(x = var_2533_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_165_cast_fp16")]; + tensor var_2538 = const()[name = tensor("op_2538"), val = tensor([20, 77, 77])]; + tensor input_437_cast_fp16 = reshape(shape = var_2538, x = attn_weights_165_cast_fp16)[name = tensor("input_437_cast_fp16")]; + tensor input_439_cast_fp16 = softmax(axis = var_5, x = input_437_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor attn_output_163_transpose_x_0 = const()[name = tensor("attn_output_163_transpose_x_0"), val = tensor(false)]; + tensor attn_output_163_transpose_y_0 = const()[name = tensor("attn_output_163_transpose_y_0"), val = tensor(false)]; + tensor attn_output_163_cast_fp16 = matmul(transpose_x = attn_output_163_transpose_x_0, transpose_y = attn_output_163_transpose_y_0, x = input_439_cast_fp16, y = value_states_111_cast_fp16)[name = tensor("attn_output_163_cast_fp16")]; + tensor var_2543 = const()[name = tensor("op_2543"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_165_cast_fp16 = reshape(shape = var_2543, x = attn_output_163_cast_fp16)[name = tensor("attn_output_165_cast_fp16")]; + tensor attn_output_167_perm_0 = const()[name = tensor("attn_output_167_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2546 = const()[name = tensor("op_2546"), val = tensor([1, 77, 1280])]; + tensor attn_output_167_cast_fp16 = transpose(perm = attn_output_167_perm_0, x = attn_output_165_cast_fp16)[name = tensor("transpose_16")]; + tensor input_441_cast_fp16 = reshape(shape = var_2546, x = attn_output_167_cast_fp16)[name = tensor("input_441_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529352320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530581184))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530581376)))]; + tensor linear_165_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized, x = input_441_cast_fp16)[name = tensor("linear_165_cast_fp16")]; + tensor input_443_cast_fp16 = add(x = input_435_cast_fp16, y = linear_165_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor input_445_axes_0 = const()[name = tensor("input_445_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530584000)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530586624)))]; + tensor input_445_cast_fp16 = layer_norm(axes = input_445_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("input_445_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530589248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535504512))), name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535504704)))]; + tensor linear_166_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized, x = input_445_cast_fp16)[name = tensor("linear_166_cast_fp16")]; + tensor input_449_mode_0 = const()[name = tensor("input_449_mode_0"), val = tensor("EXACT")]; + tensor input_449_cast_fp16 = gelu(mode = input_449_mode_0, x = linear_166_cast_fp16)[name = tensor("input_449_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535515008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540430272))), name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540430464)))]; + tensor linear_167_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized, x = input_449_cast_fp16)[name = tensor("linear_167_cast_fp16")]; + tensor input_451_cast_fp16 = add(x = input_443_cast_fp16, y = linear_167_cast_fp16)[name = tensor("input_451_cast_fp16")]; + tensor hidden_states_169_axes_0 = const()[name = tensor("hidden_states_169_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540433088)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540435712)))]; + tensor hidden_states_169_cast_fp16 = layer_norm(axes = hidden_states_169_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("hidden_states_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540438336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541667200))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541667392)))]; + tensor linear_168_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_168_cast_fp16")]; + tensor var_2585_to_fp16 = const()[name = tensor("op_2585_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_173_cast_fp16 = mul(x = linear_168_cast_fp16, y = var_2585_to_fp16)[name = tensor("tensor_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541670016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542898880))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542899072)))]; + tensor linear_169_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_169_cast_fp16")]; + tensor var_2590 = const()[name = tensor("op_2590"), val = tensor([1, -1, 20, 64])]; + tensor var_2591_cast_fp16 = reshape(shape = var_2590, x = linear_169_cast_fp16)[name = tensor("op_2591_cast_fp16")]; + tensor var_2592_perm_0 = const()[name = tensor("op_2592_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542901696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544130560))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544130752)))]; + tensor linear_170_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_170_cast_fp16")]; + tensor var_2597 = const()[name = tensor("op_2597"), val = tensor([1, -1, 20, 64])]; + tensor var_2598_cast_fp16 = reshape(shape = var_2597, x = linear_170_cast_fp16)[name = tensor("op_2598_cast_fp16")]; + tensor var_2599_perm_0 = const()[name = tensor("op_2599_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2606 = const()[name = tensor("op_2606"), val = tensor([1, 77, 20, 64])]; + tensor var_2607_cast_fp16 = reshape(shape = var_2606, x = tensor_173_cast_fp16)[name = tensor("op_2607_cast_fp16")]; + tensor var_2608_perm_0 = const()[name = tensor("op_2608_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2610 = const()[name = tensor("op_2610"), val = tensor([20, -1, 64])]; + tensor var_2608_cast_fp16 = transpose(perm = var_2608_perm_0, x = var_2607_cast_fp16)[name = tensor("transpose_15")]; + tensor query_states_57_cast_fp16 = reshape(shape = var_2610, x = var_2608_cast_fp16)[name = tensor("query_states_57_cast_fp16")]; + tensor var_2612 = const()[name = tensor("op_2612"), val = tensor([20, -1, 64])]; + tensor var_2592_cast_fp16 = transpose(perm = var_2592_perm_0, x = var_2591_cast_fp16)[name = tensor("transpose_14")]; + tensor key_states_115_cast_fp16 = reshape(shape = var_2612, x = var_2592_cast_fp16)[name = tensor("key_states_115_cast_fp16")]; + tensor var_2614 = const()[name = tensor("op_2614"), val = tensor([20, -1, 64])]; + tensor var_2599_cast_fp16 = transpose(perm = var_2599_perm_0, x = var_2598_cast_fp16)[name = tensor("transpose_13")]; + tensor value_states_115_cast_fp16 = reshape(shape = var_2614, x = var_2599_cast_fp16)[name = tensor("value_states_115_cast_fp16")]; + tensor attn_weights_169_transpose_x_1 = const()[name = tensor("attn_weights_169_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_169_transpose_y_1 = const()[name = tensor("attn_weights_169_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_169_cast_fp16 = matmul(transpose_x = attn_weights_169_transpose_x_1, transpose_y = attn_weights_169_transpose_y_1, x = query_states_57_cast_fp16, y = key_states_115_cast_fp16)[name = tensor("attn_weights_169_cast_fp16")]; + tensor var_2619 = const()[name = tensor("op_2619"), val = tensor([1, 20, 77, 77])]; + tensor var_2620_cast_fp16 = reshape(shape = var_2619, x = attn_weights_169_cast_fp16)[name = tensor("op_2620_cast_fp16")]; + tensor attn_weights_171_cast_fp16 = add(x = var_2620_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_171_cast_fp16")]; + tensor var_2625 = const()[name = tensor("op_2625"), val = tensor([20, 77, 77])]; + tensor input_453_cast_fp16 = reshape(shape = var_2625, x = attn_weights_171_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor input_455_cast_fp16 = softmax(axis = var_5, x = input_453_cast_fp16)[name = tensor("input_455_cast_fp16")]; + tensor attn_output_169_transpose_x_0 = const()[name = tensor("attn_output_169_transpose_x_0"), val = tensor(false)]; + tensor attn_output_169_transpose_y_0 = const()[name = tensor("attn_output_169_transpose_y_0"), val = tensor(false)]; + tensor attn_output_169_cast_fp16 = matmul(transpose_x = attn_output_169_transpose_x_0, transpose_y = attn_output_169_transpose_y_0, x = input_455_cast_fp16, y = value_states_115_cast_fp16)[name = tensor("attn_output_169_cast_fp16")]; + tensor var_2630 = const()[name = tensor("op_2630"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_171_cast_fp16 = reshape(shape = var_2630, x = attn_output_169_cast_fp16)[name = tensor("attn_output_171_cast_fp16")]; + tensor attn_output_173_perm_0 = const()[name = tensor("attn_output_173_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2633 = const()[name = tensor("op_2633"), val = tensor([1, 77, 1280])]; + tensor attn_output_173_cast_fp16 = transpose(perm = attn_output_173_perm_0, x = attn_output_171_cast_fp16)[name = tensor("transpose_12")]; + tensor input_457_cast_fp16 = reshape(shape = var_2633, x = attn_output_173_cast_fp16)[name = tensor("input_457_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(544133376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545362240))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545362432)))]; + tensor linear_171_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized, x = input_457_cast_fp16)[name = tensor("linear_171_cast_fp16")]; + tensor input_459_cast_fp16 = add(x = input_451_cast_fp16, y = linear_171_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor input_461_axes_0 = const()[name = tensor("input_461_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545365056)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545367680)))]; + tensor input_461_cast_fp16 = layer_norm(axes = input_461_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("input_461_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545370304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550285568))), name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550285760)))]; + tensor linear_172_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized, x = input_461_cast_fp16)[name = tensor("linear_172_cast_fp16")]; + tensor input_465_mode_0 = const()[name = tensor("input_465_mode_0"), val = tensor("EXACT")]; + tensor input_465_cast_fp16 = gelu(mode = input_465_mode_0, x = linear_172_cast_fp16)[name = tensor("input_465_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550296064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555211328))), name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555211520)))]; + tensor linear_173_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized, x = input_465_cast_fp16)[name = tensor("linear_173_cast_fp16")]; + tensor input_467_cast_fp16 = add(x = input_459_cast_fp16, y = linear_173_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor hidden_states_175_axes_0 = const()[name = tensor("hidden_states_175_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555214144)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555216768)))]; + tensor hidden_states_175_cast_fp16 = layer_norm(axes = hidden_states_175_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16, x = input_467_cast_fp16)[name = tensor("hidden_states_175_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555219392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556448256))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556448448)))]; + tensor linear_174_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_174_cast_fp16")]; + tensor var_2672_to_fp16 = const()[name = tensor("op_2672_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_179_cast_fp16 = mul(x = linear_174_cast_fp16, y = var_2672_to_fp16)[name = tensor("tensor_179_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556451072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557679936))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557680128)))]; + tensor linear_175_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_175_cast_fp16")]; + tensor var_2677 = const()[name = tensor("op_2677"), val = tensor([1, -1, 20, 64])]; + tensor var_2678_cast_fp16 = reshape(shape = var_2677, x = linear_175_cast_fp16)[name = tensor("op_2678_cast_fp16")]; + tensor var_2679_perm_0 = const()[name = tensor("op_2679_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557682752))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558911616))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558911808)))]; + tensor linear_176_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_176_cast_fp16")]; + tensor var_2684 = const()[name = tensor("op_2684"), val = tensor([1, -1, 20, 64])]; + tensor var_2685_cast_fp16 = reshape(shape = var_2684, x = linear_176_cast_fp16)[name = tensor("op_2685_cast_fp16")]; + tensor var_2686_perm_0 = const()[name = tensor("op_2686_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2693 = const()[name = tensor("op_2693"), val = tensor([1, 77, 20, 64])]; + tensor var_2694_cast_fp16 = reshape(shape = var_2693, x = tensor_179_cast_fp16)[name = tensor("op_2694_cast_fp16")]; + tensor var_2695_perm_0 = const()[name = tensor("op_2695_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2697 = const()[name = tensor("op_2697"), val = tensor([20, -1, 64])]; + tensor var_2695_cast_fp16 = transpose(perm = var_2695_perm_0, x = var_2694_cast_fp16)[name = tensor("transpose_11")]; + tensor query_states_59_cast_fp16 = reshape(shape = var_2697, x = var_2695_cast_fp16)[name = tensor("query_states_59_cast_fp16")]; + tensor var_2699 = const()[name = tensor("op_2699"), val = tensor([20, -1, 64])]; + tensor var_2679_cast_fp16 = transpose(perm = var_2679_perm_0, x = var_2678_cast_fp16)[name = tensor("transpose_10")]; + tensor key_states_119_cast_fp16 = reshape(shape = var_2699, x = var_2679_cast_fp16)[name = tensor("key_states_119_cast_fp16")]; + tensor var_2701 = const()[name = tensor("op_2701"), val = tensor([20, -1, 64])]; + tensor var_2686_cast_fp16 = transpose(perm = var_2686_perm_0, x = var_2685_cast_fp16)[name = tensor("transpose_9")]; + tensor value_states_119_cast_fp16 = reshape(shape = var_2701, x = var_2686_cast_fp16)[name = tensor("value_states_119_cast_fp16")]; + tensor attn_weights_175_transpose_x_1 = const()[name = tensor("attn_weights_175_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_175_transpose_y_1 = const()[name = tensor("attn_weights_175_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_175_cast_fp16 = matmul(transpose_x = attn_weights_175_transpose_x_1, transpose_y = attn_weights_175_transpose_y_1, x = query_states_59_cast_fp16, y = key_states_119_cast_fp16)[name = tensor("attn_weights_175_cast_fp16")]; + tensor var_2706 = const()[name = tensor("op_2706"), val = tensor([1, 20, 77, 77])]; + tensor var_2707_cast_fp16 = reshape(shape = var_2706, x = attn_weights_175_cast_fp16)[name = tensor("op_2707_cast_fp16")]; + tensor attn_weights_177_cast_fp16 = add(x = var_2707_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_177_cast_fp16")]; + tensor var_2712 = const()[name = tensor("op_2712"), val = tensor([20, 77, 77])]; + tensor input_469_cast_fp16 = reshape(shape = var_2712, x = attn_weights_177_cast_fp16)[name = tensor("input_469_cast_fp16")]; + tensor input_471_cast_fp16 = softmax(axis = var_5, x = input_469_cast_fp16)[name = tensor("input_471_cast_fp16")]; + tensor attn_output_175_transpose_x_0 = const()[name = tensor("attn_output_175_transpose_x_0"), val = tensor(false)]; + tensor attn_output_175_transpose_y_0 = const()[name = tensor("attn_output_175_transpose_y_0"), val = tensor(false)]; + tensor attn_output_175_cast_fp16 = matmul(transpose_x = attn_output_175_transpose_x_0, transpose_y = attn_output_175_transpose_y_0, x = input_471_cast_fp16, y = value_states_119_cast_fp16)[name = tensor("attn_output_175_cast_fp16")]; + tensor var_2717 = const()[name = tensor("op_2717"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_177_cast_fp16 = reshape(shape = var_2717, x = attn_output_175_cast_fp16)[name = tensor("attn_output_177_cast_fp16")]; + tensor attn_output_179_perm_0 = const()[name = tensor("attn_output_179_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2720 = const()[name = tensor("op_2720"), val = tensor([1, 77, 1280])]; + tensor attn_output_179_cast_fp16 = transpose(perm = attn_output_179_perm_0, x = attn_output_177_cast_fp16)[name = tensor("transpose_8")]; + tensor input_473_cast_fp16 = reshape(shape = var_2720, x = attn_output_179_cast_fp16)[name = tensor("input_473_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558914432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560143296))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560143488)))]; + tensor linear_177_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized, x = input_473_cast_fp16)[name = tensor("linear_177_cast_fp16")]; + tensor input_475_cast_fp16 = add(x = input_467_cast_fp16, y = linear_177_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor input_477_axes_0 = const()[name = tensor("input_477_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560146112)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560148736)))]; + tensor input_477_cast_fp16 = layer_norm(axes = input_477_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("input_477_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(560151360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565066624))), name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565066816)))]; + tensor linear_178_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized, x = input_477_cast_fp16)[name = tensor("linear_178_cast_fp16")]; + tensor input_481_mode_0 = const()[name = tensor("input_481_mode_0"), val = tensor("EXACT")]; + tensor input_481_cast_fp16 = gelu(mode = input_481_mode_0, x = linear_178_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(565077120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569992384))), name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569992576)))]; + tensor linear_179_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized, x = input_481_cast_fp16)[name = tensor("linear_179_cast_fp16")]; + tensor input_483_cast_fp16 = add(x = input_475_cast_fp16, y = linear_179_cast_fp16)[name = tensor("input_483_cast_fp16")]; + tensor hidden_states_181_axes_0 = const()[name = tensor("hidden_states_181_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569995200)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569997824)))]; + tensor hidden_states_181_cast_fp16 = layer_norm(axes = hidden_states_181_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("hidden_states_181_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(570000448))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571229312))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571229504)))]; + tensor linear_180_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_180_cast_fp16")]; + tensor var_2759_to_fp16 = const()[name = tensor("op_2759_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_185_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_2759_to_fp16)[name = tensor("tensor_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571232128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572460992))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572461184)))]; + tensor linear_181_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_181_cast_fp16")]; + tensor var_2764 = const()[name = tensor("op_2764"), val = tensor([1, -1, 20, 64])]; + tensor var_2765_cast_fp16 = reshape(shape = var_2764, x = linear_181_cast_fp16)[name = tensor("op_2765_cast_fp16")]; + tensor var_2766_perm_0 = const()[name = tensor("op_2766_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572463808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573692672))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573692864)))]; + tensor linear_182_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_182_cast_fp16")]; + tensor var_2771 = const()[name = tensor("op_2771"), val = tensor([1, -1, 20, 64])]; + tensor var_2772_cast_fp16 = reshape(shape = var_2771, x = linear_182_cast_fp16)[name = tensor("op_2772_cast_fp16")]; + tensor var_2773_perm_0 = const()[name = tensor("op_2773_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2780 = const()[name = tensor("op_2780"), val = tensor([1, 77, 20, 64])]; + tensor var_2781_cast_fp16 = reshape(shape = var_2780, x = tensor_185_cast_fp16)[name = tensor("op_2781_cast_fp16")]; + tensor var_2782_perm_0 = const()[name = tensor("op_2782_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2784 = const()[name = tensor("op_2784"), val = tensor([20, -1, 64])]; + tensor var_2782_cast_fp16 = transpose(perm = var_2782_perm_0, x = var_2781_cast_fp16)[name = tensor("transpose_7")]; + tensor query_states_61_cast_fp16 = reshape(shape = var_2784, x = var_2782_cast_fp16)[name = tensor("query_states_61_cast_fp16")]; + tensor var_2786 = const()[name = tensor("op_2786"), val = tensor([20, -1, 64])]; + tensor var_2766_cast_fp16 = transpose(perm = var_2766_perm_0, x = var_2765_cast_fp16)[name = tensor("transpose_6")]; + tensor key_states_123_cast_fp16 = reshape(shape = var_2786, x = var_2766_cast_fp16)[name = tensor("key_states_123_cast_fp16")]; + tensor var_2788 = const()[name = tensor("op_2788"), val = tensor([20, -1, 64])]; + tensor var_2773_cast_fp16 = transpose(perm = var_2773_perm_0, x = var_2772_cast_fp16)[name = tensor("transpose_5")]; + tensor value_states_123_cast_fp16 = reshape(shape = var_2788, x = var_2773_cast_fp16)[name = tensor("value_states_123_cast_fp16")]; + tensor attn_weights_181_transpose_x_1 = const()[name = tensor("attn_weights_181_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_181_transpose_y_1 = const()[name = tensor("attn_weights_181_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_181_cast_fp16 = matmul(transpose_x = attn_weights_181_transpose_x_1, transpose_y = attn_weights_181_transpose_y_1, x = query_states_61_cast_fp16, y = key_states_123_cast_fp16)[name = tensor("attn_weights_181_cast_fp16")]; + tensor var_2793 = const()[name = tensor("op_2793"), val = tensor([1, 20, 77, 77])]; + tensor var_2794_cast_fp16 = reshape(shape = var_2793, x = attn_weights_181_cast_fp16)[name = tensor("op_2794_cast_fp16")]; + tensor attn_weights_183_cast_fp16 = add(x = var_2794_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_183_cast_fp16")]; + tensor var_2799 = const()[name = tensor("op_2799"), val = tensor([20, 77, 77])]; + tensor input_485_cast_fp16 = reshape(shape = var_2799, x = attn_weights_183_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor input_487_cast_fp16 = softmax(axis = var_5, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; + tensor attn_output_181_transpose_x_0 = const()[name = tensor("attn_output_181_transpose_x_0"), val = tensor(false)]; + tensor attn_output_181_transpose_y_0 = const()[name = tensor("attn_output_181_transpose_y_0"), val = tensor(false)]; + tensor attn_output_181_cast_fp16 = matmul(transpose_x = attn_output_181_transpose_x_0, transpose_y = attn_output_181_transpose_y_0, x = input_487_cast_fp16, y = value_states_123_cast_fp16)[name = tensor("attn_output_181_cast_fp16")]; + tensor var_2804 = const()[name = tensor("op_2804"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_183_cast_fp16 = reshape(shape = var_2804, x = attn_output_181_cast_fp16)[name = tensor("attn_output_183_cast_fp16")]; + tensor attn_output_185_perm_0 = const()[name = tensor("attn_output_185_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2807 = const()[name = tensor("op_2807"), val = tensor([1, 77, 1280])]; + tensor attn_output_185_cast_fp16 = transpose(perm = attn_output_185_perm_0, x = attn_output_183_cast_fp16)[name = tensor("transpose_4")]; + tensor input_489_cast_fp16 = reshape(shape = var_2807, x = attn_output_185_cast_fp16)[name = tensor("input_489_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573695488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574924352))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574924544)))]; + tensor linear_183_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized, x = input_489_cast_fp16)[name = tensor("linear_183_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_483_cast_fp16, y = linear_183_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574927168)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574929792)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574932416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579847680))), name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579847872)))]; + tensor linear_184_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized, x = input_493_cast_fp16)[name = tensor("linear_184_cast_fp16")]; + tensor input_497_mode_0 = const()[name = tensor("input_497_mode_0"), val = tensor("EXACT")]; + tensor input_497_cast_fp16 = gelu(mode = input_497_mode_0, x = linear_184_cast_fp16)[name = tensor("input_497_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579858176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584773440))), name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584773632)))]; + tensor linear_185_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized, x = input_497_cast_fp16)[name = tensor("linear_185_cast_fp16")]; + tensor input_499_cast_fp16 = add(x = input_491_cast_fp16, y = linear_185_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor input_499_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("input_499_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor hidden_states_187_axes_0 = const()[name = tensor("hidden_states_187_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584776256)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584778880)))]; + tensor hidden_states_187_cast_fp16 = layer_norm(axes = hidden_states_187_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("hidden_states_187_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584781504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(586010368))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(586010560)))]; + tensor linear_186_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_186_cast_fp16")]; + tensor var_2846_to_fp16 = const()[name = tensor("op_2846_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_cast_fp16 = mul(x = linear_186_cast_fp16, y = var_2846_to_fp16)[name = tensor("tensor_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(586013184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587242048))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587242240)))]; + tensor linear_187_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_187_cast_fp16")]; + tensor var_2851 = const()[name = tensor("op_2851"), val = tensor([1, -1, 20, 64])]; + tensor var_2852_cast_fp16 = reshape(shape = var_2851, x = linear_187_cast_fp16)[name = tensor("op_2852_cast_fp16")]; + tensor var_2853_perm_0 = const()[name = tensor("op_2853_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587244864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588473728))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588473920)))]; + tensor linear_188_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_188_cast_fp16")]; + tensor var_2858 = const()[name = tensor("op_2858"), val = tensor([1, -1, 20, 64])]; + tensor var_2859_cast_fp16 = reshape(shape = var_2858, x = linear_188_cast_fp16)[name = tensor("op_2859_cast_fp16")]; + tensor var_2860_perm_0 = const()[name = tensor("op_2860_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2867 = const()[name = tensor("op_2867"), val = tensor([1, 77, 20, 64])]; + tensor var_2868_cast_fp16 = reshape(shape = var_2867, x = tensor_cast_fp16)[name = tensor("op_2868_cast_fp16")]; + tensor var_2869_perm_0 = const()[name = tensor("op_2869_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2871 = const()[name = tensor("op_2871"), val = tensor([20, -1, 64])]; + tensor var_2869_cast_fp16 = transpose(perm = var_2869_perm_0, x = var_2868_cast_fp16)[name = tensor("transpose_3")]; + tensor query_states_cast_fp16 = reshape(shape = var_2871, x = var_2869_cast_fp16)[name = tensor("query_states_cast_fp16")]; + tensor var_2873 = const()[name = tensor("op_2873"), val = tensor([20, -1, 64])]; + tensor var_2853_cast_fp16 = transpose(perm = var_2853_perm_0, x = var_2852_cast_fp16)[name = tensor("transpose_2")]; + tensor key_states_cast_fp16 = reshape(shape = var_2873, x = var_2853_cast_fp16)[name = tensor("key_states_cast_fp16")]; + tensor var_2875 = const()[name = tensor("op_2875"), val = tensor([20, -1, 64])]; + tensor var_2860_cast_fp16 = transpose(perm = var_2860_perm_0, x = var_2859_cast_fp16)[name = tensor("transpose_1")]; + tensor value_states_cast_fp16 = reshape(shape = var_2875, x = var_2860_cast_fp16)[name = tensor("value_states_cast_fp16")]; + tensor attn_weights_187_transpose_x_1 = const()[name = tensor("attn_weights_187_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_187_transpose_y_1 = const()[name = tensor("attn_weights_187_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_187_cast_fp16 = matmul(transpose_x = attn_weights_187_transpose_x_1, transpose_y = attn_weights_187_transpose_y_1, x = query_states_cast_fp16, y = key_states_cast_fp16)[name = tensor("attn_weights_187_cast_fp16")]; + tensor var_2880 = const()[name = tensor("op_2880"), val = tensor([1, 20, 77, 77])]; + tensor var_2881_cast_fp16 = reshape(shape = var_2880, x = attn_weights_187_cast_fp16)[name = tensor("op_2881_cast_fp16")]; + tensor attn_weights_189_cast_fp16 = add(x = var_2881_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_189_cast_fp16")]; + tensor var_2886 = const()[name = tensor("op_2886"), val = tensor([20, 77, 77])]; + tensor input_501_cast_fp16 = reshape(shape = var_2886, x = attn_weights_189_cast_fp16)[name = tensor("input_501_cast_fp16")]; + tensor input_503_cast_fp16 = softmax(axis = var_5, x = input_501_cast_fp16)[name = tensor("input_503_cast_fp16")]; + tensor attn_output_187_transpose_x_0 = const()[name = tensor("attn_output_187_transpose_x_0"), val = tensor(false)]; + tensor attn_output_187_transpose_y_0 = const()[name = tensor("attn_output_187_transpose_y_0"), val = tensor(false)]; + tensor attn_output_187_cast_fp16 = matmul(transpose_x = attn_output_187_transpose_x_0, transpose_y = attn_output_187_transpose_y_0, x = input_503_cast_fp16, y = value_states_cast_fp16)[name = tensor("attn_output_187_cast_fp16")]; + tensor var_2891 = const()[name = tensor("op_2891"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_189_cast_fp16 = reshape(shape = var_2891, x = attn_output_187_cast_fp16)[name = tensor("attn_output_189_cast_fp16")]; + tensor attn_output_perm_0 = const()[name = tensor("attn_output_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2894 = const()[name = tensor("op_2894"), val = tensor([1, 77, 1280])]; + tensor attn_output_cast_fp16 = transpose(perm = attn_output_perm_0, x = attn_output_189_cast_fp16)[name = tensor("transpose_0")]; + tensor input_505_cast_fp16 = reshape(shape = var_2894, x = attn_output_cast_fp16)[name = tensor("input_505_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588476544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589705408))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589705600)))]; + tensor linear_189_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized, x = input_505_cast_fp16)[name = tensor("linear_189_cast_fp16")]; + tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_189_cast_fp16)[name = tensor("input_507_cast_fp16")]; + tensor input_509_axes_0 = const()[name = tensor("input_509_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589708224)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589710848)))]; + tensor input_509_cast_fp16 = layer_norm(axes = input_509_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("input_509_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589713472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594628736))), name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594628928)))]; + tensor linear_190_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized, x = input_509_cast_fp16)[name = tensor("linear_190_cast_fp16")]; + tensor input_513_mode_0 = const()[name = tensor("input_513_mode_0"), val = tensor("EXACT")]; + tensor input_513_cast_fp16 = gelu(mode = input_513_mode_0, x = linear_190_cast_fp16)[name = tensor("input_513_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594639232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599554496))), name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599554688)))]; + tensor linear_191_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized, x = input_513_cast_fp16)[name = tensor("linear_191_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_507_cast_fp16, y = linear_191_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor last_hidden_state_axes_0 = const()[name = tensor("last_hidden_state_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_final_layer_norm_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599557312)))]; + tensor text_encoder_text_model_final_layer_norm_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599559936)))]; + tensor last_hidden_state_cast_fp16 = layer_norm(axes = last_hidden_state_axes_0, beta = text_encoder_text_model_final_layer_norm_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_final_layer_norm_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("last_hidden_state_cast_fp16")]; + tensor var_2922 = const()[name = tensor("op_2922"), val = tensor([0])]; + tensor var_2924 = reduce_argmax(axis = var_5, keep_dims = var_6, x = cast_1)[name = tensor("op_2924")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (var_2922, var_2924))[name = tensor("stack_0")]; + tensor input_transpose_batch_dims_0 = const()[name = tensor("input_transpose_batch_dims_0"), val = tensor(0)]; + tensor input_transpose_cast_fp16 = gather_nd(batch_dims = input_transpose_batch_dims_0, indices = stack_0, x = last_hidden_state_cast_fp16)[name = tensor("input_transpose_cast_fp16")]; + tensor text_encoder_text_projection_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599562560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(600791424))), name = tensor("text_encoder_text_projection_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor linear_192_bias_0_to_fp16 = const()[name = tensor("linear_192_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(600791616)))]; + tensor linear_192_cast_fp16 = linear(bias = linear_192_bias_0_to_fp16, weight = text_encoder_text_projection_weight_to_fp16_palettized, x = input_transpose_cast_fp16)[name = tensor("linear_192_cast_fp16")]; + tensor linear_192_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("linear_192_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor hidden_embeds = cast(dtype = input_499_cast_fp16_to_fp32_dtype_0, x = input_499_cast_fp16)[name = tensor("cast_0")]; + tensor pooled_outputs = cast(dtype = linear_192_cast_fp16_to_fp32_dtype_0, x = linear_192_cast_fp16)[name = tensor("cast_1")]; + } -> (hidden_embeds, pooled_outputs); +} \ No newline at end of file