diff --git "a/text_encoder_2/coreml_model.mlmodelc/model.mil" "b/text_encoder_2/coreml_model.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/text_encoder_2/coreml_model.mlmodelc/model.mil" @@ -0,0 +1,2211 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3304.5.2"}, {"coremlc-version", "3304.6.2"}})] +{ + func main(tensor input_ids) { + tensor var_5 = const()[name = tensor("op_5"), val = tensor(-1)]; + tensor var_6 = const()[name = tensor("op_6"), val = tensor(false)]; + tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; + tensor inputs_embeds_axis_0 = const()[name = tensor("inputs_embeds_axis_0"), val = tensor(0)]; + tensor inputs_embeds_batch_dims_0 = const()[name = tensor("inputs_embeds_batch_dims_0"), val = tensor(0)]; + tensor text_encoder_text_model_embeddings_token_embedding_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_embeddings_token_embedding_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor cast_2 = cast(dtype = cast_1_dtype_0, x = input_ids)[name = tensor("cast_2")]; + tensor inputs_embeds_cast_fp16 = gather(axis = inputs_embeds_axis_0, batch_dims = inputs_embeds_batch_dims_0, indices = cast_2, x = text_encoder_text_model_embeddings_token_embedding_weight_to_fp16)[name = tensor("inputs_embeds_cast_fp16")]; + tensor position_embeddings_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126484608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126558592))), name = tensor("position_embeddings_to_fp16_palettized"), shape = tensor([1, 77, 1280])]; + tensor input_3_cast_fp16 = add(x = inputs_embeds_cast_fp16, y = position_embeddings_to_fp16_palettized)[name = tensor("input_3_cast_fp16")]; + tensor hidden_states_1_axes_0 = const()[name = tensor("hidden_states_1_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126558784)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126561408)))]; + tensor var_15_to_fp16 = const()[name = tensor("op_15_to_fp16"), val = tensor(0x1.5p-17)]; + tensor hidden_states_1_cast_fp16 = layer_norm(axes = hidden_states_1_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("hidden_states_1_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126564032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127792896))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127793088)))]; + tensor linear_0_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_148_to_fp16 = const()[name = tensor("op_148_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_5_cast_fp16 = mul(x = linear_0_cast_fp16, y = var_148_to_fp16)[name = tensor("tensor_5_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127795712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129024576))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129024768)))]; + tensor linear_1_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_153 = const()[name = tensor("op_153"), val = tensor([1, -1, 20, 64])]; + tensor var_154_cast_fp16 = reshape(shape = var_153, x = linear_1_cast_fp16)[name = tensor("op_154_cast_fp16")]; + tensor var_155_perm_0 = const()[name = tensor("op_155_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129027392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130256256))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130256448)))]; + tensor linear_2_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_1_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_160 = const()[name = tensor("op_160"), val = tensor([1, -1, 20, 64])]; + tensor var_161_cast_fp16 = reshape(shape = var_160, x = linear_2_cast_fp16)[name = tensor("op_161_cast_fp16")]; + tensor var_162_perm_0 = const()[name = tensor("op_162_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_169 = const()[name = tensor("op_169"), val = tensor([1, 77, 20, 64])]; + tensor var_170_cast_fp16 = reshape(shape = var_169, x = tensor_5_cast_fp16)[name = tensor("op_170_cast_fp16")]; + tensor var_171_perm_0 = const()[name = tensor("op_171_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_173 = const()[name = tensor("op_173"), val = tensor([20, -1, 64])]; + tensor transpose_127 = transpose(perm = var_171_perm_0, x = var_170_cast_fp16)[name = tensor("transpose_127")]; + tensor query_states_1_cast_fp16 = reshape(shape = var_173, x = transpose_127)[name = tensor("query_states_1_cast_fp16")]; + tensor var_175 = const()[name = tensor("op_175"), val = tensor([20, -1, 64])]; + tensor transpose_126 = transpose(perm = var_155_perm_0, x = var_154_cast_fp16)[name = tensor("transpose_126")]; + tensor key_states_3_cast_fp16 = reshape(shape = var_175, x = transpose_126)[name = tensor("key_states_3_cast_fp16")]; + tensor var_177 = const()[name = tensor("op_177"), val = tensor([20, -1, 64])]; + tensor transpose_125 = transpose(perm = var_162_perm_0, x = var_161_cast_fp16)[name = tensor("transpose_125")]; + tensor value_states_3_cast_fp16 = reshape(shape = var_177, x = transpose_125)[name = tensor("value_states_3_cast_fp16")]; + tensor attn_weights_1_transpose_x_1 = const()[name = tensor("attn_weights_1_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_1_transpose_y_1 = const()[name = tensor("attn_weights_1_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_1_cast_fp16 = matmul(transpose_x = attn_weights_1_transpose_x_1, transpose_y = attn_weights_1_transpose_y_1, x = query_states_1_cast_fp16, y = key_states_3_cast_fp16)[name = tensor("attn_weights_1_cast_fp16")]; + tensor var_182 = const()[name = tensor("op_182"), val = tensor([1, 20, 77, 77])]; + tensor var_183_cast_fp16 = reshape(shape = var_182, x = attn_weights_1_cast_fp16)[name = tensor("op_183_cast_fp16")]; + tensor op_58_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130259072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130263616))), name = tensor("op_58_to_fp16_palettized"), shape = tensor([1, 1, 77, 77])]; + tensor attn_weights_3_cast_fp16 = add(x = var_183_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_3_cast_fp16")]; + tensor var_188 = const()[name = tensor("op_188"), val = tensor([20, 77, 77])]; + tensor input_5_cast_fp16 = reshape(shape = var_188, x = attn_weights_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = softmax(axis = var_5, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor attn_output_1_transpose_x_0 = const()[name = tensor("attn_output_1_transpose_x_0"), val = tensor(false)]; + tensor attn_output_1_transpose_y_0 = const()[name = tensor("attn_output_1_transpose_y_0"), val = tensor(false)]; + tensor attn_output_1_cast_fp16 = matmul(transpose_x = attn_output_1_transpose_x_0, transpose_y = attn_output_1_transpose_y_0, x = input_7_cast_fp16, y = value_states_3_cast_fp16)[name = tensor("attn_output_1_cast_fp16")]; + tensor var_193 = const()[name = tensor("op_193"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_3_cast_fp16 = reshape(shape = var_193, x = attn_output_1_cast_fp16)[name = tensor("attn_output_3_cast_fp16")]; + tensor attn_output_5_perm_0 = const()[name = tensor("attn_output_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_196 = const()[name = tensor("op_196"), val = tensor([1, 77, 1280])]; + tensor transpose_124 = transpose(perm = attn_output_5_perm_0, x = attn_output_3_cast_fp16)[name = tensor("transpose_124")]; + tensor input_9_cast_fp16 = reshape(shape = var_196, x = transpose_124)[name = tensor("input_9_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130263808))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131492672))), name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131492864)))]; + tensor linear_3_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16_palettized, x = input_9_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor input_11_cast_fp16 = add(x = input_3_cast_fp16, y = linear_3_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_axes_0 = const()[name = tensor("input_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131495488)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131498112)))]; + tensor input_13_cast_fp16 = layer_norm(axes = input_13_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131500736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136416000))), name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136416192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136420096))), name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_4_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16_palettized, x = input_13_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("EXACT")]; + tensor input_17_cast_fp16 = gelu(mode = input_17_mode_0, x = linear_4_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136420288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141335552))), name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141335744)))]; + tensor linear_5_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16_palettized, x = input_17_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor input_19_cast_fp16 = add(x = input_11_cast_fp16, y = linear_5_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor hidden_states_7_axes_0 = const()[name = tensor("hidden_states_7_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141338368)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141340992)))]; + tensor hidden_states_7_cast_fp16 = layer_norm(axes = hidden_states_7_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("hidden_states_7_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141343616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142572480))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142572672)))]; + tensor linear_6_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_6_cast_fp16")]; + tensor var_235_to_fp16 = const()[name = tensor("op_235_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_11_cast_fp16 = mul(x = linear_6_cast_fp16, y = var_235_to_fp16)[name = tensor("tensor_11_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142575296))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143804160))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143804352)))]; + tensor linear_7_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor var_240 = const()[name = tensor("op_240"), val = tensor([1, -1, 20, 64])]; + tensor var_241_cast_fp16 = reshape(shape = var_240, x = linear_7_cast_fp16)[name = tensor("op_241_cast_fp16")]; + tensor var_242_perm_0 = const()[name = tensor("op_242_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143806976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145035840))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145036032)))]; + tensor linear_8_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_7_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor var_247 = const()[name = tensor("op_247"), val = tensor([1, -1, 20, 64])]; + tensor var_248_cast_fp16 = reshape(shape = var_247, x = linear_8_cast_fp16)[name = tensor("op_248_cast_fp16")]; + tensor var_249_perm_0 = const()[name = tensor("op_249_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_256 = const()[name = tensor("op_256"), val = tensor([1, 77, 20, 64])]; + tensor var_257_cast_fp16 = reshape(shape = var_256, x = tensor_11_cast_fp16)[name = tensor("op_257_cast_fp16")]; + tensor var_258_perm_0 = const()[name = tensor("op_258_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_260 = const()[name = tensor("op_260"), val = tensor([20, -1, 64])]; + tensor transpose_123 = transpose(perm = var_258_perm_0, x = var_257_cast_fp16)[name = tensor("transpose_123")]; + tensor query_states_3_cast_fp16 = reshape(shape = var_260, x = transpose_123)[name = tensor("query_states_3_cast_fp16")]; + tensor var_262 = const()[name = tensor("op_262"), val = tensor([20, -1, 64])]; + tensor transpose_122 = transpose(perm = var_242_perm_0, x = var_241_cast_fp16)[name = tensor("transpose_122")]; + tensor key_states_7_cast_fp16 = reshape(shape = var_262, x = transpose_122)[name = tensor("key_states_7_cast_fp16")]; + tensor var_264 = const()[name = tensor("op_264"), val = tensor([20, -1, 64])]; + tensor transpose_121 = transpose(perm = var_249_perm_0, x = var_248_cast_fp16)[name = tensor("transpose_121")]; + tensor value_states_7_cast_fp16 = reshape(shape = var_264, x = transpose_121)[name = tensor("value_states_7_cast_fp16")]; + tensor attn_weights_7_transpose_x_1 = const()[name = tensor("attn_weights_7_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_7_transpose_y_1 = const()[name = tensor("attn_weights_7_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_7_cast_fp16 = matmul(transpose_x = attn_weights_7_transpose_x_1, transpose_y = attn_weights_7_transpose_y_1, x = query_states_3_cast_fp16, y = key_states_7_cast_fp16)[name = tensor("attn_weights_7_cast_fp16")]; + tensor var_269 = const()[name = tensor("op_269"), val = tensor([1, 20, 77, 77])]; + tensor var_270_cast_fp16 = reshape(shape = var_269, x = attn_weights_7_cast_fp16)[name = tensor("op_270_cast_fp16")]; + tensor attn_weights_9_cast_fp16 = add(x = var_270_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_9_cast_fp16")]; + tensor var_275 = const()[name = tensor("op_275"), val = tensor([20, 77, 77])]; + tensor input_21_cast_fp16 = reshape(shape = var_275, x = attn_weights_9_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor input_23_cast_fp16 = softmax(axis = var_5, x = input_21_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor attn_output_7_transpose_x_0 = const()[name = tensor("attn_output_7_transpose_x_0"), val = tensor(false)]; + tensor attn_output_7_transpose_y_0 = const()[name = tensor("attn_output_7_transpose_y_0"), val = tensor(false)]; + tensor attn_output_7_cast_fp16 = matmul(transpose_x = attn_output_7_transpose_x_0, transpose_y = attn_output_7_transpose_y_0, x = input_23_cast_fp16, y = value_states_7_cast_fp16)[name = tensor("attn_output_7_cast_fp16")]; + tensor var_280 = const()[name = tensor("op_280"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_9_cast_fp16 = reshape(shape = var_280, x = attn_output_7_cast_fp16)[name = tensor("attn_output_9_cast_fp16")]; + tensor attn_output_11_perm_0 = const()[name = tensor("attn_output_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_283 = const()[name = tensor("op_283"), val = tensor([1, 77, 1280])]; + tensor transpose_120 = transpose(perm = attn_output_11_perm_0, x = attn_output_9_cast_fp16)[name = tensor("transpose_120")]; + tensor input_25_cast_fp16 = reshape(shape = var_283, x = transpose_120)[name = tensor("input_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145038656))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146267520))), name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146267712)))]; + tensor linear_9_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16_palettized, x = input_25_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor input_27_cast_fp16 = add(x = input_19_cast_fp16, y = linear_9_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor input_29_axes_0 = const()[name = tensor("input_29_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146270336)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146272960)))]; + tensor input_29_cast_fp16 = layer_norm(axes = input_29_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("input_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146275584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151190848))), name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151191040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151194944))), name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_10_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16_palettized, x = input_29_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_33_mode_0 = const()[name = tensor("input_33_mode_0"), val = tensor("EXACT")]; + tensor input_33_cast_fp16 = gelu(mode = input_33_mode_0, x = linear_10_cast_fp16)[name = tensor("input_33_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151195136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156110400))), name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156110592)))]; + tensor linear_11_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16_palettized, x = input_33_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor input_35_cast_fp16 = add(x = input_27_cast_fp16, y = linear_11_cast_fp16)[name = tensor("input_35_cast_fp16")]; + tensor hidden_states_13_axes_0 = const()[name = tensor("hidden_states_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156113216)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156115840)))]; + tensor hidden_states_13_cast_fp16 = layer_norm(axes = hidden_states_13_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("hidden_states_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156118464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157347328))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157347520)))]; + tensor linear_12_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_322_to_fp16 = const()[name = tensor("op_322_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_17_cast_fp16 = mul(x = linear_12_cast_fp16, y = var_322_to_fp16)[name = tensor("tensor_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157350144))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158579008))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158579200)))]; + tensor linear_13_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_327 = const()[name = tensor("op_327"), val = tensor([1, -1, 20, 64])]; + tensor var_328_cast_fp16 = reshape(shape = var_327, x = linear_13_cast_fp16)[name = tensor("op_328_cast_fp16")]; + tensor var_329_perm_0 = const()[name = tensor("op_329_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158581824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159810688))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159810880)))]; + tensor linear_14_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_13_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_334 = const()[name = tensor("op_334"), val = tensor([1, -1, 20, 64])]; + tensor var_335_cast_fp16 = reshape(shape = var_334, x = linear_14_cast_fp16)[name = tensor("op_335_cast_fp16")]; + tensor var_336_perm_0 = const()[name = tensor("op_336_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_343 = const()[name = tensor("op_343"), val = tensor([1, 77, 20, 64])]; + tensor var_344_cast_fp16 = reshape(shape = var_343, x = tensor_17_cast_fp16)[name = tensor("op_344_cast_fp16")]; + tensor var_345_perm_0 = const()[name = tensor("op_345_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_347 = const()[name = tensor("op_347"), val = tensor([20, -1, 64])]; + tensor transpose_119 = transpose(perm = var_345_perm_0, x = var_344_cast_fp16)[name = tensor("transpose_119")]; + tensor query_states_5_cast_fp16 = reshape(shape = var_347, x = transpose_119)[name = tensor("query_states_5_cast_fp16")]; + tensor var_349 = const()[name = tensor("op_349"), val = tensor([20, -1, 64])]; + tensor transpose_118 = transpose(perm = var_329_perm_0, x = var_328_cast_fp16)[name = tensor("transpose_118")]; + tensor key_states_11_cast_fp16 = reshape(shape = var_349, x = transpose_118)[name = tensor("key_states_11_cast_fp16")]; + tensor var_351 = const()[name = tensor("op_351"), val = tensor([20, -1, 64])]; + tensor transpose_117 = transpose(perm = var_336_perm_0, x = var_335_cast_fp16)[name = tensor("transpose_117")]; + tensor value_states_11_cast_fp16 = reshape(shape = var_351, x = transpose_117)[name = tensor("value_states_11_cast_fp16")]; + tensor attn_weights_13_transpose_x_1 = const()[name = tensor("attn_weights_13_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_13_transpose_y_1 = const()[name = tensor("attn_weights_13_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_13_cast_fp16 = matmul(transpose_x = attn_weights_13_transpose_x_1, transpose_y = attn_weights_13_transpose_y_1, x = query_states_5_cast_fp16, y = key_states_11_cast_fp16)[name = tensor("attn_weights_13_cast_fp16")]; + tensor var_356 = const()[name = tensor("op_356"), val = tensor([1, 20, 77, 77])]; + tensor var_357_cast_fp16 = reshape(shape = var_356, x = attn_weights_13_cast_fp16)[name = tensor("op_357_cast_fp16")]; + tensor attn_weights_15_cast_fp16 = add(x = var_357_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_15_cast_fp16")]; + tensor var_362 = const()[name = tensor("op_362"), val = tensor([20, 77, 77])]; + tensor input_37_cast_fp16 = reshape(shape = var_362, x = attn_weights_15_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor input_39_cast_fp16 = softmax(axis = var_5, x = input_37_cast_fp16)[name = tensor("input_39_cast_fp16")]; + tensor attn_output_13_transpose_x_0 = const()[name = tensor("attn_output_13_transpose_x_0"), val = tensor(false)]; + tensor attn_output_13_transpose_y_0 = const()[name = tensor("attn_output_13_transpose_y_0"), val = tensor(false)]; + tensor attn_output_13_cast_fp16 = matmul(transpose_x = attn_output_13_transpose_x_0, transpose_y = attn_output_13_transpose_y_0, x = input_39_cast_fp16, y = value_states_11_cast_fp16)[name = tensor("attn_output_13_cast_fp16")]; + tensor var_367 = const()[name = tensor("op_367"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_15_cast_fp16 = reshape(shape = var_367, x = attn_output_13_cast_fp16)[name = tensor("attn_output_15_cast_fp16")]; + tensor attn_output_17_perm_0 = const()[name = tensor("attn_output_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, 77, 1280])]; + tensor transpose_116 = transpose(perm = attn_output_17_perm_0, x = attn_output_15_cast_fp16)[name = tensor("transpose_116")]; + tensor input_41_cast_fp16 = reshape(shape = var_370, x = transpose_116)[name = tensor("input_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159813504))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161042368))), name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161042560)))]; + tensor linear_15_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16_palettized, x = input_41_cast_fp16)[name = tensor("linear_15_cast_fp16")]; + tensor input_43_cast_fp16 = add(x = input_35_cast_fp16, y = linear_15_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor input_45_axes_0 = const()[name = tensor("input_45_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161045184)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161047808)))]; + tensor input_45_cast_fp16 = layer_norm(axes = input_45_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("input_45_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161050432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165965696))), name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165965888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165969792))), name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_16_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16_palettized, x = input_45_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_49_mode_0 = const()[name = tensor("input_49_mode_0"), val = tensor("EXACT")]; + tensor input_49_cast_fp16 = gelu(mode = input_49_mode_0, x = linear_16_cast_fp16)[name = tensor("input_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165969984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170885248))), name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170885440)))]; + tensor linear_17_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16_palettized, x = input_49_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_51_cast_fp16 = add(x = input_43_cast_fp16, y = linear_17_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor hidden_states_19_axes_0 = const()[name = tensor("hidden_states_19_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170888064)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170890688)))]; + tensor hidden_states_19_cast_fp16 = layer_norm(axes = hidden_states_19_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16, x = input_51_cast_fp16)[name = tensor("hidden_states_19_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170893312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172122176))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172122368)))]; + tensor linear_18_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_409_to_fp16 = const()[name = tensor("op_409_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_23_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_409_to_fp16)[name = tensor("tensor_23_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172124992))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173353856))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173354048)))]; + tensor linear_19_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor var_414 = const()[name = tensor("op_414"), val = tensor([1, -1, 20, 64])]; + tensor var_415_cast_fp16 = reshape(shape = var_414, x = linear_19_cast_fp16)[name = tensor("op_415_cast_fp16")]; + tensor var_416_perm_0 = const()[name = tensor("op_416_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173356672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174585536))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174585728)))]; + tensor linear_20_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_19_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_421 = const()[name = tensor("op_421"), val = tensor([1, -1, 20, 64])]; + tensor var_422_cast_fp16 = reshape(shape = var_421, x = linear_20_cast_fp16)[name = tensor("op_422_cast_fp16")]; + tensor var_423_perm_0 = const()[name = tensor("op_423_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_430 = const()[name = tensor("op_430"), val = tensor([1, 77, 20, 64])]; + tensor var_431_cast_fp16 = reshape(shape = var_430, x = tensor_23_cast_fp16)[name = tensor("op_431_cast_fp16")]; + tensor var_432_perm_0 = const()[name = tensor("op_432_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_434 = const()[name = tensor("op_434"), val = tensor([20, -1, 64])]; + tensor transpose_115 = transpose(perm = var_432_perm_0, x = var_431_cast_fp16)[name = tensor("transpose_115")]; + tensor query_states_7_cast_fp16 = reshape(shape = var_434, x = transpose_115)[name = tensor("query_states_7_cast_fp16")]; + tensor var_436 = const()[name = tensor("op_436"), val = tensor([20, -1, 64])]; + tensor transpose_114 = transpose(perm = var_416_perm_0, x = var_415_cast_fp16)[name = tensor("transpose_114")]; + tensor key_states_15_cast_fp16 = reshape(shape = var_436, x = transpose_114)[name = tensor("key_states_15_cast_fp16")]; + tensor var_438 = const()[name = tensor("op_438"), val = tensor([20, -1, 64])]; + tensor transpose_113 = transpose(perm = var_423_perm_0, x = var_422_cast_fp16)[name = tensor("transpose_113")]; + tensor value_states_15_cast_fp16 = reshape(shape = var_438, x = transpose_113)[name = tensor("value_states_15_cast_fp16")]; + tensor attn_weights_19_transpose_x_1 = const()[name = tensor("attn_weights_19_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_19_transpose_y_1 = const()[name = tensor("attn_weights_19_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_19_cast_fp16 = matmul(transpose_x = attn_weights_19_transpose_x_1, transpose_y = attn_weights_19_transpose_y_1, x = query_states_7_cast_fp16, y = key_states_15_cast_fp16)[name = tensor("attn_weights_19_cast_fp16")]; + tensor var_443 = const()[name = tensor("op_443"), val = tensor([1, 20, 77, 77])]; + tensor var_444_cast_fp16 = reshape(shape = var_443, x = attn_weights_19_cast_fp16)[name = tensor("op_444_cast_fp16")]; + tensor attn_weights_21_cast_fp16 = add(x = var_444_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_21_cast_fp16")]; + tensor var_449 = const()[name = tensor("op_449"), val = tensor([20, 77, 77])]; + tensor input_53_cast_fp16 = reshape(shape = var_449, x = attn_weights_21_cast_fp16)[name = tensor("input_53_cast_fp16")]; + tensor input_55_cast_fp16 = softmax(axis = var_5, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; + tensor attn_output_19_transpose_x_0 = const()[name = tensor("attn_output_19_transpose_x_0"), val = tensor(false)]; + tensor attn_output_19_transpose_y_0 = const()[name = tensor("attn_output_19_transpose_y_0"), val = tensor(false)]; + tensor attn_output_19_cast_fp16 = matmul(transpose_x = attn_output_19_transpose_x_0, transpose_y = attn_output_19_transpose_y_0, x = input_55_cast_fp16, y = value_states_15_cast_fp16)[name = tensor("attn_output_19_cast_fp16")]; + tensor var_454 = const()[name = tensor("op_454"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_21_cast_fp16 = reshape(shape = var_454, x = attn_output_19_cast_fp16)[name = tensor("attn_output_21_cast_fp16")]; + tensor attn_output_23_perm_0 = const()[name = tensor("attn_output_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_457 = const()[name = tensor("op_457"), val = tensor([1, 77, 1280])]; + tensor transpose_112 = transpose(perm = attn_output_23_perm_0, x = attn_output_21_cast_fp16)[name = tensor("transpose_112")]; + tensor input_57_cast_fp16 = reshape(shape = var_457, x = transpose_112)[name = tensor("input_57_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174588352))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175817216))), name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175817408)))]; + tensor linear_21_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16_palettized, x = input_57_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor input_59_cast_fp16 = add(x = input_51_cast_fp16, y = linear_21_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor input_61_axes_0 = const()[name = tensor("input_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175820032)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175822656)))]; + tensor input_61_cast_fp16 = layer_norm(axes = input_61_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175825280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180740544))), name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180740736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180744640))), name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_22_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16_palettized, x = input_61_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor input_65_mode_0 = const()[name = tensor("input_65_mode_0"), val = tensor("EXACT")]; + tensor input_65_cast_fp16 = gelu(mode = input_65_mode_0, x = linear_22_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180744832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185660096))), name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185660288)))]; + tensor linear_23_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16_palettized, x = input_65_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor input_67_cast_fp16 = add(x = input_59_cast_fp16, y = linear_23_cast_fp16)[name = tensor("input_67_cast_fp16")]; + tensor hidden_states_25_axes_0 = const()[name = tensor("hidden_states_25_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185662912)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185665536)))]; + tensor hidden_states_25_cast_fp16 = layer_norm(axes = hidden_states_25_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("hidden_states_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185668160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186897024))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186897216)))]; + tensor linear_24_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_24_cast_fp16")]; + tensor var_496_to_fp16 = const()[name = tensor("op_496_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_29_cast_fp16 = mul(x = linear_24_cast_fp16, y = var_496_to_fp16)[name = tensor("tensor_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186899840))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188128704))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188128896)))]; + tensor linear_25_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor var_501 = const()[name = tensor("op_501"), val = tensor([1, -1, 20, 64])]; + tensor var_502_cast_fp16 = reshape(shape = var_501, x = linear_25_cast_fp16)[name = tensor("op_502_cast_fp16")]; + tensor var_503_perm_0 = const()[name = tensor("op_503_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188131520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189360384))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189360576)))]; + tensor linear_26_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_25_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor var_508 = const()[name = tensor("op_508"), val = tensor([1, -1, 20, 64])]; + tensor var_509_cast_fp16 = reshape(shape = var_508, x = linear_26_cast_fp16)[name = tensor("op_509_cast_fp16")]; + tensor var_510_perm_0 = const()[name = tensor("op_510_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_517 = const()[name = tensor("op_517"), val = tensor([1, 77, 20, 64])]; + tensor var_518_cast_fp16 = reshape(shape = var_517, x = tensor_29_cast_fp16)[name = tensor("op_518_cast_fp16")]; + tensor var_519_perm_0 = const()[name = tensor("op_519_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_521 = const()[name = tensor("op_521"), val = tensor([20, -1, 64])]; + tensor transpose_111 = transpose(perm = var_519_perm_0, x = var_518_cast_fp16)[name = tensor("transpose_111")]; + tensor query_states_9_cast_fp16 = reshape(shape = var_521, x = transpose_111)[name = tensor("query_states_9_cast_fp16")]; + tensor var_523 = const()[name = tensor("op_523"), val = tensor([20, -1, 64])]; + tensor transpose_110 = transpose(perm = var_503_perm_0, x = var_502_cast_fp16)[name = tensor("transpose_110")]; + tensor key_states_19_cast_fp16 = reshape(shape = var_523, x = transpose_110)[name = tensor("key_states_19_cast_fp16")]; + tensor var_525 = const()[name = tensor("op_525"), val = tensor([20, -1, 64])]; + tensor transpose_109 = transpose(perm = var_510_perm_0, x = var_509_cast_fp16)[name = tensor("transpose_109")]; + tensor value_states_19_cast_fp16 = reshape(shape = var_525, x = transpose_109)[name = tensor("value_states_19_cast_fp16")]; + tensor attn_weights_25_transpose_x_1 = const()[name = tensor("attn_weights_25_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_25_transpose_y_1 = const()[name = tensor("attn_weights_25_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_25_cast_fp16 = matmul(transpose_x = attn_weights_25_transpose_x_1, transpose_y = attn_weights_25_transpose_y_1, x = query_states_9_cast_fp16, y = key_states_19_cast_fp16)[name = tensor("attn_weights_25_cast_fp16")]; + tensor var_530 = const()[name = tensor("op_530"), val = tensor([1, 20, 77, 77])]; + tensor var_531_cast_fp16 = reshape(shape = var_530, x = attn_weights_25_cast_fp16)[name = tensor("op_531_cast_fp16")]; + tensor attn_weights_27_cast_fp16 = add(x = var_531_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_27_cast_fp16")]; + tensor var_536 = const()[name = tensor("op_536"), val = tensor([20, 77, 77])]; + tensor input_69_cast_fp16 = reshape(shape = var_536, x = attn_weights_27_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor input_71_cast_fp16 = softmax(axis = var_5, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; + tensor attn_output_25_transpose_x_0 = const()[name = tensor("attn_output_25_transpose_x_0"), val = tensor(false)]; + tensor attn_output_25_transpose_y_0 = const()[name = tensor("attn_output_25_transpose_y_0"), val = tensor(false)]; + tensor attn_output_25_cast_fp16 = matmul(transpose_x = attn_output_25_transpose_x_0, transpose_y = attn_output_25_transpose_y_0, x = input_71_cast_fp16, y = value_states_19_cast_fp16)[name = tensor("attn_output_25_cast_fp16")]; + tensor var_541 = const()[name = tensor("op_541"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_27_cast_fp16 = reshape(shape = var_541, x = attn_output_25_cast_fp16)[name = tensor("attn_output_27_cast_fp16")]; + tensor attn_output_29_perm_0 = const()[name = tensor("attn_output_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, 77, 1280])]; + tensor transpose_108 = transpose(perm = attn_output_29_perm_0, x = attn_output_27_cast_fp16)[name = tensor("transpose_108")]; + tensor input_73_cast_fp16 = reshape(shape = var_544, x = transpose_108)[name = tensor("input_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189363200))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190592064))), name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190592256)))]; + tensor linear_27_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16_palettized, x = input_73_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_67_cast_fp16, y = linear_27_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190594880)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190597504)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190600128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195515392))), name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195515584))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195519488))), name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_28_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16_palettized, x = input_77_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_81_mode_0 = const()[name = tensor("input_81_mode_0"), val = tensor("EXACT")]; + tensor input_81_cast_fp16 = gelu(mode = input_81_mode_0, x = linear_28_cast_fp16)[name = tensor("input_81_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195519680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200434944))), name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200435136)))]; + tensor linear_29_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16_palettized, x = input_81_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor input_83_cast_fp16 = add(x = input_75_cast_fp16, y = linear_29_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor hidden_states_31_axes_0 = const()[name = tensor("hidden_states_31_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200437760)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200440384)))]; + tensor hidden_states_31_cast_fp16 = layer_norm(axes = hidden_states_31_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("hidden_states_31_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200443008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201671872))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201672064)))]; + tensor linear_30_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_583_to_fp16 = const()[name = tensor("op_583_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_35_cast_fp16 = mul(x = linear_30_cast_fp16, y = var_583_to_fp16)[name = tensor("tensor_35_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201674688))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202903552))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202903744)))]; + tensor linear_31_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_588 = const()[name = tensor("op_588"), val = tensor([1, -1, 20, 64])]; + tensor var_589_cast_fp16 = reshape(shape = var_588, x = linear_31_cast_fp16)[name = tensor("op_589_cast_fp16")]; + tensor var_590_perm_0 = const()[name = tensor("op_590_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202906368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204135232))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204135424)))]; + tensor linear_32_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_31_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_595 = const()[name = tensor("op_595"), val = tensor([1, -1, 20, 64])]; + tensor var_596_cast_fp16 = reshape(shape = var_595, x = linear_32_cast_fp16)[name = tensor("op_596_cast_fp16")]; + tensor var_597_perm_0 = const()[name = tensor("op_597_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_604 = const()[name = tensor("op_604"), val = tensor([1, 77, 20, 64])]; + tensor var_605_cast_fp16 = reshape(shape = var_604, x = tensor_35_cast_fp16)[name = tensor("op_605_cast_fp16")]; + tensor var_606_perm_0 = const()[name = tensor("op_606_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_608 = const()[name = tensor("op_608"), val = tensor([20, -1, 64])]; + tensor transpose_107 = transpose(perm = var_606_perm_0, x = var_605_cast_fp16)[name = tensor("transpose_107")]; + tensor query_states_11_cast_fp16 = reshape(shape = var_608, x = transpose_107)[name = tensor("query_states_11_cast_fp16")]; + tensor var_610 = const()[name = tensor("op_610"), val = tensor([20, -1, 64])]; + tensor transpose_106 = transpose(perm = var_590_perm_0, x = var_589_cast_fp16)[name = tensor("transpose_106")]; + tensor key_states_23_cast_fp16 = reshape(shape = var_610, x = transpose_106)[name = tensor("key_states_23_cast_fp16")]; + tensor var_612 = const()[name = tensor("op_612"), val = tensor([20, -1, 64])]; + tensor transpose_105 = transpose(perm = var_597_perm_0, x = var_596_cast_fp16)[name = tensor("transpose_105")]; + tensor value_states_23_cast_fp16 = reshape(shape = var_612, x = transpose_105)[name = tensor("value_states_23_cast_fp16")]; + tensor attn_weights_31_transpose_x_1 = const()[name = tensor("attn_weights_31_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_31_transpose_y_1 = const()[name = tensor("attn_weights_31_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_31_cast_fp16 = matmul(transpose_x = attn_weights_31_transpose_x_1, transpose_y = attn_weights_31_transpose_y_1, x = query_states_11_cast_fp16, y = key_states_23_cast_fp16)[name = tensor("attn_weights_31_cast_fp16")]; + tensor var_617 = const()[name = tensor("op_617"), val = tensor([1, 20, 77, 77])]; + tensor var_618_cast_fp16 = reshape(shape = var_617, x = attn_weights_31_cast_fp16)[name = tensor("op_618_cast_fp16")]; + tensor attn_weights_33_cast_fp16 = add(x = var_618_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_33_cast_fp16")]; + tensor var_623 = const()[name = tensor("op_623"), val = tensor([20, 77, 77])]; + tensor input_85_cast_fp16 = reshape(shape = var_623, x = attn_weights_33_cast_fp16)[name = tensor("input_85_cast_fp16")]; + tensor input_87_cast_fp16 = softmax(axis = var_5, x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; + tensor attn_output_31_transpose_x_0 = const()[name = tensor("attn_output_31_transpose_x_0"), val = tensor(false)]; + tensor attn_output_31_transpose_y_0 = const()[name = tensor("attn_output_31_transpose_y_0"), val = tensor(false)]; + tensor attn_output_31_cast_fp16 = matmul(transpose_x = attn_output_31_transpose_x_0, transpose_y = attn_output_31_transpose_y_0, x = input_87_cast_fp16, y = value_states_23_cast_fp16)[name = tensor("attn_output_31_cast_fp16")]; + tensor var_628 = const()[name = tensor("op_628"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_33_cast_fp16 = reshape(shape = var_628, x = attn_output_31_cast_fp16)[name = tensor("attn_output_33_cast_fp16")]; + tensor attn_output_35_perm_0 = const()[name = tensor("attn_output_35_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_631 = const()[name = tensor("op_631"), val = tensor([1, 77, 1280])]; + tensor transpose_104 = transpose(perm = attn_output_35_perm_0, x = attn_output_33_cast_fp16)[name = tensor("transpose_104")]; + tensor input_89_cast_fp16 = reshape(shape = var_631, x = transpose_104)[name = tensor("input_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204138048))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205366912))), name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205367104)))]; + tensor linear_33_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16_palettized, x = input_89_cast_fp16)[name = tensor("linear_33_cast_fp16")]; + tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_33_cast_fp16)[name = tensor("input_91_cast_fp16")]; + tensor input_93_axes_0 = const()[name = tensor("input_93_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205369728)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205372352)))]; + tensor input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205374976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210290240))), name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210290432))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210294336))), name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_34_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16_palettized, x = input_93_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_97_mode_0 = const()[name = tensor("input_97_mode_0"), val = tensor("EXACT")]; + tensor input_97_cast_fp16 = gelu(mode = input_97_mode_0, x = linear_34_cast_fp16)[name = tensor("input_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210294528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215209792))), name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215209984)))]; + tensor linear_35_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16_palettized, x = input_97_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_35_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor hidden_states_37_axes_0 = const()[name = tensor("hidden_states_37_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215212608)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215215232)))]; + tensor hidden_states_37_cast_fp16 = layer_norm(axes = hidden_states_37_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("hidden_states_37_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215217856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216446720))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216446912)))]; + tensor linear_36_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_670_to_fp16 = const()[name = tensor("op_670_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_41_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_670_to_fp16)[name = tensor("tensor_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(216449536))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217678400))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217678592)))]; + tensor linear_37_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor var_675 = const()[name = tensor("op_675"), val = tensor([1, -1, 20, 64])]; + tensor var_676_cast_fp16 = reshape(shape = var_675, x = linear_37_cast_fp16)[name = tensor("op_676_cast_fp16")]; + tensor var_677_perm_0 = const()[name = tensor("op_677_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(217681216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218910080))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218910272)))]; + tensor linear_38_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_37_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_682 = const()[name = tensor("op_682"), val = tensor([1, -1, 20, 64])]; + tensor var_683_cast_fp16 = reshape(shape = var_682, x = linear_38_cast_fp16)[name = tensor("op_683_cast_fp16")]; + tensor var_684_perm_0 = const()[name = tensor("op_684_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_691 = const()[name = tensor("op_691"), val = tensor([1, 77, 20, 64])]; + tensor var_692_cast_fp16 = reshape(shape = var_691, x = tensor_41_cast_fp16)[name = tensor("op_692_cast_fp16")]; + tensor var_693_perm_0 = const()[name = tensor("op_693_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_695 = const()[name = tensor("op_695"), val = tensor([20, -1, 64])]; + tensor transpose_103 = transpose(perm = var_693_perm_0, x = var_692_cast_fp16)[name = tensor("transpose_103")]; + tensor query_states_13_cast_fp16 = reshape(shape = var_695, x = transpose_103)[name = tensor("query_states_13_cast_fp16")]; + tensor var_697 = const()[name = tensor("op_697"), val = tensor([20, -1, 64])]; + tensor transpose_102 = transpose(perm = var_677_perm_0, x = var_676_cast_fp16)[name = tensor("transpose_102")]; + tensor key_states_27_cast_fp16 = reshape(shape = var_697, x = transpose_102)[name = tensor("key_states_27_cast_fp16")]; + tensor var_699 = const()[name = tensor("op_699"), val = tensor([20, -1, 64])]; + tensor transpose_101 = transpose(perm = var_684_perm_0, x = var_683_cast_fp16)[name = tensor("transpose_101")]; + tensor value_states_27_cast_fp16 = reshape(shape = var_699, x = transpose_101)[name = tensor("value_states_27_cast_fp16")]; + tensor attn_weights_37_transpose_x_1 = const()[name = tensor("attn_weights_37_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_37_transpose_y_1 = const()[name = tensor("attn_weights_37_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_37_cast_fp16 = matmul(transpose_x = attn_weights_37_transpose_x_1, transpose_y = attn_weights_37_transpose_y_1, x = query_states_13_cast_fp16, y = key_states_27_cast_fp16)[name = tensor("attn_weights_37_cast_fp16")]; + tensor var_704 = const()[name = tensor("op_704"), val = tensor([1, 20, 77, 77])]; + tensor var_705_cast_fp16 = reshape(shape = var_704, x = attn_weights_37_cast_fp16)[name = tensor("op_705_cast_fp16")]; + tensor attn_weights_39_cast_fp16 = add(x = var_705_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_39_cast_fp16")]; + tensor var_710 = const()[name = tensor("op_710"), val = tensor([20, 77, 77])]; + tensor input_101_cast_fp16 = reshape(shape = var_710, x = attn_weights_39_cast_fp16)[name = tensor("input_101_cast_fp16")]; + tensor input_103_cast_fp16 = softmax(axis = var_5, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor attn_output_37_transpose_x_0 = const()[name = tensor("attn_output_37_transpose_x_0"), val = tensor(false)]; + tensor attn_output_37_transpose_y_0 = const()[name = tensor("attn_output_37_transpose_y_0"), val = tensor(false)]; + tensor attn_output_37_cast_fp16 = matmul(transpose_x = attn_output_37_transpose_x_0, transpose_y = attn_output_37_transpose_y_0, x = input_103_cast_fp16, y = value_states_27_cast_fp16)[name = tensor("attn_output_37_cast_fp16")]; + tensor var_715 = const()[name = tensor("op_715"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_39_cast_fp16 = reshape(shape = var_715, x = attn_output_37_cast_fp16)[name = tensor("attn_output_39_cast_fp16")]; + tensor attn_output_41_perm_0 = const()[name = tensor("attn_output_41_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, 77, 1280])]; + tensor transpose_100 = transpose(perm = attn_output_41_perm_0, x = attn_output_39_cast_fp16)[name = tensor("transpose_100")]; + tensor input_105_cast_fp16 = reshape(shape = var_718, x = transpose_100)[name = tensor("input_105_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218912896))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220141760))), name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220141952)))]; + tensor linear_39_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16_palettized, x = input_105_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor input_107_cast_fp16 = add(x = input_99_cast_fp16, y = linear_39_cast_fp16)[name = tensor("input_107_cast_fp16")]; + tensor input_109_axes_0 = const()[name = tensor("input_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220144576)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220147200)))]; + tensor input_109_cast_fp16 = layer_norm(axes = input_109_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(220149824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225065088))), name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225065280))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225069184))), name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_40_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16_palettized, x = input_109_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor input_113_mode_0 = const()[name = tensor("input_113_mode_0"), val = tensor("EXACT")]; + tensor input_113_cast_fp16 = gelu(mode = input_113_mode_0, x = linear_40_cast_fp16)[name = tensor("input_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(225069376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229984640))), name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229984832)))]; + tensor linear_41_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16_palettized, x = input_113_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor input_115_cast_fp16 = add(x = input_107_cast_fp16, y = linear_41_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor hidden_states_43_axes_0 = const()[name = tensor("hidden_states_43_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229987456)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229990080)))]; + tensor hidden_states_43_cast_fp16 = layer_norm(axes = hidden_states_43_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("hidden_states_43_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(229992704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231221568))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231221760)))]; + tensor linear_42_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_42_cast_fp16")]; + tensor var_757_to_fp16 = const()[name = tensor("op_757_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_47_cast_fp16 = mul(x = linear_42_cast_fp16, y = var_757_to_fp16)[name = tensor("tensor_47_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231224384))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232453248))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232453440)))]; + tensor linear_43_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor var_762 = const()[name = tensor("op_762"), val = tensor([1, -1, 20, 64])]; + tensor var_763_cast_fp16 = reshape(shape = var_762, x = linear_43_cast_fp16)[name = tensor("op_763_cast_fp16")]; + tensor var_764_perm_0 = const()[name = tensor("op_764_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232456064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233684928))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233685120)))]; + tensor linear_44_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_43_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor var_769 = const()[name = tensor("op_769"), val = tensor([1, -1, 20, 64])]; + tensor var_770_cast_fp16 = reshape(shape = var_769, x = linear_44_cast_fp16)[name = tensor("op_770_cast_fp16")]; + tensor var_771_perm_0 = const()[name = tensor("op_771_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_778 = const()[name = tensor("op_778"), val = tensor([1, 77, 20, 64])]; + tensor var_779_cast_fp16 = reshape(shape = var_778, x = tensor_47_cast_fp16)[name = tensor("op_779_cast_fp16")]; + tensor var_780_perm_0 = const()[name = tensor("op_780_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_782 = const()[name = tensor("op_782"), val = tensor([20, -1, 64])]; + tensor transpose_99 = transpose(perm = var_780_perm_0, x = var_779_cast_fp16)[name = tensor("transpose_99")]; + tensor query_states_15_cast_fp16 = reshape(shape = var_782, x = transpose_99)[name = tensor("query_states_15_cast_fp16")]; + tensor var_784 = const()[name = tensor("op_784"), val = tensor([20, -1, 64])]; + tensor transpose_98 = transpose(perm = var_764_perm_0, x = var_763_cast_fp16)[name = tensor("transpose_98")]; + tensor key_states_31_cast_fp16 = reshape(shape = var_784, x = transpose_98)[name = tensor("key_states_31_cast_fp16")]; + tensor var_786 = const()[name = tensor("op_786"), val = tensor([20, -1, 64])]; + tensor transpose_97 = transpose(perm = var_771_perm_0, x = var_770_cast_fp16)[name = tensor("transpose_97")]; + tensor value_states_31_cast_fp16 = reshape(shape = var_786, x = transpose_97)[name = tensor("value_states_31_cast_fp16")]; + tensor attn_weights_43_transpose_x_1 = const()[name = tensor("attn_weights_43_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_43_transpose_y_1 = const()[name = tensor("attn_weights_43_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_43_cast_fp16 = matmul(transpose_x = attn_weights_43_transpose_x_1, transpose_y = attn_weights_43_transpose_y_1, x = query_states_15_cast_fp16, y = key_states_31_cast_fp16)[name = tensor("attn_weights_43_cast_fp16")]; + tensor var_791 = const()[name = tensor("op_791"), val = tensor([1, 20, 77, 77])]; + tensor var_792_cast_fp16 = reshape(shape = var_791, x = attn_weights_43_cast_fp16)[name = tensor("op_792_cast_fp16")]; + tensor attn_weights_45_cast_fp16 = add(x = var_792_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_45_cast_fp16")]; + tensor var_797 = const()[name = tensor("op_797"), val = tensor([20, 77, 77])]; + tensor input_117_cast_fp16 = reshape(shape = var_797, x = attn_weights_45_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor input_119_cast_fp16 = softmax(axis = var_5, x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; + tensor attn_output_43_transpose_x_0 = const()[name = tensor("attn_output_43_transpose_x_0"), val = tensor(false)]; + tensor attn_output_43_transpose_y_0 = const()[name = tensor("attn_output_43_transpose_y_0"), val = tensor(false)]; + tensor attn_output_43_cast_fp16 = matmul(transpose_x = attn_output_43_transpose_x_0, transpose_y = attn_output_43_transpose_y_0, x = input_119_cast_fp16, y = value_states_31_cast_fp16)[name = tensor("attn_output_43_cast_fp16")]; + tensor var_802 = const()[name = tensor("op_802"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_45_cast_fp16 = reshape(shape = var_802, x = attn_output_43_cast_fp16)[name = tensor("attn_output_45_cast_fp16")]; + tensor attn_output_47_perm_0 = const()[name = tensor("attn_output_47_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_805 = const()[name = tensor("op_805"), val = tensor([1, 77, 1280])]; + tensor transpose_96 = transpose(perm = attn_output_47_perm_0, x = attn_output_45_cast_fp16)[name = tensor("transpose_96")]; + tensor input_121_cast_fp16 = reshape(shape = var_805, x = transpose_96)[name = tensor("input_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233687744))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234916608))), name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234916800)))]; + tensor linear_45_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16_palettized, x = input_121_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor input_123_cast_fp16 = add(x = input_115_cast_fp16, y = linear_45_cast_fp16)[name = tensor("input_123_cast_fp16")]; + tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234919424)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234922048)))]; + tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234924672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239839936))), name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239840128))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239844032))), name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_46_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16_palettized, x = input_125_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_129_mode_0 = const()[name = tensor("input_129_mode_0"), val = tensor("EXACT")]; + tensor input_129_cast_fp16 = gelu(mode = input_129_mode_0, x = linear_46_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(239844224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244759488))), name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244759680)))]; + tensor linear_47_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16_palettized, x = input_129_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor input_131_cast_fp16 = add(x = input_123_cast_fp16, y = linear_47_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor hidden_states_49_axes_0 = const()[name = tensor("hidden_states_49_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244762304)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244764928)))]; + tensor hidden_states_49_cast_fp16 = layer_norm(axes = hidden_states_49_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("hidden_states_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244767552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245996416))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245996608)))]; + tensor linear_48_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_844_to_fp16 = const()[name = tensor("op_844_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_53_cast_fp16 = mul(x = linear_48_cast_fp16, y = var_844_to_fp16)[name = tensor("tensor_53_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(245999232))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247228096))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247228288)))]; + tensor linear_49_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_849 = const()[name = tensor("op_849"), val = tensor([1, -1, 20, 64])]; + tensor var_850_cast_fp16 = reshape(shape = var_849, x = linear_49_cast_fp16)[name = tensor("op_850_cast_fp16")]; + tensor var_851_perm_0 = const()[name = tensor("op_851_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247230912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248459776))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248459968)))]; + tensor linear_50_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_49_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_856 = const()[name = tensor("op_856"), val = tensor([1, -1, 20, 64])]; + tensor var_857_cast_fp16 = reshape(shape = var_856, x = linear_50_cast_fp16)[name = tensor("op_857_cast_fp16")]; + tensor var_858_perm_0 = const()[name = tensor("op_858_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_865 = const()[name = tensor("op_865"), val = tensor([1, 77, 20, 64])]; + tensor var_866_cast_fp16 = reshape(shape = var_865, x = tensor_53_cast_fp16)[name = tensor("op_866_cast_fp16")]; + tensor var_867_perm_0 = const()[name = tensor("op_867_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_869 = const()[name = tensor("op_869"), val = tensor([20, -1, 64])]; + tensor transpose_95 = transpose(perm = var_867_perm_0, x = var_866_cast_fp16)[name = tensor("transpose_95")]; + tensor query_states_17_cast_fp16 = reshape(shape = var_869, x = transpose_95)[name = tensor("query_states_17_cast_fp16")]; + tensor var_871 = const()[name = tensor("op_871"), val = tensor([20, -1, 64])]; + tensor transpose_94 = transpose(perm = var_851_perm_0, x = var_850_cast_fp16)[name = tensor("transpose_94")]; + tensor key_states_35_cast_fp16 = reshape(shape = var_871, x = transpose_94)[name = tensor("key_states_35_cast_fp16")]; + tensor var_873 = const()[name = tensor("op_873"), val = tensor([20, -1, 64])]; + tensor transpose_93 = transpose(perm = var_858_perm_0, x = var_857_cast_fp16)[name = tensor("transpose_93")]; + tensor value_states_35_cast_fp16 = reshape(shape = var_873, x = transpose_93)[name = tensor("value_states_35_cast_fp16")]; + tensor attn_weights_49_transpose_x_1 = const()[name = tensor("attn_weights_49_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_49_transpose_y_1 = const()[name = tensor("attn_weights_49_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_49_cast_fp16 = matmul(transpose_x = attn_weights_49_transpose_x_1, transpose_y = attn_weights_49_transpose_y_1, x = query_states_17_cast_fp16, y = key_states_35_cast_fp16)[name = tensor("attn_weights_49_cast_fp16")]; + tensor var_878 = const()[name = tensor("op_878"), val = tensor([1, 20, 77, 77])]; + tensor var_879_cast_fp16 = reshape(shape = var_878, x = attn_weights_49_cast_fp16)[name = tensor("op_879_cast_fp16")]; + tensor attn_weights_51_cast_fp16 = add(x = var_879_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_51_cast_fp16")]; + tensor var_884 = const()[name = tensor("op_884"), val = tensor([20, 77, 77])]; + tensor input_133_cast_fp16 = reshape(shape = var_884, x = attn_weights_51_cast_fp16)[name = tensor("input_133_cast_fp16")]; + tensor input_135_cast_fp16 = softmax(axis = var_5, x = input_133_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor attn_output_49_transpose_x_0 = const()[name = tensor("attn_output_49_transpose_x_0"), val = tensor(false)]; + tensor attn_output_49_transpose_y_0 = const()[name = tensor("attn_output_49_transpose_y_0"), val = tensor(false)]; + tensor attn_output_49_cast_fp16 = matmul(transpose_x = attn_output_49_transpose_x_0, transpose_y = attn_output_49_transpose_y_0, x = input_135_cast_fp16, y = value_states_35_cast_fp16)[name = tensor("attn_output_49_cast_fp16")]; + tensor var_889 = const()[name = tensor("op_889"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_51_cast_fp16 = reshape(shape = var_889, x = attn_output_49_cast_fp16)[name = tensor("attn_output_51_cast_fp16")]; + tensor attn_output_53_perm_0 = const()[name = tensor("attn_output_53_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, 77, 1280])]; + tensor transpose_92 = transpose(perm = attn_output_53_perm_0, x = attn_output_51_cast_fp16)[name = tensor("transpose_92")]; + tensor input_137_cast_fp16 = reshape(shape = var_892, x = transpose_92)[name = tensor("input_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248462592))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249691456))), name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249691648)))]; + tensor linear_51_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16_palettized, x = input_137_cast_fp16)[name = tensor("linear_51_cast_fp16")]; + tensor input_139_cast_fp16 = add(x = input_131_cast_fp16, y = linear_51_cast_fp16)[name = tensor("input_139_cast_fp16")]; + tensor input_141_axes_0 = const()[name = tensor("input_141_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249694272)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249696896)))]; + tensor input_141_cast_fp16 = layer_norm(axes = input_141_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(249699520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254614784))), name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254614976))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254618880))), name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_52_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16_palettized, x = input_141_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_145_mode_0 = const()[name = tensor("input_145_mode_0"), val = tensor("EXACT")]; + tensor input_145_cast_fp16 = gelu(mode = input_145_mode_0, x = linear_52_cast_fp16)[name = tensor("input_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254619072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259534336))), name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259534528)))]; + tensor linear_53_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16_palettized, x = input_145_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_147_cast_fp16 = add(x = input_139_cast_fp16, y = linear_53_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor hidden_states_55_axes_0 = const()[name = tensor("hidden_states_55_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259537152)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259539776)))]; + tensor hidden_states_55_cast_fp16 = layer_norm(axes = hidden_states_55_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("hidden_states_55_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259542400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260771264))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260771456)))]; + tensor linear_54_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_931_to_fp16 = const()[name = tensor("op_931_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_59_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_931_to_fp16)[name = tensor("tensor_59_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260774080))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262002944))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262003136)))]; + tensor linear_55_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor var_936 = const()[name = tensor("op_936"), val = tensor([1, -1, 20, 64])]; + tensor var_937_cast_fp16 = reshape(shape = var_936, x = linear_55_cast_fp16)[name = tensor("op_937_cast_fp16")]; + tensor var_938_perm_0 = const()[name = tensor("op_938_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262005760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263234624))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263234816)))]; + tensor linear_56_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_55_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_943 = const()[name = tensor("op_943"), val = tensor([1, -1, 20, 64])]; + tensor var_944_cast_fp16 = reshape(shape = var_943, x = linear_56_cast_fp16)[name = tensor("op_944_cast_fp16")]; + tensor var_945_perm_0 = const()[name = tensor("op_945_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_952 = const()[name = tensor("op_952"), val = tensor([1, 77, 20, 64])]; + tensor var_953_cast_fp16 = reshape(shape = var_952, x = tensor_59_cast_fp16)[name = tensor("op_953_cast_fp16")]; + tensor var_954_perm_0 = const()[name = tensor("op_954_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_956 = const()[name = tensor("op_956"), val = tensor([20, -1, 64])]; + tensor transpose_91 = transpose(perm = var_954_perm_0, x = var_953_cast_fp16)[name = tensor("transpose_91")]; + tensor query_states_19_cast_fp16 = reshape(shape = var_956, x = transpose_91)[name = tensor("query_states_19_cast_fp16")]; + tensor var_958 = const()[name = tensor("op_958"), val = tensor([20, -1, 64])]; + tensor transpose_90 = transpose(perm = var_938_perm_0, x = var_937_cast_fp16)[name = tensor("transpose_90")]; + tensor key_states_39_cast_fp16 = reshape(shape = var_958, x = transpose_90)[name = tensor("key_states_39_cast_fp16")]; + tensor var_960 = const()[name = tensor("op_960"), val = tensor([20, -1, 64])]; + tensor transpose_89 = transpose(perm = var_945_perm_0, x = var_944_cast_fp16)[name = tensor("transpose_89")]; + tensor value_states_39_cast_fp16 = reshape(shape = var_960, x = transpose_89)[name = tensor("value_states_39_cast_fp16")]; + tensor attn_weights_55_transpose_x_1 = const()[name = tensor("attn_weights_55_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_55_transpose_y_1 = const()[name = tensor("attn_weights_55_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_55_cast_fp16 = matmul(transpose_x = attn_weights_55_transpose_x_1, transpose_y = attn_weights_55_transpose_y_1, x = query_states_19_cast_fp16, y = key_states_39_cast_fp16)[name = tensor("attn_weights_55_cast_fp16")]; + tensor var_965 = const()[name = tensor("op_965"), val = tensor([1, 20, 77, 77])]; + tensor var_966_cast_fp16 = reshape(shape = var_965, x = attn_weights_55_cast_fp16)[name = tensor("op_966_cast_fp16")]; + tensor attn_weights_57_cast_fp16 = add(x = var_966_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_57_cast_fp16")]; + tensor var_971 = const()[name = tensor("op_971"), val = tensor([20, 77, 77])]; + tensor input_149_cast_fp16 = reshape(shape = var_971, x = attn_weights_57_cast_fp16)[name = tensor("input_149_cast_fp16")]; + tensor input_151_cast_fp16 = softmax(axis = var_5, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor attn_output_55_transpose_x_0 = const()[name = tensor("attn_output_55_transpose_x_0"), val = tensor(false)]; + tensor attn_output_55_transpose_y_0 = const()[name = tensor("attn_output_55_transpose_y_0"), val = tensor(false)]; + tensor attn_output_55_cast_fp16 = matmul(transpose_x = attn_output_55_transpose_x_0, transpose_y = attn_output_55_transpose_y_0, x = input_151_cast_fp16, y = value_states_39_cast_fp16)[name = tensor("attn_output_55_cast_fp16")]; + tensor var_976 = const()[name = tensor("op_976"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_57_cast_fp16 = reshape(shape = var_976, x = attn_output_55_cast_fp16)[name = tensor("attn_output_57_cast_fp16")]; + tensor attn_output_59_perm_0 = const()[name = tensor("attn_output_59_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_979 = const()[name = tensor("op_979"), val = tensor([1, 77, 1280])]; + tensor transpose_88 = transpose(perm = attn_output_59_perm_0, x = attn_output_57_cast_fp16)[name = tensor("transpose_88")]; + tensor input_153_cast_fp16 = reshape(shape = var_979, x = transpose_88)[name = tensor("input_153_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263237440))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264466304))), name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264466496)))]; + tensor linear_57_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16_palettized, x = input_153_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor input_155_cast_fp16 = add(x = input_147_cast_fp16, y = linear_57_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor input_157_axes_0 = const()[name = tensor("input_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264469120)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264471744)))]; + tensor input_157_cast_fp16 = layer_norm(axes = input_157_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16, x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(264474368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269389632))), name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269389824))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269393728))), name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_58_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16_palettized, x = input_157_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor input_161_mode_0 = const()[name = tensor("input_161_mode_0"), val = tensor("EXACT")]; + tensor input_161_cast_fp16 = gelu(mode = input_161_mode_0, x = linear_58_cast_fp16)[name = tensor("input_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(269393920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274309184))), name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274309376)))]; + tensor linear_59_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16_palettized, x = input_161_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor input_163_cast_fp16 = add(x = input_155_cast_fp16, y = linear_59_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor hidden_states_61_axes_0 = const()[name = tensor("hidden_states_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274312000)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274314624)))]; + tensor hidden_states_61_cast_fp16 = layer_norm(axes = hidden_states_61_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("hidden_states_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(274317248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275546112))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275546304)))]; + tensor linear_60_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_60_cast_fp16")]; + tensor var_1018_to_fp16 = const()[name = tensor("op_1018_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_65_cast_fp16 = mul(x = linear_60_cast_fp16, y = var_1018_to_fp16)[name = tensor("tensor_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275548928))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276777792))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276777984)))]; + tensor linear_61_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor var_1023 = const()[name = tensor("op_1023"), val = tensor([1, -1, 20, 64])]; + tensor var_1024_cast_fp16 = reshape(shape = var_1023, x = linear_61_cast_fp16)[name = tensor("op_1024_cast_fp16")]; + tensor var_1025_perm_0 = const()[name = tensor("op_1025_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(276780608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278009472))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278009664)))]; + tensor linear_62_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_61_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor var_1030 = const()[name = tensor("op_1030"), val = tensor([1, -1, 20, 64])]; + tensor var_1031_cast_fp16 = reshape(shape = var_1030, x = linear_62_cast_fp16)[name = tensor("op_1031_cast_fp16")]; + tensor var_1032_perm_0 = const()[name = tensor("op_1032_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1039 = const()[name = tensor("op_1039"), val = tensor([1, 77, 20, 64])]; + tensor var_1040_cast_fp16 = reshape(shape = var_1039, x = tensor_65_cast_fp16)[name = tensor("op_1040_cast_fp16")]; + tensor var_1041_perm_0 = const()[name = tensor("op_1041_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1043 = const()[name = tensor("op_1043"), val = tensor([20, -1, 64])]; + tensor transpose_87 = transpose(perm = var_1041_perm_0, x = var_1040_cast_fp16)[name = tensor("transpose_87")]; + tensor query_states_21_cast_fp16 = reshape(shape = var_1043, x = transpose_87)[name = tensor("query_states_21_cast_fp16")]; + tensor var_1045 = const()[name = tensor("op_1045"), val = tensor([20, -1, 64])]; + tensor transpose_86 = transpose(perm = var_1025_perm_0, x = var_1024_cast_fp16)[name = tensor("transpose_86")]; + tensor key_states_43_cast_fp16 = reshape(shape = var_1045, x = transpose_86)[name = tensor("key_states_43_cast_fp16")]; + tensor var_1047 = const()[name = tensor("op_1047"), val = tensor([20, -1, 64])]; + tensor transpose_85 = transpose(perm = var_1032_perm_0, x = var_1031_cast_fp16)[name = tensor("transpose_85")]; + tensor value_states_43_cast_fp16 = reshape(shape = var_1047, x = transpose_85)[name = tensor("value_states_43_cast_fp16")]; + tensor attn_weights_61_transpose_x_1 = const()[name = tensor("attn_weights_61_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_61_transpose_y_1 = const()[name = tensor("attn_weights_61_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_61_cast_fp16 = matmul(transpose_x = attn_weights_61_transpose_x_1, transpose_y = attn_weights_61_transpose_y_1, x = query_states_21_cast_fp16, y = key_states_43_cast_fp16)[name = tensor("attn_weights_61_cast_fp16")]; + tensor var_1052 = const()[name = tensor("op_1052"), val = tensor([1, 20, 77, 77])]; + tensor var_1053_cast_fp16 = reshape(shape = var_1052, x = attn_weights_61_cast_fp16)[name = tensor("op_1053_cast_fp16")]; + tensor attn_weights_63_cast_fp16 = add(x = var_1053_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_63_cast_fp16")]; + tensor var_1058 = const()[name = tensor("op_1058"), val = tensor([20, 77, 77])]; + tensor input_165_cast_fp16 = reshape(shape = var_1058, x = attn_weights_63_cast_fp16)[name = tensor("input_165_cast_fp16")]; + tensor input_167_cast_fp16 = softmax(axis = var_5, x = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor attn_output_61_transpose_x_0 = const()[name = tensor("attn_output_61_transpose_x_0"), val = tensor(false)]; + tensor attn_output_61_transpose_y_0 = const()[name = tensor("attn_output_61_transpose_y_0"), val = tensor(false)]; + tensor attn_output_61_cast_fp16 = matmul(transpose_x = attn_output_61_transpose_x_0, transpose_y = attn_output_61_transpose_y_0, x = input_167_cast_fp16, y = value_states_43_cast_fp16)[name = tensor("attn_output_61_cast_fp16")]; + tensor var_1063 = const()[name = tensor("op_1063"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_63_cast_fp16 = reshape(shape = var_1063, x = attn_output_61_cast_fp16)[name = tensor("attn_output_63_cast_fp16")]; + tensor attn_output_65_perm_0 = const()[name = tensor("attn_output_65_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, 77, 1280])]; + tensor transpose_84 = transpose(perm = attn_output_65_perm_0, x = attn_output_63_cast_fp16)[name = tensor("transpose_84")]; + tensor input_169_cast_fp16 = reshape(shape = var_1066, x = transpose_84)[name = tensor("input_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278012288))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279241152))), name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279241344)))]; + tensor linear_63_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16_palettized, x = input_169_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor input_171_cast_fp16 = add(x = input_163_cast_fp16, y = linear_63_cast_fp16)[name = tensor("input_171_cast_fp16")]; + tensor input_173_axes_0 = const()[name = tensor("input_173_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279243968)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279246592)))]; + tensor input_173_cast_fp16 = layer_norm(axes = input_173_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279249216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284164480))), name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284164672))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284168576))), name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_64_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16_palettized, x = input_173_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_177_mode_0 = const()[name = tensor("input_177_mode_0"), val = tensor("EXACT")]; + tensor input_177_cast_fp16 = gelu(mode = input_177_mode_0, x = linear_64_cast_fp16)[name = tensor("input_177_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284168768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289084032))), name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289084224)))]; + tensor linear_65_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16_palettized, x = input_177_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_171_cast_fp16, y = linear_65_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor hidden_states_67_axes_0 = const()[name = tensor("hidden_states_67_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289086848)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289089472)))]; + tensor hidden_states_67_cast_fp16 = layer_norm(axes = hidden_states_67_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("hidden_states_67_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(289092096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290320960))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290321152)))]; + tensor linear_66_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1105_to_fp16 = const()[name = tensor("op_1105_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_71_cast_fp16 = mul(x = linear_66_cast_fp16, y = var_1105_to_fp16)[name = tensor("tensor_71_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290323776))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291552640))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291552832)))]; + tensor linear_67_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1110 = const()[name = tensor("op_1110"), val = tensor([1, -1, 20, 64])]; + tensor var_1111_cast_fp16 = reshape(shape = var_1110, x = linear_67_cast_fp16)[name = tensor("op_1111_cast_fp16")]; + tensor var_1112_perm_0 = const()[name = tensor("op_1112_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291555456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292784320))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292784512)))]; + tensor linear_68_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_67_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1117 = const()[name = tensor("op_1117"), val = tensor([1, -1, 20, 64])]; + tensor var_1118_cast_fp16 = reshape(shape = var_1117, x = linear_68_cast_fp16)[name = tensor("op_1118_cast_fp16")]; + tensor var_1119_perm_0 = const()[name = tensor("op_1119_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1126 = const()[name = tensor("op_1126"), val = tensor([1, 77, 20, 64])]; + tensor var_1127_cast_fp16 = reshape(shape = var_1126, x = tensor_71_cast_fp16)[name = tensor("op_1127_cast_fp16")]; + tensor var_1128_perm_0 = const()[name = tensor("op_1128_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1130 = const()[name = tensor("op_1130"), val = tensor([20, -1, 64])]; + tensor transpose_83 = transpose(perm = var_1128_perm_0, x = var_1127_cast_fp16)[name = tensor("transpose_83")]; + tensor query_states_23_cast_fp16 = reshape(shape = var_1130, x = transpose_83)[name = tensor("query_states_23_cast_fp16")]; + tensor var_1132 = const()[name = tensor("op_1132"), val = tensor([20, -1, 64])]; + tensor transpose_82 = transpose(perm = var_1112_perm_0, x = var_1111_cast_fp16)[name = tensor("transpose_82")]; + tensor key_states_47_cast_fp16 = reshape(shape = var_1132, x = transpose_82)[name = tensor("key_states_47_cast_fp16")]; + tensor var_1134 = const()[name = tensor("op_1134"), val = tensor([20, -1, 64])]; + tensor transpose_81 = transpose(perm = var_1119_perm_0, x = var_1118_cast_fp16)[name = tensor("transpose_81")]; + tensor value_states_47_cast_fp16 = reshape(shape = var_1134, x = transpose_81)[name = tensor("value_states_47_cast_fp16")]; + tensor attn_weights_67_transpose_x_1 = const()[name = tensor("attn_weights_67_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_67_transpose_y_1 = const()[name = tensor("attn_weights_67_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_67_cast_fp16 = matmul(transpose_x = attn_weights_67_transpose_x_1, transpose_y = attn_weights_67_transpose_y_1, x = query_states_23_cast_fp16, y = key_states_47_cast_fp16)[name = tensor("attn_weights_67_cast_fp16")]; + tensor var_1139 = const()[name = tensor("op_1139"), val = tensor([1, 20, 77, 77])]; + tensor var_1140_cast_fp16 = reshape(shape = var_1139, x = attn_weights_67_cast_fp16)[name = tensor("op_1140_cast_fp16")]; + tensor attn_weights_69_cast_fp16 = add(x = var_1140_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_69_cast_fp16")]; + tensor var_1145 = const()[name = tensor("op_1145"), val = tensor([20, 77, 77])]; + tensor input_181_cast_fp16 = reshape(shape = var_1145, x = attn_weights_69_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor input_183_cast_fp16 = softmax(axis = var_5, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor attn_output_67_transpose_x_0 = const()[name = tensor("attn_output_67_transpose_x_0"), val = tensor(false)]; + tensor attn_output_67_transpose_y_0 = const()[name = tensor("attn_output_67_transpose_y_0"), val = tensor(false)]; + tensor attn_output_67_cast_fp16 = matmul(transpose_x = attn_output_67_transpose_x_0, transpose_y = attn_output_67_transpose_y_0, x = input_183_cast_fp16, y = value_states_47_cast_fp16)[name = tensor("attn_output_67_cast_fp16")]; + tensor var_1150 = const()[name = tensor("op_1150"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_69_cast_fp16 = reshape(shape = var_1150, x = attn_output_67_cast_fp16)[name = tensor("attn_output_69_cast_fp16")]; + tensor attn_output_71_perm_0 = const()[name = tensor("attn_output_71_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1153 = const()[name = tensor("op_1153"), val = tensor([1, 77, 1280])]; + tensor transpose_80 = transpose(perm = attn_output_71_perm_0, x = attn_output_69_cast_fp16)[name = tensor("transpose_80")]; + tensor input_185_cast_fp16 = reshape(shape = var_1153, x = transpose_80)[name = tensor("input_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(292787136))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294016000))), name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294016192)))]; + tensor linear_69_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16_palettized, x = input_185_cast_fp16)[name = tensor("linear_69_cast_fp16")]; + tensor input_187_cast_fp16 = add(x = input_179_cast_fp16, y = linear_69_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor input_189_axes_0 = const()[name = tensor("input_189_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294018816)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294021440)))]; + tensor input_189_cast_fp16 = layer_norm(axes = input_189_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("input_189_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(294024064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298939328))), name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298939520))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298943424))), name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_70_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16_palettized, x = input_189_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_193_mode_0 = const()[name = tensor("input_193_mode_0"), val = tensor("EXACT")]; + tensor input_193_cast_fp16 = gelu(mode = input_193_mode_0, x = linear_70_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(298943616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303858880))), name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303859072)))]; + tensor linear_71_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16_palettized, x = input_193_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_71_cast_fp16)[name = tensor("input_195_cast_fp16")]; + tensor hidden_states_73_axes_0 = const()[name = tensor("hidden_states_73_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303861696)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303864320)))]; + tensor hidden_states_73_cast_fp16 = layer_norm(axes = hidden_states_73_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("hidden_states_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(303866944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305095808))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305096000)))]; + tensor linear_72_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1192_to_fp16 = const()[name = tensor("op_1192_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_77_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1192_to_fp16)[name = tensor("tensor_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305098624))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306327488))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306327680)))]; + tensor linear_73_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor var_1197 = const()[name = tensor("op_1197"), val = tensor([1, -1, 20, 64])]; + tensor var_1198_cast_fp16 = reshape(shape = var_1197, x = linear_73_cast_fp16)[name = tensor("op_1198_cast_fp16")]; + tensor var_1199_perm_0 = const()[name = tensor("op_1199_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306330304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307559168))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307559360)))]; + tensor linear_74_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_73_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1204 = const()[name = tensor("op_1204"), val = tensor([1, -1, 20, 64])]; + tensor var_1205_cast_fp16 = reshape(shape = var_1204, x = linear_74_cast_fp16)[name = tensor("op_1205_cast_fp16")]; + tensor var_1206_perm_0 = const()[name = tensor("op_1206_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1213 = const()[name = tensor("op_1213"), val = tensor([1, 77, 20, 64])]; + tensor var_1214_cast_fp16 = reshape(shape = var_1213, x = tensor_77_cast_fp16)[name = tensor("op_1214_cast_fp16")]; + tensor var_1215_perm_0 = const()[name = tensor("op_1215_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1217 = const()[name = tensor("op_1217"), val = tensor([20, -1, 64])]; + tensor transpose_79 = transpose(perm = var_1215_perm_0, x = var_1214_cast_fp16)[name = tensor("transpose_79")]; + tensor query_states_25_cast_fp16 = reshape(shape = var_1217, x = transpose_79)[name = tensor("query_states_25_cast_fp16")]; + tensor var_1219 = const()[name = tensor("op_1219"), val = tensor([20, -1, 64])]; + tensor transpose_78 = transpose(perm = var_1199_perm_0, x = var_1198_cast_fp16)[name = tensor("transpose_78")]; + tensor key_states_51_cast_fp16 = reshape(shape = var_1219, x = transpose_78)[name = tensor("key_states_51_cast_fp16")]; + tensor var_1221 = const()[name = tensor("op_1221"), val = tensor([20, -1, 64])]; + tensor transpose_77 = transpose(perm = var_1206_perm_0, x = var_1205_cast_fp16)[name = tensor("transpose_77")]; + tensor value_states_51_cast_fp16 = reshape(shape = var_1221, x = transpose_77)[name = tensor("value_states_51_cast_fp16")]; + tensor attn_weights_73_transpose_x_1 = const()[name = tensor("attn_weights_73_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_73_transpose_y_1 = const()[name = tensor("attn_weights_73_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_73_cast_fp16 = matmul(transpose_x = attn_weights_73_transpose_x_1, transpose_y = attn_weights_73_transpose_y_1, x = query_states_25_cast_fp16, y = key_states_51_cast_fp16)[name = tensor("attn_weights_73_cast_fp16")]; + tensor var_1226 = const()[name = tensor("op_1226"), val = tensor([1, 20, 77, 77])]; + tensor var_1227_cast_fp16 = reshape(shape = var_1226, x = attn_weights_73_cast_fp16)[name = tensor("op_1227_cast_fp16")]; + tensor attn_weights_75_cast_fp16 = add(x = var_1227_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_75_cast_fp16")]; + tensor var_1232 = const()[name = tensor("op_1232"), val = tensor([20, 77, 77])]; + tensor input_197_cast_fp16 = reshape(shape = var_1232, x = attn_weights_75_cast_fp16)[name = tensor("input_197_cast_fp16")]; + tensor input_199_cast_fp16 = softmax(axis = var_5, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor attn_output_73_transpose_x_0 = const()[name = tensor("attn_output_73_transpose_x_0"), val = tensor(false)]; + tensor attn_output_73_transpose_y_0 = const()[name = tensor("attn_output_73_transpose_y_0"), val = tensor(false)]; + tensor attn_output_73_cast_fp16 = matmul(transpose_x = attn_output_73_transpose_x_0, transpose_y = attn_output_73_transpose_y_0, x = input_199_cast_fp16, y = value_states_51_cast_fp16)[name = tensor("attn_output_73_cast_fp16")]; + tensor var_1237 = const()[name = tensor("op_1237"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_75_cast_fp16 = reshape(shape = var_1237, x = attn_output_73_cast_fp16)[name = tensor("attn_output_75_cast_fp16")]; + tensor attn_output_77_perm_0 = const()[name = tensor("attn_output_77_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, 77, 1280])]; + tensor transpose_76 = transpose(perm = attn_output_77_perm_0, x = attn_output_75_cast_fp16)[name = tensor("transpose_76")]; + tensor input_201_cast_fp16 = reshape(shape = var_1240, x = transpose_76)[name = tensor("input_201_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307561984))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308790848))), name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308791040)))]; + tensor linear_75_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16_palettized, x = input_201_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_195_cast_fp16, y = linear_75_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor input_205_axes_0 = const()[name = tensor("input_205_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308793664)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308796288)))]; + tensor input_205_cast_fp16 = layer_norm(axes = input_205_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("input_205_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308798912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313714176))), name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313714368))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313718272))), name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_76_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16_palettized, x = input_205_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor input_209_mode_0 = const()[name = tensor("input_209_mode_0"), val = tensor("EXACT")]; + tensor input_209_cast_fp16 = gelu(mode = input_209_mode_0, x = linear_76_cast_fp16)[name = tensor("input_209_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(313718464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318633728))), name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318633920)))]; + tensor linear_77_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16_palettized, x = input_209_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor input_211_cast_fp16 = add(x = input_203_cast_fp16, y = linear_77_cast_fp16)[name = tensor("input_211_cast_fp16")]; + tensor hidden_states_79_axes_0 = const()[name = tensor("hidden_states_79_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318636544)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318639168)))]; + tensor hidden_states_79_cast_fp16 = layer_norm(axes = hidden_states_79_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("hidden_states_79_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(318641792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319870656))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319870848)))]; + tensor linear_78_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_78_cast_fp16")]; + tensor var_1279_to_fp16 = const()[name = tensor("op_1279_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_83_cast_fp16 = mul(x = linear_78_cast_fp16, y = var_1279_to_fp16)[name = tensor("tensor_83_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(319873472))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321102336))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321102528)))]; + tensor linear_79_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor var_1284 = const()[name = tensor("op_1284"), val = tensor([1, -1, 20, 64])]; + tensor var_1285_cast_fp16 = reshape(shape = var_1284, x = linear_79_cast_fp16)[name = tensor("op_1285_cast_fp16")]; + tensor var_1286_perm_0 = const()[name = tensor("op_1286_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(321105152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322334016))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322334208)))]; + tensor linear_80_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_79_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor var_1291 = const()[name = tensor("op_1291"), val = tensor([1, -1, 20, 64])]; + tensor var_1292_cast_fp16 = reshape(shape = var_1291, x = linear_80_cast_fp16)[name = tensor("op_1292_cast_fp16")]; + tensor var_1293_perm_0 = const()[name = tensor("op_1293_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1300 = const()[name = tensor("op_1300"), val = tensor([1, 77, 20, 64])]; + tensor var_1301_cast_fp16 = reshape(shape = var_1300, x = tensor_83_cast_fp16)[name = tensor("op_1301_cast_fp16")]; + tensor var_1302_perm_0 = const()[name = tensor("op_1302_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1304 = const()[name = tensor("op_1304"), val = tensor([20, -1, 64])]; + tensor transpose_75 = transpose(perm = var_1302_perm_0, x = var_1301_cast_fp16)[name = tensor("transpose_75")]; + tensor query_states_27_cast_fp16 = reshape(shape = var_1304, x = transpose_75)[name = tensor("query_states_27_cast_fp16")]; + tensor var_1306 = const()[name = tensor("op_1306"), val = tensor([20, -1, 64])]; + tensor transpose_74 = transpose(perm = var_1286_perm_0, x = var_1285_cast_fp16)[name = tensor("transpose_74")]; + tensor key_states_55_cast_fp16 = reshape(shape = var_1306, x = transpose_74)[name = tensor("key_states_55_cast_fp16")]; + tensor var_1308 = const()[name = tensor("op_1308"), val = tensor([20, -1, 64])]; + tensor transpose_73 = transpose(perm = var_1293_perm_0, x = var_1292_cast_fp16)[name = tensor("transpose_73")]; + tensor value_states_55_cast_fp16 = reshape(shape = var_1308, x = transpose_73)[name = tensor("value_states_55_cast_fp16")]; + tensor attn_weights_79_transpose_x_1 = const()[name = tensor("attn_weights_79_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_79_transpose_y_1 = const()[name = tensor("attn_weights_79_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_79_cast_fp16 = matmul(transpose_x = attn_weights_79_transpose_x_1, transpose_y = attn_weights_79_transpose_y_1, x = query_states_27_cast_fp16, y = key_states_55_cast_fp16)[name = tensor("attn_weights_79_cast_fp16")]; + tensor var_1313 = const()[name = tensor("op_1313"), val = tensor([1, 20, 77, 77])]; + tensor var_1314_cast_fp16 = reshape(shape = var_1313, x = attn_weights_79_cast_fp16)[name = tensor("op_1314_cast_fp16")]; + tensor attn_weights_81_cast_fp16 = add(x = var_1314_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_81_cast_fp16")]; + tensor var_1319 = const()[name = tensor("op_1319"), val = tensor([20, 77, 77])]; + tensor input_213_cast_fp16 = reshape(shape = var_1319, x = attn_weights_81_cast_fp16)[name = tensor("input_213_cast_fp16")]; + tensor input_215_cast_fp16 = softmax(axis = var_5, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor attn_output_79_transpose_x_0 = const()[name = tensor("attn_output_79_transpose_x_0"), val = tensor(false)]; + tensor attn_output_79_transpose_y_0 = const()[name = tensor("attn_output_79_transpose_y_0"), val = tensor(false)]; + tensor attn_output_79_cast_fp16 = matmul(transpose_x = attn_output_79_transpose_x_0, transpose_y = attn_output_79_transpose_y_0, x = input_215_cast_fp16, y = value_states_55_cast_fp16)[name = tensor("attn_output_79_cast_fp16")]; + tensor var_1324 = const()[name = tensor("op_1324"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_81_cast_fp16 = reshape(shape = var_1324, x = attn_output_79_cast_fp16)[name = tensor("attn_output_81_cast_fp16")]; + tensor attn_output_83_perm_0 = const()[name = tensor("attn_output_83_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1327 = const()[name = tensor("op_1327"), val = tensor([1, 77, 1280])]; + tensor transpose_72 = transpose(perm = attn_output_83_perm_0, x = attn_output_81_cast_fp16)[name = tensor("transpose_72")]; + tensor input_217_cast_fp16 = reshape(shape = var_1327, x = transpose_72)[name = tensor("input_217_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(322336832))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323565696))), name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323565888)))]; + tensor linear_81_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16_palettized, x = input_217_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor input_219_cast_fp16 = add(x = input_211_cast_fp16, y = linear_81_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323568512)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323571136)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323573760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328489024))), name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328489216))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328493120))), name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_82_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16_palettized, x = input_221_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_225_mode_0 = const()[name = tensor("input_225_mode_0"), val = tensor("EXACT")]; + tensor input_225_cast_fp16 = gelu(mode = input_225_mode_0, x = linear_82_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328493312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333408576))), name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333408768)))]; + tensor linear_83_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16_palettized, x = input_225_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor input_227_cast_fp16 = add(x = input_219_cast_fp16, y = linear_83_cast_fp16)[name = tensor("input_227_cast_fp16")]; + tensor hidden_states_85_axes_0 = const()[name = tensor("hidden_states_85_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333411392)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333414016)))]; + tensor hidden_states_85_cast_fp16 = layer_norm(axes = hidden_states_85_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("hidden_states_85_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333416640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334645504))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334645696)))]; + tensor linear_84_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_1366_to_fp16 = const()[name = tensor("op_1366_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_89_cast_fp16 = mul(x = linear_84_cast_fp16, y = var_1366_to_fp16)[name = tensor("tensor_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(334648320))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335877184))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335877376)))]; + tensor linear_85_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_1371 = const()[name = tensor("op_1371"), val = tensor([1, -1, 20, 64])]; + tensor var_1372_cast_fp16 = reshape(shape = var_1371, x = linear_85_cast_fp16)[name = tensor("op_1372_cast_fp16")]; + tensor var_1373_perm_0 = const()[name = tensor("op_1373_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335880000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337108864))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337109056)))]; + tensor linear_86_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_85_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_1378 = const()[name = tensor("op_1378"), val = tensor([1, -1, 20, 64])]; + tensor var_1379_cast_fp16 = reshape(shape = var_1378, x = linear_86_cast_fp16)[name = tensor("op_1379_cast_fp16")]; + tensor var_1380_perm_0 = const()[name = tensor("op_1380_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1387 = const()[name = tensor("op_1387"), val = tensor([1, 77, 20, 64])]; + tensor var_1388_cast_fp16 = reshape(shape = var_1387, x = tensor_89_cast_fp16)[name = tensor("op_1388_cast_fp16")]; + tensor var_1389_perm_0 = const()[name = tensor("op_1389_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1391 = const()[name = tensor("op_1391"), val = tensor([20, -1, 64])]; + tensor transpose_71 = transpose(perm = var_1389_perm_0, x = var_1388_cast_fp16)[name = tensor("transpose_71")]; + tensor query_states_29_cast_fp16 = reshape(shape = var_1391, x = transpose_71)[name = tensor("query_states_29_cast_fp16")]; + tensor var_1393 = const()[name = tensor("op_1393"), val = tensor([20, -1, 64])]; + tensor transpose_70 = transpose(perm = var_1373_perm_0, x = var_1372_cast_fp16)[name = tensor("transpose_70")]; + tensor key_states_59_cast_fp16 = reshape(shape = var_1393, x = transpose_70)[name = tensor("key_states_59_cast_fp16")]; + tensor var_1395 = const()[name = tensor("op_1395"), val = tensor([20, -1, 64])]; + tensor transpose_69 = transpose(perm = var_1380_perm_0, x = var_1379_cast_fp16)[name = tensor("transpose_69")]; + tensor value_states_59_cast_fp16 = reshape(shape = var_1395, x = transpose_69)[name = tensor("value_states_59_cast_fp16")]; + tensor attn_weights_85_transpose_x_1 = const()[name = tensor("attn_weights_85_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_85_transpose_y_1 = const()[name = tensor("attn_weights_85_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_85_cast_fp16 = matmul(transpose_x = attn_weights_85_transpose_x_1, transpose_y = attn_weights_85_transpose_y_1, x = query_states_29_cast_fp16, y = key_states_59_cast_fp16)[name = tensor("attn_weights_85_cast_fp16")]; + tensor var_1400 = const()[name = tensor("op_1400"), val = tensor([1, 20, 77, 77])]; + tensor var_1401_cast_fp16 = reshape(shape = var_1400, x = attn_weights_85_cast_fp16)[name = tensor("op_1401_cast_fp16")]; + tensor attn_weights_87_cast_fp16 = add(x = var_1401_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_87_cast_fp16")]; + tensor var_1406 = const()[name = tensor("op_1406"), val = tensor([20, 77, 77])]; + tensor input_229_cast_fp16 = reshape(shape = var_1406, x = attn_weights_87_cast_fp16)[name = tensor("input_229_cast_fp16")]; + tensor input_231_cast_fp16 = softmax(axis = var_5, x = input_229_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor attn_output_85_transpose_x_0 = const()[name = tensor("attn_output_85_transpose_x_0"), val = tensor(false)]; + tensor attn_output_85_transpose_y_0 = const()[name = tensor("attn_output_85_transpose_y_0"), val = tensor(false)]; + tensor attn_output_85_cast_fp16 = matmul(transpose_x = attn_output_85_transpose_x_0, transpose_y = attn_output_85_transpose_y_0, x = input_231_cast_fp16, y = value_states_59_cast_fp16)[name = tensor("attn_output_85_cast_fp16")]; + tensor var_1411 = const()[name = tensor("op_1411"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_87_cast_fp16 = reshape(shape = var_1411, x = attn_output_85_cast_fp16)[name = tensor("attn_output_87_cast_fp16")]; + tensor attn_output_89_perm_0 = const()[name = tensor("attn_output_89_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, 77, 1280])]; + tensor transpose_68 = transpose(perm = attn_output_89_perm_0, x = attn_output_87_cast_fp16)[name = tensor("transpose_68")]; + tensor input_233_cast_fp16 = reshape(shape = var_1414, x = transpose_68)[name = tensor("input_233_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(337111680))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338340544))), name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338340736)))]; + tensor linear_87_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16_palettized, x = input_233_cast_fp16)[name = tensor("linear_87_cast_fp16")]; + tensor input_235_cast_fp16 = add(x = input_227_cast_fp16, y = linear_87_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor input_237_axes_0 = const()[name = tensor("input_237_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338343360)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338345984)))]; + tensor input_237_cast_fp16 = layer_norm(axes = input_237_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("input_237_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338348608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343263872))), name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343264064))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343267968))), name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_88_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16_palettized, x = input_237_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_241_mode_0 = const()[name = tensor("input_241_mode_0"), val = tensor("EXACT")]; + tensor input_241_cast_fp16 = gelu(mode = input_241_mode_0, x = linear_88_cast_fp16)[name = tensor("input_241_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(343268160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348183424))), name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348183616)))]; + tensor linear_89_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16_palettized, x = input_241_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_243_cast_fp16 = add(x = input_235_cast_fp16, y = linear_89_cast_fp16)[name = tensor("input_243_cast_fp16")]; + tensor hidden_states_91_axes_0 = const()[name = tensor("hidden_states_91_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348186240)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348188864)))]; + tensor hidden_states_91_cast_fp16 = layer_norm(axes = hidden_states_91_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("hidden_states_91_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348191488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349420352))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349420544)))]; + tensor linear_90_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_1453_to_fp16 = const()[name = tensor("op_1453_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_95_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1453_to_fp16)[name = tensor("tensor_95_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349423168))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350652032))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350652224)))]; + tensor linear_91_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor var_1458 = const()[name = tensor("op_1458"), val = tensor([1, -1, 20, 64])]; + tensor var_1459_cast_fp16 = reshape(shape = var_1458, x = linear_91_cast_fp16)[name = tensor("op_1459_cast_fp16")]; + tensor var_1460_perm_0 = const()[name = tensor("op_1460_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(350654848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351883712))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351883904)))]; + tensor linear_92_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_91_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_1465 = const()[name = tensor("op_1465"), val = tensor([1, -1, 20, 64])]; + tensor var_1466_cast_fp16 = reshape(shape = var_1465, x = linear_92_cast_fp16)[name = tensor("op_1466_cast_fp16")]; + tensor var_1467_perm_0 = const()[name = tensor("op_1467_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1474 = const()[name = tensor("op_1474"), val = tensor([1, 77, 20, 64])]; + tensor var_1475_cast_fp16 = reshape(shape = var_1474, x = tensor_95_cast_fp16)[name = tensor("op_1475_cast_fp16")]; + tensor var_1476_perm_0 = const()[name = tensor("op_1476_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1478 = const()[name = tensor("op_1478"), val = tensor([20, -1, 64])]; + tensor transpose_67 = transpose(perm = var_1476_perm_0, x = var_1475_cast_fp16)[name = tensor("transpose_67")]; + tensor query_states_31_cast_fp16 = reshape(shape = var_1478, x = transpose_67)[name = tensor("query_states_31_cast_fp16")]; + tensor var_1480 = const()[name = tensor("op_1480"), val = tensor([20, -1, 64])]; + tensor transpose_66 = transpose(perm = var_1460_perm_0, x = var_1459_cast_fp16)[name = tensor("transpose_66")]; + tensor key_states_63_cast_fp16 = reshape(shape = var_1480, x = transpose_66)[name = tensor("key_states_63_cast_fp16")]; + tensor var_1482 = const()[name = tensor("op_1482"), val = tensor([20, -1, 64])]; + tensor transpose_65 = transpose(perm = var_1467_perm_0, x = var_1466_cast_fp16)[name = tensor("transpose_65")]; + tensor value_states_63_cast_fp16 = reshape(shape = var_1482, x = transpose_65)[name = tensor("value_states_63_cast_fp16")]; + tensor attn_weights_91_transpose_x_1 = const()[name = tensor("attn_weights_91_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_91_transpose_y_1 = const()[name = tensor("attn_weights_91_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_91_cast_fp16 = matmul(transpose_x = attn_weights_91_transpose_x_1, transpose_y = attn_weights_91_transpose_y_1, x = query_states_31_cast_fp16, y = key_states_63_cast_fp16)[name = tensor("attn_weights_91_cast_fp16")]; + tensor var_1487 = const()[name = tensor("op_1487"), val = tensor([1, 20, 77, 77])]; + tensor var_1488_cast_fp16 = reshape(shape = var_1487, x = attn_weights_91_cast_fp16)[name = tensor("op_1488_cast_fp16")]; + tensor attn_weights_93_cast_fp16 = add(x = var_1488_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_93_cast_fp16")]; + tensor var_1493 = const()[name = tensor("op_1493"), val = tensor([20, 77, 77])]; + tensor input_245_cast_fp16 = reshape(shape = var_1493, x = attn_weights_93_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor input_247_cast_fp16 = softmax(axis = var_5, x = input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; + tensor attn_output_91_transpose_x_0 = const()[name = tensor("attn_output_91_transpose_x_0"), val = tensor(false)]; + tensor attn_output_91_transpose_y_0 = const()[name = tensor("attn_output_91_transpose_y_0"), val = tensor(false)]; + tensor attn_output_91_cast_fp16 = matmul(transpose_x = attn_output_91_transpose_x_0, transpose_y = attn_output_91_transpose_y_0, x = input_247_cast_fp16, y = value_states_63_cast_fp16)[name = tensor("attn_output_91_cast_fp16")]; + tensor var_1498 = const()[name = tensor("op_1498"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_93_cast_fp16 = reshape(shape = var_1498, x = attn_output_91_cast_fp16)[name = tensor("attn_output_93_cast_fp16")]; + tensor attn_output_95_perm_0 = const()[name = tensor("attn_output_95_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1501 = const()[name = tensor("op_1501"), val = tensor([1, 77, 1280])]; + tensor transpose_64 = transpose(perm = attn_output_95_perm_0, x = attn_output_93_cast_fp16)[name = tensor("transpose_64")]; + tensor input_249_cast_fp16 = reshape(shape = var_1501, x = transpose_64)[name = tensor("input_249_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(351886528))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353115392))), name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353115584)))]; + tensor linear_93_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16_palettized, x = input_249_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor input_251_cast_fp16 = add(x = input_243_cast_fp16, y = linear_93_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor input_253_axes_0 = const()[name = tensor("input_253_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353118208)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353120832)))]; + tensor input_253_cast_fp16 = layer_norm(axes = input_253_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("input_253_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(353123456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358038720))), name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358038912))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358042816))), name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_94_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16_palettized, x = input_253_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor input_257_mode_0 = const()[name = tensor("input_257_mode_0"), val = tensor("EXACT")]; + tensor input_257_cast_fp16 = gelu(mode = input_257_mode_0, x = linear_94_cast_fp16)[name = tensor("input_257_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(358043008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362958272))), name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362958464)))]; + tensor linear_95_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16_palettized, x = input_257_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor input_259_cast_fp16 = add(x = input_251_cast_fp16, y = linear_95_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor hidden_states_97_axes_0 = const()[name = tensor("hidden_states_97_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362961088)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362963712)))]; + tensor hidden_states_97_cast_fp16 = layer_norm(axes = hidden_states_97_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16, x = input_259_cast_fp16)[name = tensor("hidden_states_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362966336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364195200))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364195392)))]; + tensor linear_96_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_96_cast_fp16")]; + tensor var_1540_to_fp16 = const()[name = tensor("op_1540_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_101_cast_fp16 = mul(x = linear_96_cast_fp16, y = var_1540_to_fp16)[name = tensor("tensor_101_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364198016))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365426880))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365427072)))]; + tensor linear_97_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor var_1545 = const()[name = tensor("op_1545"), val = tensor([1, -1, 20, 64])]; + tensor var_1546_cast_fp16 = reshape(shape = var_1545, x = linear_97_cast_fp16)[name = tensor("op_1546_cast_fp16")]; + tensor var_1547_perm_0 = const()[name = tensor("op_1547_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(365429696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366658560))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366658752)))]; + tensor linear_98_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_97_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor var_1552 = const()[name = tensor("op_1552"), val = tensor([1, -1, 20, 64])]; + tensor var_1553_cast_fp16 = reshape(shape = var_1552, x = linear_98_cast_fp16)[name = tensor("op_1553_cast_fp16")]; + tensor var_1554_perm_0 = const()[name = tensor("op_1554_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1561 = const()[name = tensor("op_1561"), val = tensor([1, 77, 20, 64])]; + tensor var_1562_cast_fp16 = reshape(shape = var_1561, x = tensor_101_cast_fp16)[name = tensor("op_1562_cast_fp16")]; + tensor var_1563_perm_0 = const()[name = tensor("op_1563_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1565 = const()[name = tensor("op_1565"), val = tensor([20, -1, 64])]; + tensor transpose_63 = transpose(perm = var_1563_perm_0, x = var_1562_cast_fp16)[name = tensor("transpose_63")]; + tensor query_states_33_cast_fp16 = reshape(shape = var_1565, x = transpose_63)[name = tensor("query_states_33_cast_fp16")]; + tensor var_1567 = const()[name = tensor("op_1567"), val = tensor([20, -1, 64])]; + tensor transpose_62 = transpose(perm = var_1547_perm_0, x = var_1546_cast_fp16)[name = tensor("transpose_62")]; + tensor key_states_67_cast_fp16 = reshape(shape = var_1567, x = transpose_62)[name = tensor("key_states_67_cast_fp16")]; + tensor var_1569 = const()[name = tensor("op_1569"), val = tensor([20, -1, 64])]; + tensor transpose_61 = transpose(perm = var_1554_perm_0, x = var_1553_cast_fp16)[name = tensor("transpose_61")]; + tensor value_states_67_cast_fp16 = reshape(shape = var_1569, x = transpose_61)[name = tensor("value_states_67_cast_fp16")]; + tensor attn_weights_97_transpose_x_1 = const()[name = tensor("attn_weights_97_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_97_transpose_y_1 = const()[name = tensor("attn_weights_97_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_97_cast_fp16 = matmul(transpose_x = attn_weights_97_transpose_x_1, transpose_y = attn_weights_97_transpose_y_1, x = query_states_33_cast_fp16, y = key_states_67_cast_fp16)[name = tensor("attn_weights_97_cast_fp16")]; + tensor var_1574 = const()[name = tensor("op_1574"), val = tensor([1, 20, 77, 77])]; + tensor var_1575_cast_fp16 = reshape(shape = var_1574, x = attn_weights_97_cast_fp16)[name = tensor("op_1575_cast_fp16")]; + tensor attn_weights_99_cast_fp16 = add(x = var_1575_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_99_cast_fp16")]; + tensor var_1580 = const()[name = tensor("op_1580"), val = tensor([20, 77, 77])]; + tensor input_261_cast_fp16 = reshape(shape = var_1580, x = attn_weights_99_cast_fp16)[name = tensor("input_261_cast_fp16")]; + tensor input_263_cast_fp16 = softmax(axis = var_5, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; + tensor attn_output_97_transpose_x_0 = const()[name = tensor("attn_output_97_transpose_x_0"), val = tensor(false)]; + tensor attn_output_97_transpose_y_0 = const()[name = tensor("attn_output_97_transpose_y_0"), val = tensor(false)]; + tensor attn_output_97_cast_fp16 = matmul(transpose_x = attn_output_97_transpose_x_0, transpose_y = attn_output_97_transpose_y_0, x = input_263_cast_fp16, y = value_states_67_cast_fp16)[name = tensor("attn_output_97_cast_fp16")]; + tensor var_1585 = const()[name = tensor("op_1585"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_99_cast_fp16 = reshape(shape = var_1585, x = attn_output_97_cast_fp16)[name = tensor("attn_output_99_cast_fp16")]; + tensor attn_output_101_perm_0 = const()[name = tensor("attn_output_101_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, 77, 1280])]; + tensor transpose_60 = transpose(perm = attn_output_101_perm_0, x = attn_output_99_cast_fp16)[name = tensor("transpose_60")]; + tensor input_265_cast_fp16 = reshape(shape = var_1588, x = transpose_60)[name = tensor("input_265_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366661376))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367890240))), name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367890432)))]; + tensor linear_99_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16_palettized, x = input_265_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor input_267_cast_fp16 = add(x = input_259_cast_fp16, y = linear_99_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor input_269_axes_0 = const()[name = tensor("input_269_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367893056)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367895680)))]; + tensor input_269_cast_fp16 = layer_norm(axes = input_269_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(367898304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372813568))), name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372813760))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372817664))), name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_100_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16_palettized, x = input_269_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_273_mode_0 = const()[name = tensor("input_273_mode_0"), val = tensor("EXACT")]; + tensor input_273_cast_fp16 = gelu(mode = input_273_mode_0, x = linear_100_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372817856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377733120))), name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377733312)))]; + tensor linear_101_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16_palettized, x = input_273_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor input_275_cast_fp16 = add(x = input_267_cast_fp16, y = linear_101_cast_fp16)[name = tensor("input_275_cast_fp16")]; + tensor hidden_states_103_axes_0 = const()[name = tensor("hidden_states_103_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377735936)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377738560)))]; + tensor hidden_states_103_cast_fp16 = layer_norm(axes = hidden_states_103_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("hidden_states_103_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377741184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378970048))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378970240)))]; + tensor linear_102_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_1627_to_fp16 = const()[name = tensor("op_1627_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_107_cast_fp16 = mul(x = linear_102_cast_fp16, y = var_1627_to_fp16)[name = tensor("tensor_107_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378972864))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380201728))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380201920)))]; + tensor linear_103_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_1632 = const()[name = tensor("op_1632"), val = tensor([1, -1, 20, 64])]; + tensor var_1633_cast_fp16 = reshape(shape = var_1632, x = linear_103_cast_fp16)[name = tensor("op_1633_cast_fp16")]; + tensor var_1634_perm_0 = const()[name = tensor("op_1634_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380204544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381433408))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381433600)))]; + tensor linear_104_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_103_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_1639 = const()[name = tensor("op_1639"), val = tensor([1, -1, 20, 64])]; + tensor var_1640_cast_fp16 = reshape(shape = var_1639, x = linear_104_cast_fp16)[name = tensor("op_1640_cast_fp16")]; + tensor var_1641_perm_0 = const()[name = tensor("op_1641_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1648 = const()[name = tensor("op_1648"), val = tensor([1, 77, 20, 64])]; + tensor var_1649_cast_fp16 = reshape(shape = var_1648, x = tensor_107_cast_fp16)[name = tensor("op_1649_cast_fp16")]; + tensor var_1650_perm_0 = const()[name = tensor("op_1650_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1652 = const()[name = tensor("op_1652"), val = tensor([20, -1, 64])]; + tensor transpose_59 = transpose(perm = var_1650_perm_0, x = var_1649_cast_fp16)[name = tensor("transpose_59")]; + tensor query_states_35_cast_fp16 = reshape(shape = var_1652, x = transpose_59)[name = tensor("query_states_35_cast_fp16")]; + tensor var_1654 = const()[name = tensor("op_1654"), val = tensor([20, -1, 64])]; + tensor transpose_58 = transpose(perm = var_1634_perm_0, x = var_1633_cast_fp16)[name = tensor("transpose_58")]; + tensor key_states_71_cast_fp16 = reshape(shape = var_1654, x = transpose_58)[name = tensor("key_states_71_cast_fp16")]; + tensor var_1656 = const()[name = tensor("op_1656"), val = tensor([20, -1, 64])]; + tensor transpose_57 = transpose(perm = var_1641_perm_0, x = var_1640_cast_fp16)[name = tensor("transpose_57")]; + tensor value_states_71_cast_fp16 = reshape(shape = var_1656, x = transpose_57)[name = tensor("value_states_71_cast_fp16")]; + tensor attn_weights_103_transpose_x_1 = const()[name = tensor("attn_weights_103_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_103_transpose_y_1 = const()[name = tensor("attn_weights_103_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_103_cast_fp16 = matmul(transpose_x = attn_weights_103_transpose_x_1, transpose_y = attn_weights_103_transpose_y_1, x = query_states_35_cast_fp16, y = key_states_71_cast_fp16)[name = tensor("attn_weights_103_cast_fp16")]; + tensor var_1661 = const()[name = tensor("op_1661"), val = tensor([1, 20, 77, 77])]; + tensor var_1662_cast_fp16 = reshape(shape = var_1661, x = attn_weights_103_cast_fp16)[name = tensor("op_1662_cast_fp16")]; + tensor attn_weights_105_cast_fp16 = add(x = var_1662_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_105_cast_fp16")]; + tensor var_1667 = const()[name = tensor("op_1667"), val = tensor([20, 77, 77])]; + tensor input_277_cast_fp16 = reshape(shape = var_1667, x = attn_weights_105_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor input_279_cast_fp16 = softmax(axis = var_5, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; + tensor attn_output_103_transpose_x_0 = const()[name = tensor("attn_output_103_transpose_x_0"), val = tensor(false)]; + tensor attn_output_103_transpose_y_0 = const()[name = tensor("attn_output_103_transpose_y_0"), val = tensor(false)]; + tensor attn_output_103_cast_fp16 = matmul(transpose_x = attn_output_103_transpose_x_0, transpose_y = attn_output_103_transpose_y_0, x = input_279_cast_fp16, y = value_states_71_cast_fp16)[name = tensor("attn_output_103_cast_fp16")]; + tensor var_1672 = const()[name = tensor("op_1672"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_105_cast_fp16 = reshape(shape = var_1672, x = attn_output_103_cast_fp16)[name = tensor("attn_output_105_cast_fp16")]; + tensor attn_output_107_perm_0 = const()[name = tensor("attn_output_107_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1675 = const()[name = tensor("op_1675"), val = tensor([1, 77, 1280])]; + tensor transpose_56 = transpose(perm = attn_output_107_perm_0, x = attn_output_105_cast_fp16)[name = tensor("transpose_56")]; + tensor input_281_cast_fp16 = reshape(shape = var_1675, x = transpose_56)[name = tensor("input_281_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381436224))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382665088))), name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382665280)))]; + tensor linear_105_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16_palettized, x = input_281_cast_fp16)[name = tensor("linear_105_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_275_cast_fp16, y = linear_105_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382667904)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382670528)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(382673152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387588416))), name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387588608))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387592512))), name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_106_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16_palettized, x = input_285_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_289_mode_0 = const()[name = tensor("input_289_mode_0"), val = tensor("EXACT")]; + tensor input_289_cast_fp16 = gelu(mode = input_289_mode_0, x = linear_106_cast_fp16)[name = tensor("input_289_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(387592704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392507968))), name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392508160)))]; + tensor linear_107_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16_palettized, x = input_289_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_291_cast_fp16 = add(x = input_283_cast_fp16, y = linear_107_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor hidden_states_109_axes_0 = const()[name = tensor("hidden_states_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392510784)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392513408)))]; + tensor hidden_states_109_cast_fp16 = layer_norm(axes = hidden_states_109_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("hidden_states_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(392516032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393744896))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393745088)))]; + tensor linear_108_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_1714_to_fp16 = const()[name = tensor("op_1714_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_113_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_1714_to_fp16)[name = tensor("tensor_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393747712))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(394976576))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(394976768)))]; + tensor linear_109_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor var_1719 = const()[name = tensor("op_1719"), val = tensor([1, -1, 20, 64])]; + tensor var_1720_cast_fp16 = reshape(shape = var_1719, x = linear_109_cast_fp16)[name = tensor("op_1720_cast_fp16")]; + tensor var_1721_perm_0 = const()[name = tensor("op_1721_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(394979392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396208256))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396208448)))]; + tensor linear_110_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_109_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_1726 = const()[name = tensor("op_1726"), val = tensor([1, -1, 20, 64])]; + tensor var_1727_cast_fp16 = reshape(shape = var_1726, x = linear_110_cast_fp16)[name = tensor("op_1727_cast_fp16")]; + tensor var_1728_perm_0 = const()[name = tensor("op_1728_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1735 = const()[name = tensor("op_1735"), val = tensor([1, 77, 20, 64])]; + tensor var_1736_cast_fp16 = reshape(shape = var_1735, x = tensor_113_cast_fp16)[name = tensor("op_1736_cast_fp16")]; + tensor var_1737_perm_0 = const()[name = tensor("op_1737_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1739 = const()[name = tensor("op_1739"), val = tensor([20, -1, 64])]; + tensor transpose_55 = transpose(perm = var_1737_perm_0, x = var_1736_cast_fp16)[name = tensor("transpose_55")]; + tensor query_states_37_cast_fp16 = reshape(shape = var_1739, x = transpose_55)[name = tensor("query_states_37_cast_fp16")]; + tensor var_1741 = const()[name = tensor("op_1741"), val = tensor([20, -1, 64])]; + tensor transpose_54 = transpose(perm = var_1721_perm_0, x = var_1720_cast_fp16)[name = tensor("transpose_54")]; + tensor key_states_75_cast_fp16 = reshape(shape = var_1741, x = transpose_54)[name = tensor("key_states_75_cast_fp16")]; + tensor var_1743 = const()[name = tensor("op_1743"), val = tensor([20, -1, 64])]; + tensor transpose_53 = transpose(perm = var_1728_perm_0, x = var_1727_cast_fp16)[name = tensor("transpose_53")]; + tensor value_states_75_cast_fp16 = reshape(shape = var_1743, x = transpose_53)[name = tensor("value_states_75_cast_fp16")]; + tensor attn_weights_109_transpose_x_1 = const()[name = tensor("attn_weights_109_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_109_transpose_y_1 = const()[name = tensor("attn_weights_109_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_109_cast_fp16 = matmul(transpose_x = attn_weights_109_transpose_x_1, transpose_y = attn_weights_109_transpose_y_1, x = query_states_37_cast_fp16, y = key_states_75_cast_fp16)[name = tensor("attn_weights_109_cast_fp16")]; + tensor var_1748 = const()[name = tensor("op_1748"), val = tensor([1, 20, 77, 77])]; + tensor var_1749_cast_fp16 = reshape(shape = var_1748, x = attn_weights_109_cast_fp16)[name = tensor("op_1749_cast_fp16")]; + tensor attn_weights_111_cast_fp16 = add(x = var_1749_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_111_cast_fp16")]; + tensor var_1754 = const()[name = tensor("op_1754"), val = tensor([20, 77, 77])]; + tensor input_293_cast_fp16 = reshape(shape = var_1754, x = attn_weights_111_cast_fp16)[name = tensor("input_293_cast_fp16")]; + tensor input_295_cast_fp16 = softmax(axis = var_5, x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; + tensor attn_output_109_transpose_x_0 = const()[name = tensor("attn_output_109_transpose_x_0"), val = tensor(false)]; + tensor attn_output_109_transpose_y_0 = const()[name = tensor("attn_output_109_transpose_y_0"), val = tensor(false)]; + tensor attn_output_109_cast_fp16 = matmul(transpose_x = attn_output_109_transpose_x_0, transpose_y = attn_output_109_transpose_y_0, x = input_295_cast_fp16, y = value_states_75_cast_fp16)[name = tensor("attn_output_109_cast_fp16")]; + tensor var_1759 = const()[name = tensor("op_1759"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_111_cast_fp16 = reshape(shape = var_1759, x = attn_output_109_cast_fp16)[name = tensor("attn_output_111_cast_fp16")]; + tensor attn_output_113_perm_0 = const()[name = tensor("attn_output_113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, 77, 1280])]; + tensor transpose_52 = transpose(perm = attn_output_113_perm_0, x = attn_output_111_cast_fp16)[name = tensor("transpose_52")]; + tensor input_297_cast_fp16 = reshape(shape = var_1762, x = transpose_52)[name = tensor("input_297_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(396211072))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397439936))), name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397440128)))]; + tensor linear_111_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16_palettized, x = input_297_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_111_cast_fp16)[name = tensor("input_299_cast_fp16")]; + tensor input_301_axes_0 = const()[name = tensor("input_301_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397442752)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397445376)))]; + tensor input_301_cast_fp16 = layer_norm(axes = input_301_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397448000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402363264))), name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402363456))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402367360))), name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_112_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16_palettized, x = input_301_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor input_305_mode_0 = const()[name = tensor("input_305_mode_0"), val = tensor("EXACT")]; + tensor input_305_cast_fp16 = gelu(mode = input_305_mode_0, x = linear_112_cast_fp16)[name = tensor("input_305_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402367552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407282816))), name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407283008)))]; + tensor linear_113_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16_palettized, x = input_305_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_299_cast_fp16, y = linear_113_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor hidden_states_115_axes_0 = const()[name = tensor("hidden_states_115_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407285632)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407288256)))]; + tensor hidden_states_115_cast_fp16 = layer_norm(axes = hidden_states_115_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("hidden_states_115_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407290880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408519744))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408519936)))]; + tensor linear_114_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_114_cast_fp16")]; + tensor var_1801_to_fp16 = const()[name = tensor("op_1801_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_119_cast_fp16 = mul(x = linear_114_cast_fp16, y = var_1801_to_fp16)[name = tensor("tensor_119_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408522560))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409751424))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409751616)))]; + tensor linear_115_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor var_1806 = const()[name = tensor("op_1806"), val = tensor([1, -1, 20, 64])]; + tensor var_1807_cast_fp16 = reshape(shape = var_1806, x = linear_115_cast_fp16)[name = tensor("op_1807_cast_fp16")]; + tensor var_1808_perm_0 = const()[name = tensor("op_1808_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(409754240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(410983104))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(410983296)))]; + tensor linear_116_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_115_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor var_1813 = const()[name = tensor("op_1813"), val = tensor([1, -1, 20, 64])]; + tensor var_1814_cast_fp16 = reshape(shape = var_1813, x = linear_116_cast_fp16)[name = tensor("op_1814_cast_fp16")]; + tensor var_1815_perm_0 = const()[name = tensor("op_1815_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1822 = const()[name = tensor("op_1822"), val = tensor([1, 77, 20, 64])]; + tensor var_1823_cast_fp16 = reshape(shape = var_1822, x = tensor_119_cast_fp16)[name = tensor("op_1823_cast_fp16")]; + tensor var_1824_perm_0 = const()[name = tensor("op_1824_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1826 = const()[name = tensor("op_1826"), val = tensor([20, -1, 64])]; + tensor transpose_51 = transpose(perm = var_1824_perm_0, x = var_1823_cast_fp16)[name = tensor("transpose_51")]; + tensor query_states_39_cast_fp16 = reshape(shape = var_1826, x = transpose_51)[name = tensor("query_states_39_cast_fp16")]; + tensor var_1828 = const()[name = tensor("op_1828"), val = tensor([20, -1, 64])]; + tensor transpose_50 = transpose(perm = var_1808_perm_0, x = var_1807_cast_fp16)[name = tensor("transpose_50")]; + tensor key_states_79_cast_fp16 = reshape(shape = var_1828, x = transpose_50)[name = tensor("key_states_79_cast_fp16")]; + tensor var_1830 = const()[name = tensor("op_1830"), val = tensor([20, -1, 64])]; + tensor transpose_49 = transpose(perm = var_1815_perm_0, x = var_1814_cast_fp16)[name = tensor("transpose_49")]; + tensor value_states_79_cast_fp16 = reshape(shape = var_1830, x = transpose_49)[name = tensor("value_states_79_cast_fp16")]; + tensor attn_weights_115_transpose_x_1 = const()[name = tensor("attn_weights_115_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_115_transpose_y_1 = const()[name = tensor("attn_weights_115_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_115_cast_fp16 = matmul(transpose_x = attn_weights_115_transpose_x_1, transpose_y = attn_weights_115_transpose_y_1, x = query_states_39_cast_fp16, y = key_states_79_cast_fp16)[name = tensor("attn_weights_115_cast_fp16")]; + tensor var_1835 = const()[name = tensor("op_1835"), val = tensor([1, 20, 77, 77])]; + tensor var_1836_cast_fp16 = reshape(shape = var_1835, x = attn_weights_115_cast_fp16)[name = tensor("op_1836_cast_fp16")]; + tensor attn_weights_117_cast_fp16 = add(x = var_1836_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_117_cast_fp16")]; + tensor var_1841 = const()[name = tensor("op_1841"), val = tensor([20, 77, 77])]; + tensor input_309_cast_fp16 = reshape(shape = var_1841, x = attn_weights_117_cast_fp16)[name = tensor("input_309_cast_fp16")]; + tensor input_311_cast_fp16 = softmax(axis = var_5, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor attn_output_115_transpose_x_0 = const()[name = tensor("attn_output_115_transpose_x_0"), val = tensor(false)]; + tensor attn_output_115_transpose_y_0 = const()[name = tensor("attn_output_115_transpose_y_0"), val = tensor(false)]; + tensor attn_output_115_cast_fp16 = matmul(transpose_x = attn_output_115_transpose_x_0, transpose_y = attn_output_115_transpose_y_0, x = input_311_cast_fp16, y = value_states_79_cast_fp16)[name = tensor("attn_output_115_cast_fp16")]; + tensor var_1846 = const()[name = tensor("op_1846"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_117_cast_fp16 = reshape(shape = var_1846, x = attn_output_115_cast_fp16)[name = tensor("attn_output_117_cast_fp16")]; + tensor attn_output_119_perm_0 = const()[name = tensor("attn_output_119_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1849 = const()[name = tensor("op_1849"), val = tensor([1, 77, 1280])]; + tensor transpose_48 = transpose(perm = attn_output_119_perm_0, x = attn_output_117_cast_fp16)[name = tensor("transpose_48")]; + tensor input_313_cast_fp16 = reshape(shape = var_1849, x = transpose_48)[name = tensor("input_313_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(410985920))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412214784))), name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412214976)))]; + tensor linear_117_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16_palettized, x = input_313_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor input_315_cast_fp16 = add(x = input_307_cast_fp16, y = linear_117_cast_fp16)[name = tensor("input_315_cast_fp16")]; + tensor input_317_axes_0 = const()[name = tensor("input_317_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412217600)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412220224)))]; + tensor input_317_cast_fp16 = layer_norm(axes = input_317_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412222848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417138112))), name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417138304))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417142208))), name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_118_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16_palettized, x = input_317_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_321_mode_0 = const()[name = tensor("input_321_mode_0"), val = tensor("EXACT")]; + tensor input_321_cast_fp16 = gelu(mode = input_321_mode_0, x = linear_118_cast_fp16)[name = tensor("input_321_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417142400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422057664))), name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422057856)))]; + tensor linear_119_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16_palettized, x = input_321_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor input_323_cast_fp16 = add(x = input_315_cast_fp16, y = linear_119_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor hidden_states_121_axes_0 = const()[name = tensor("hidden_states_121_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422060480)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422063104)))]; + tensor hidden_states_121_cast_fp16 = layer_norm(axes = hidden_states_121_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("hidden_states_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(422065728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423294592))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423294784)))]; + tensor linear_120_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_1888_to_fp16 = const()[name = tensor("op_1888_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_125_cast_fp16 = mul(x = linear_120_cast_fp16, y = var_1888_to_fp16)[name = tensor("tensor_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(423297408))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424526272))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424526464)))]; + tensor linear_121_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_1893 = const()[name = tensor("op_1893"), val = tensor([1, -1, 20, 64])]; + tensor var_1894_cast_fp16 = reshape(shape = var_1893, x = linear_121_cast_fp16)[name = tensor("op_1894_cast_fp16")]; + tensor var_1895_perm_0 = const()[name = tensor("op_1895_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(424529088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425757952))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425758144)))]; + tensor linear_122_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_121_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_1900 = const()[name = tensor("op_1900"), val = tensor([1, -1, 20, 64])]; + tensor var_1901_cast_fp16 = reshape(shape = var_1900, x = linear_122_cast_fp16)[name = tensor("op_1901_cast_fp16")]; + tensor var_1902_perm_0 = const()[name = tensor("op_1902_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1909 = const()[name = tensor("op_1909"), val = tensor([1, 77, 20, 64])]; + tensor var_1910_cast_fp16 = reshape(shape = var_1909, x = tensor_125_cast_fp16)[name = tensor("op_1910_cast_fp16")]; + tensor var_1911_perm_0 = const()[name = tensor("op_1911_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1913 = const()[name = tensor("op_1913"), val = tensor([20, -1, 64])]; + tensor transpose_47 = transpose(perm = var_1911_perm_0, x = var_1910_cast_fp16)[name = tensor("transpose_47")]; + tensor query_states_41_cast_fp16 = reshape(shape = var_1913, x = transpose_47)[name = tensor("query_states_41_cast_fp16")]; + tensor var_1915 = const()[name = tensor("op_1915"), val = tensor([20, -1, 64])]; + tensor transpose_46 = transpose(perm = var_1895_perm_0, x = var_1894_cast_fp16)[name = tensor("transpose_46")]; + tensor key_states_83_cast_fp16 = reshape(shape = var_1915, x = transpose_46)[name = tensor("key_states_83_cast_fp16")]; + tensor var_1917 = const()[name = tensor("op_1917"), val = tensor([20, -1, 64])]; + tensor transpose_45 = transpose(perm = var_1902_perm_0, x = var_1901_cast_fp16)[name = tensor("transpose_45")]; + tensor value_states_83_cast_fp16 = reshape(shape = var_1917, x = transpose_45)[name = tensor("value_states_83_cast_fp16")]; + tensor attn_weights_121_transpose_x_1 = const()[name = tensor("attn_weights_121_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_121_transpose_y_1 = const()[name = tensor("attn_weights_121_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_121_cast_fp16 = matmul(transpose_x = attn_weights_121_transpose_x_1, transpose_y = attn_weights_121_transpose_y_1, x = query_states_41_cast_fp16, y = key_states_83_cast_fp16)[name = tensor("attn_weights_121_cast_fp16")]; + tensor var_1922 = const()[name = tensor("op_1922"), val = tensor([1, 20, 77, 77])]; + tensor var_1923_cast_fp16 = reshape(shape = var_1922, x = attn_weights_121_cast_fp16)[name = tensor("op_1923_cast_fp16")]; + tensor attn_weights_123_cast_fp16 = add(x = var_1923_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_123_cast_fp16")]; + tensor var_1928 = const()[name = tensor("op_1928"), val = tensor([20, 77, 77])]; + tensor input_325_cast_fp16 = reshape(shape = var_1928, x = attn_weights_123_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor input_327_cast_fp16 = softmax(axis = var_5, x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; + tensor attn_output_121_transpose_x_0 = const()[name = tensor("attn_output_121_transpose_x_0"), val = tensor(false)]; + tensor attn_output_121_transpose_y_0 = const()[name = tensor("attn_output_121_transpose_y_0"), val = tensor(false)]; + tensor attn_output_121_cast_fp16 = matmul(transpose_x = attn_output_121_transpose_x_0, transpose_y = attn_output_121_transpose_y_0, x = input_327_cast_fp16, y = value_states_83_cast_fp16)[name = tensor("attn_output_121_cast_fp16")]; + tensor var_1933 = const()[name = tensor("op_1933"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_123_cast_fp16 = reshape(shape = var_1933, x = attn_output_121_cast_fp16)[name = tensor("attn_output_123_cast_fp16")]; + tensor attn_output_125_perm_0 = const()[name = tensor("attn_output_125_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, 77, 1280])]; + tensor transpose_44 = transpose(perm = attn_output_125_perm_0, x = attn_output_123_cast_fp16)[name = tensor("transpose_44")]; + tensor input_329_cast_fp16 = reshape(shape = var_1936, x = transpose_44)[name = tensor("input_329_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425760768))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426989632))), name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426989824)))]; + tensor linear_123_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16_palettized, x = input_329_cast_fp16)[name = tensor("linear_123_cast_fp16")]; + tensor input_331_cast_fp16 = add(x = input_323_cast_fp16, y = linear_123_cast_fp16)[name = tensor("input_331_cast_fp16")]; + tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426992448)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426995072)))]; + tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426997696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431912960))), name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431913152))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431917056))), name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_124_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16_palettized, x = input_333_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_337_mode_0 = const()[name = tensor("input_337_mode_0"), val = tensor("EXACT")]; + tensor input_337_cast_fp16 = gelu(mode = input_337_mode_0, x = linear_124_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(431917248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436832512))), name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436832704)))]; + tensor linear_125_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16_palettized, x = input_337_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_339_cast_fp16 = add(x = input_331_cast_fp16, y = linear_125_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor hidden_states_127_axes_0 = const()[name = tensor("hidden_states_127_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436835328)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436837952)))]; + tensor hidden_states_127_cast_fp16 = layer_norm(axes = hidden_states_127_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("hidden_states_127_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(436840576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438069440))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438069632)))]; + tensor linear_126_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_1975_to_fp16 = const()[name = tensor("op_1975_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_131_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_1975_to_fp16)[name = tensor("tensor_131_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(438072256))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439301120))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439301312)))]; + tensor linear_127_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor var_1980 = const()[name = tensor("op_1980"), val = tensor([1, -1, 20, 64])]; + tensor var_1981_cast_fp16 = reshape(shape = var_1980, x = linear_127_cast_fp16)[name = tensor("op_1981_cast_fp16")]; + tensor var_1982_perm_0 = const()[name = tensor("op_1982_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(439303936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440532800))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440532992)))]; + tensor linear_128_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_127_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_1987 = const()[name = tensor("op_1987"), val = tensor([1, -1, 20, 64])]; + tensor var_1988_cast_fp16 = reshape(shape = var_1987, x = linear_128_cast_fp16)[name = tensor("op_1988_cast_fp16")]; + tensor var_1989_perm_0 = const()[name = tensor("op_1989_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1996 = const()[name = tensor("op_1996"), val = tensor([1, 77, 20, 64])]; + tensor var_1997_cast_fp16 = reshape(shape = var_1996, x = tensor_131_cast_fp16)[name = tensor("op_1997_cast_fp16")]; + tensor var_1998_perm_0 = const()[name = tensor("op_1998_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2000 = const()[name = tensor("op_2000"), val = tensor([20, -1, 64])]; + tensor transpose_43 = transpose(perm = var_1998_perm_0, x = var_1997_cast_fp16)[name = tensor("transpose_43")]; + tensor query_states_43_cast_fp16 = reshape(shape = var_2000, x = transpose_43)[name = tensor("query_states_43_cast_fp16")]; + tensor var_2002 = const()[name = tensor("op_2002"), val = tensor([20, -1, 64])]; + tensor transpose_42 = transpose(perm = var_1982_perm_0, x = var_1981_cast_fp16)[name = tensor("transpose_42")]; + tensor key_states_87_cast_fp16 = reshape(shape = var_2002, x = transpose_42)[name = tensor("key_states_87_cast_fp16")]; + tensor var_2004 = const()[name = tensor("op_2004"), val = tensor([20, -1, 64])]; + tensor transpose_41 = transpose(perm = var_1989_perm_0, x = var_1988_cast_fp16)[name = tensor("transpose_41")]; + tensor value_states_87_cast_fp16 = reshape(shape = var_2004, x = transpose_41)[name = tensor("value_states_87_cast_fp16")]; + tensor attn_weights_127_transpose_x_1 = const()[name = tensor("attn_weights_127_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_127_transpose_y_1 = const()[name = tensor("attn_weights_127_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_127_cast_fp16 = matmul(transpose_x = attn_weights_127_transpose_x_1, transpose_y = attn_weights_127_transpose_y_1, x = query_states_43_cast_fp16, y = key_states_87_cast_fp16)[name = tensor("attn_weights_127_cast_fp16")]; + tensor var_2009 = const()[name = tensor("op_2009"), val = tensor([1, 20, 77, 77])]; + tensor var_2010_cast_fp16 = reshape(shape = var_2009, x = attn_weights_127_cast_fp16)[name = tensor("op_2010_cast_fp16")]; + tensor attn_weights_129_cast_fp16 = add(x = var_2010_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_129_cast_fp16")]; + tensor var_2015 = const()[name = tensor("op_2015"), val = tensor([20, 77, 77])]; + tensor input_341_cast_fp16 = reshape(shape = var_2015, x = attn_weights_129_cast_fp16)[name = tensor("input_341_cast_fp16")]; + tensor input_343_cast_fp16 = softmax(axis = var_5, x = input_341_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor attn_output_127_transpose_x_0 = const()[name = tensor("attn_output_127_transpose_x_0"), val = tensor(false)]; + tensor attn_output_127_transpose_y_0 = const()[name = tensor("attn_output_127_transpose_y_0"), val = tensor(false)]; + tensor attn_output_127_cast_fp16 = matmul(transpose_x = attn_output_127_transpose_x_0, transpose_y = attn_output_127_transpose_y_0, x = input_343_cast_fp16, y = value_states_87_cast_fp16)[name = tensor("attn_output_127_cast_fp16")]; + tensor var_2020 = const()[name = tensor("op_2020"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_129_cast_fp16 = reshape(shape = var_2020, x = attn_output_127_cast_fp16)[name = tensor("attn_output_129_cast_fp16")]; + tensor attn_output_131_perm_0 = const()[name = tensor("attn_output_131_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2023 = const()[name = tensor("op_2023"), val = tensor([1, 77, 1280])]; + tensor transpose_40 = transpose(perm = attn_output_131_perm_0, x = attn_output_129_cast_fp16)[name = tensor("transpose_40")]; + tensor input_345_cast_fp16 = reshape(shape = var_2023, x = transpose_40)[name = tensor("input_345_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(440535616))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441764480))), name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441764672)))]; + tensor linear_129_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16_palettized, x = input_345_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor input_347_cast_fp16 = add(x = input_339_cast_fp16, y = linear_129_cast_fp16)[name = tensor("input_347_cast_fp16")]; + tensor input_349_axes_0 = const()[name = tensor("input_349_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441767296)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441769920)))]; + tensor input_349_cast_fp16 = layer_norm(axes = input_349_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441772544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446687808))), name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446688000))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446691904))), name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_130_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16_palettized, x = input_349_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor input_353_mode_0 = const()[name = tensor("input_353_mode_0"), val = tensor("EXACT")]; + tensor input_353_cast_fp16 = gelu(mode = input_353_mode_0, x = linear_130_cast_fp16)[name = tensor("input_353_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(446692096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451607360))), name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451607552)))]; + tensor linear_131_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16_palettized, x = input_353_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor input_355_cast_fp16 = add(x = input_347_cast_fp16, y = linear_131_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor hidden_states_133_axes_0 = const()[name = tensor("hidden_states_133_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451610176)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451612800)))]; + tensor hidden_states_133_cast_fp16 = layer_norm(axes = hidden_states_133_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("hidden_states_133_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451615424))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452844288))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452844480)))]; + tensor linear_132_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_132_cast_fp16")]; + tensor var_2062_to_fp16 = const()[name = tensor("op_2062_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_137_cast_fp16 = mul(x = linear_132_cast_fp16, y = var_2062_to_fp16)[name = tensor("tensor_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452847104))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454075968))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454076160)))]; + tensor linear_133_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor var_2067 = const()[name = tensor("op_2067"), val = tensor([1, -1, 20, 64])]; + tensor var_2068_cast_fp16 = reshape(shape = var_2067, x = linear_133_cast_fp16)[name = tensor("op_2068_cast_fp16")]; + tensor var_2069_perm_0 = const()[name = tensor("op_2069_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454078784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455307648))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455307840)))]; + tensor linear_134_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_133_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor var_2074 = const()[name = tensor("op_2074"), val = tensor([1, -1, 20, 64])]; + tensor var_2075_cast_fp16 = reshape(shape = var_2074, x = linear_134_cast_fp16)[name = tensor("op_2075_cast_fp16")]; + tensor var_2076_perm_0 = const()[name = tensor("op_2076_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2083 = const()[name = tensor("op_2083"), val = tensor([1, 77, 20, 64])]; + tensor var_2084_cast_fp16 = reshape(shape = var_2083, x = tensor_137_cast_fp16)[name = tensor("op_2084_cast_fp16")]; + tensor var_2085_perm_0 = const()[name = tensor("op_2085_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2087 = const()[name = tensor("op_2087"), val = tensor([20, -1, 64])]; + tensor transpose_39 = transpose(perm = var_2085_perm_0, x = var_2084_cast_fp16)[name = tensor("transpose_39")]; + tensor query_states_45_cast_fp16 = reshape(shape = var_2087, x = transpose_39)[name = tensor("query_states_45_cast_fp16")]; + tensor var_2089 = const()[name = tensor("op_2089"), val = tensor([20, -1, 64])]; + tensor transpose_38 = transpose(perm = var_2069_perm_0, x = var_2068_cast_fp16)[name = tensor("transpose_38")]; + tensor key_states_91_cast_fp16 = reshape(shape = var_2089, x = transpose_38)[name = tensor("key_states_91_cast_fp16")]; + tensor var_2091 = const()[name = tensor("op_2091"), val = tensor([20, -1, 64])]; + tensor transpose_37 = transpose(perm = var_2076_perm_0, x = var_2075_cast_fp16)[name = tensor("transpose_37")]; + tensor value_states_91_cast_fp16 = reshape(shape = var_2091, x = transpose_37)[name = tensor("value_states_91_cast_fp16")]; + tensor attn_weights_133_transpose_x_1 = const()[name = tensor("attn_weights_133_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_133_transpose_y_1 = const()[name = tensor("attn_weights_133_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_133_cast_fp16 = matmul(transpose_x = attn_weights_133_transpose_x_1, transpose_y = attn_weights_133_transpose_y_1, x = query_states_45_cast_fp16, y = key_states_91_cast_fp16)[name = tensor("attn_weights_133_cast_fp16")]; + tensor var_2096 = const()[name = tensor("op_2096"), val = tensor([1, 20, 77, 77])]; + tensor var_2097_cast_fp16 = reshape(shape = var_2096, x = attn_weights_133_cast_fp16)[name = tensor("op_2097_cast_fp16")]; + tensor attn_weights_135_cast_fp16 = add(x = var_2097_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_135_cast_fp16")]; + tensor var_2102 = const()[name = tensor("op_2102"), val = tensor([20, 77, 77])]; + tensor input_357_cast_fp16 = reshape(shape = var_2102, x = attn_weights_135_cast_fp16)[name = tensor("input_357_cast_fp16")]; + tensor input_359_cast_fp16 = softmax(axis = var_5, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor attn_output_133_transpose_x_0 = const()[name = tensor("attn_output_133_transpose_x_0"), val = tensor(false)]; + tensor attn_output_133_transpose_y_0 = const()[name = tensor("attn_output_133_transpose_y_0"), val = tensor(false)]; + tensor attn_output_133_cast_fp16 = matmul(transpose_x = attn_output_133_transpose_x_0, transpose_y = attn_output_133_transpose_y_0, x = input_359_cast_fp16, y = value_states_91_cast_fp16)[name = tensor("attn_output_133_cast_fp16")]; + tensor var_2107 = const()[name = tensor("op_2107"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_135_cast_fp16 = reshape(shape = var_2107, x = attn_output_133_cast_fp16)[name = tensor("attn_output_135_cast_fp16")]; + tensor attn_output_137_perm_0 = const()[name = tensor("attn_output_137_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, 77, 1280])]; + tensor transpose_36 = transpose(perm = attn_output_137_perm_0, x = attn_output_135_cast_fp16)[name = tensor("transpose_36")]; + tensor input_361_cast_fp16 = reshape(shape = var_2110, x = transpose_36)[name = tensor("input_361_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(455310464))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456539328))), name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456539520)))]; + tensor linear_135_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16_palettized, x = input_361_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor input_363_cast_fp16 = add(x = input_355_cast_fp16, y = linear_135_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor input_365_axes_0 = const()[name = tensor("input_365_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456542144)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456544768)))]; + tensor input_365_cast_fp16 = layer_norm(axes = input_365_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16, x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456547392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461462656))), name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461462848))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461466752))), name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_136_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16_palettized, x = input_365_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_369_mode_0 = const()[name = tensor("input_369_mode_0"), val = tensor("EXACT")]; + tensor input_369_cast_fp16 = gelu(mode = input_369_mode_0, x = linear_136_cast_fp16)[name = tensor("input_369_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461466944))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466382208))), name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466382400)))]; + tensor linear_137_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16_palettized, x = input_369_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor input_371_cast_fp16 = add(x = input_363_cast_fp16, y = linear_137_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor hidden_states_139_axes_0 = const()[name = tensor("hidden_states_139_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466385024)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466387648)))]; + tensor hidden_states_139_cast_fp16 = layer_norm(axes = hidden_states_139_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("hidden_states_139_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(466390272))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467619136))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467619328)))]; + tensor linear_138_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_2149_to_fp16 = const()[name = tensor("op_2149_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_143_cast_fp16 = mul(x = linear_138_cast_fp16, y = var_2149_to_fp16)[name = tensor("tensor_143_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467621952))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468850816))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468851008)))]; + tensor linear_139_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_2154 = const()[name = tensor("op_2154"), val = tensor([1, -1, 20, 64])]; + tensor var_2155_cast_fp16 = reshape(shape = var_2154, x = linear_139_cast_fp16)[name = tensor("op_2155_cast_fp16")]; + tensor var_2156_perm_0 = const()[name = tensor("op_2156_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(468853632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470082496))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470082688)))]; + tensor linear_140_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_139_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_2161 = const()[name = tensor("op_2161"), val = tensor([1, -1, 20, 64])]; + tensor var_2162_cast_fp16 = reshape(shape = var_2161, x = linear_140_cast_fp16)[name = tensor("op_2162_cast_fp16")]; + tensor var_2163_perm_0 = const()[name = tensor("op_2163_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2170 = const()[name = tensor("op_2170"), val = tensor([1, 77, 20, 64])]; + tensor var_2171_cast_fp16 = reshape(shape = var_2170, x = tensor_143_cast_fp16)[name = tensor("op_2171_cast_fp16")]; + tensor var_2172_perm_0 = const()[name = tensor("op_2172_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2174 = const()[name = tensor("op_2174"), val = tensor([20, -1, 64])]; + tensor transpose_35 = transpose(perm = var_2172_perm_0, x = var_2171_cast_fp16)[name = tensor("transpose_35")]; + tensor query_states_47_cast_fp16 = reshape(shape = var_2174, x = transpose_35)[name = tensor("query_states_47_cast_fp16")]; + tensor var_2176 = const()[name = tensor("op_2176"), val = tensor([20, -1, 64])]; + tensor transpose_34 = transpose(perm = var_2156_perm_0, x = var_2155_cast_fp16)[name = tensor("transpose_34")]; + tensor key_states_95_cast_fp16 = reshape(shape = var_2176, x = transpose_34)[name = tensor("key_states_95_cast_fp16")]; + tensor var_2178 = const()[name = tensor("op_2178"), val = tensor([20, -1, 64])]; + tensor transpose_33 = transpose(perm = var_2163_perm_0, x = var_2162_cast_fp16)[name = tensor("transpose_33")]; + tensor value_states_95_cast_fp16 = reshape(shape = var_2178, x = transpose_33)[name = tensor("value_states_95_cast_fp16")]; + tensor attn_weights_139_transpose_x_1 = const()[name = tensor("attn_weights_139_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_139_transpose_y_1 = const()[name = tensor("attn_weights_139_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_139_cast_fp16 = matmul(transpose_x = attn_weights_139_transpose_x_1, transpose_y = attn_weights_139_transpose_y_1, x = query_states_47_cast_fp16, y = key_states_95_cast_fp16)[name = tensor("attn_weights_139_cast_fp16")]; + tensor var_2183 = const()[name = tensor("op_2183"), val = tensor([1, 20, 77, 77])]; + tensor var_2184_cast_fp16 = reshape(shape = var_2183, x = attn_weights_139_cast_fp16)[name = tensor("op_2184_cast_fp16")]; + tensor attn_weights_141_cast_fp16 = add(x = var_2184_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_141_cast_fp16")]; + tensor var_2189 = const()[name = tensor("op_2189"), val = tensor([20, 77, 77])]; + tensor input_373_cast_fp16 = reshape(shape = var_2189, x = attn_weights_141_cast_fp16)[name = tensor("input_373_cast_fp16")]; + tensor input_375_cast_fp16 = softmax(axis = var_5, x = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor attn_output_139_transpose_x_0 = const()[name = tensor("attn_output_139_transpose_x_0"), val = tensor(false)]; + tensor attn_output_139_transpose_y_0 = const()[name = tensor("attn_output_139_transpose_y_0"), val = tensor(false)]; + tensor attn_output_139_cast_fp16 = matmul(transpose_x = attn_output_139_transpose_x_0, transpose_y = attn_output_139_transpose_y_0, x = input_375_cast_fp16, y = value_states_95_cast_fp16)[name = tensor("attn_output_139_cast_fp16")]; + tensor var_2194 = const()[name = tensor("op_2194"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_141_cast_fp16 = reshape(shape = var_2194, x = attn_output_139_cast_fp16)[name = tensor("attn_output_141_cast_fp16")]; + tensor attn_output_143_perm_0 = const()[name = tensor("attn_output_143_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2197 = const()[name = tensor("op_2197"), val = tensor([1, 77, 1280])]; + tensor transpose_32 = transpose(perm = attn_output_143_perm_0, x = attn_output_141_cast_fp16)[name = tensor("transpose_32")]; + tensor input_377_cast_fp16 = reshape(shape = var_2197, x = transpose_32)[name = tensor("input_377_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470085312))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471314176))), name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471314368)))]; + tensor linear_141_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16_palettized, x = input_377_cast_fp16)[name = tensor("linear_141_cast_fp16")]; + tensor input_379_cast_fp16 = add(x = input_371_cast_fp16, y = linear_141_cast_fp16)[name = tensor("input_379_cast_fp16")]; + tensor input_381_axes_0 = const()[name = tensor("input_381_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471316992)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471319616)))]; + tensor input_381_cast_fp16 = layer_norm(axes = input_381_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(471322240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476237504))), name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476237696))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476241600))), name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_142_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16_palettized, x = input_381_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_385_mode_0 = const()[name = tensor("input_385_mode_0"), val = tensor("EXACT")]; + tensor input_385_cast_fp16 = gelu(mode = input_385_mode_0, x = linear_142_cast_fp16)[name = tensor("input_385_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476241792))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481157056))), name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481157248)))]; + tensor linear_143_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16_palettized, x = input_385_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_379_cast_fp16, y = linear_143_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor hidden_states_145_axes_0 = const()[name = tensor("hidden_states_145_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481159872)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481162496)))]; + tensor hidden_states_145_cast_fp16 = layer_norm(axes = hidden_states_145_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("hidden_states_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481165120))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482393984))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482394176)))]; + tensor linear_144_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_2236_to_fp16 = const()[name = tensor("op_2236_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_149_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2236_to_fp16)[name = tensor("tensor_149_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(482396800))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483625664))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483625856)))]; + tensor linear_145_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor var_2241 = const()[name = tensor("op_2241"), val = tensor([1, -1, 20, 64])]; + tensor var_2242_cast_fp16 = reshape(shape = var_2241, x = linear_145_cast_fp16)[name = tensor("op_2242_cast_fp16")]; + tensor var_2243_perm_0 = const()[name = tensor("op_2243_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(483628480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(484857344))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(484857536)))]; + tensor linear_146_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_145_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_2248 = const()[name = tensor("op_2248"), val = tensor([1, -1, 20, 64])]; + tensor var_2249_cast_fp16 = reshape(shape = var_2248, x = linear_146_cast_fp16)[name = tensor("op_2249_cast_fp16")]; + tensor var_2250_perm_0 = const()[name = tensor("op_2250_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2257 = const()[name = tensor("op_2257"), val = tensor([1, 77, 20, 64])]; + tensor var_2258_cast_fp16 = reshape(shape = var_2257, x = tensor_149_cast_fp16)[name = tensor("op_2258_cast_fp16")]; + tensor var_2259_perm_0 = const()[name = tensor("op_2259_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2261 = const()[name = tensor("op_2261"), val = tensor([20, -1, 64])]; + tensor transpose_31 = transpose(perm = var_2259_perm_0, x = var_2258_cast_fp16)[name = tensor("transpose_31")]; + tensor query_states_49_cast_fp16 = reshape(shape = var_2261, x = transpose_31)[name = tensor("query_states_49_cast_fp16")]; + tensor var_2263 = const()[name = tensor("op_2263"), val = tensor([20, -1, 64])]; + tensor transpose_30 = transpose(perm = var_2243_perm_0, x = var_2242_cast_fp16)[name = tensor("transpose_30")]; + tensor key_states_99_cast_fp16 = reshape(shape = var_2263, x = transpose_30)[name = tensor("key_states_99_cast_fp16")]; + tensor var_2265 = const()[name = tensor("op_2265"), val = tensor([20, -1, 64])]; + tensor transpose_29 = transpose(perm = var_2250_perm_0, x = var_2249_cast_fp16)[name = tensor("transpose_29")]; + tensor value_states_99_cast_fp16 = reshape(shape = var_2265, x = transpose_29)[name = tensor("value_states_99_cast_fp16")]; + tensor attn_weights_145_transpose_x_1 = const()[name = tensor("attn_weights_145_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_145_transpose_y_1 = const()[name = tensor("attn_weights_145_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_145_cast_fp16 = matmul(transpose_x = attn_weights_145_transpose_x_1, transpose_y = attn_weights_145_transpose_y_1, x = query_states_49_cast_fp16, y = key_states_99_cast_fp16)[name = tensor("attn_weights_145_cast_fp16")]; + tensor var_2270 = const()[name = tensor("op_2270"), val = tensor([1, 20, 77, 77])]; + tensor var_2271_cast_fp16 = reshape(shape = var_2270, x = attn_weights_145_cast_fp16)[name = tensor("op_2271_cast_fp16")]; + tensor attn_weights_147_cast_fp16 = add(x = var_2271_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_147_cast_fp16")]; + tensor var_2276 = const()[name = tensor("op_2276"), val = tensor([20, 77, 77])]; + tensor input_389_cast_fp16 = reshape(shape = var_2276, x = attn_weights_147_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor input_391_cast_fp16 = softmax(axis = var_5, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor attn_output_145_transpose_x_0 = const()[name = tensor("attn_output_145_transpose_x_0"), val = tensor(false)]; + tensor attn_output_145_transpose_y_0 = const()[name = tensor("attn_output_145_transpose_y_0"), val = tensor(false)]; + tensor attn_output_145_cast_fp16 = matmul(transpose_x = attn_output_145_transpose_x_0, transpose_y = attn_output_145_transpose_y_0, x = input_391_cast_fp16, y = value_states_99_cast_fp16)[name = tensor("attn_output_145_cast_fp16")]; + tensor var_2281 = const()[name = tensor("op_2281"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_147_cast_fp16 = reshape(shape = var_2281, x = attn_output_145_cast_fp16)[name = tensor("attn_output_147_cast_fp16")]; + tensor attn_output_149_perm_0 = const()[name = tensor("attn_output_149_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, 77, 1280])]; + tensor transpose_28 = transpose(perm = attn_output_149_perm_0, x = attn_output_147_cast_fp16)[name = tensor("transpose_28")]; + tensor input_393_cast_fp16 = reshape(shape = var_2284, x = transpose_28)[name = tensor("input_393_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(484860160))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486089024))), name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486089216)))]; + tensor linear_147_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16_palettized, x = input_393_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor input_395_cast_fp16 = add(x = input_387_cast_fp16, y = linear_147_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor input_397_axes_0 = const()[name = tensor("input_397_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486091840)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486094464)))]; + tensor input_397_cast_fp16 = layer_norm(axes = input_397_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("input_397_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(486097088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491012352))), name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491012544))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491016448))), name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_148_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16_palettized, x = input_397_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor input_401_mode_0 = const()[name = tensor("input_401_mode_0"), val = tensor("EXACT")]; + tensor input_401_cast_fp16 = gelu(mode = input_401_mode_0, x = linear_148_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(491016640))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495931904))), name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495932096)))]; + tensor linear_149_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16_palettized, x = input_401_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_149_cast_fp16)[name = tensor("input_403_cast_fp16")]; + tensor hidden_states_151_axes_0 = const()[name = tensor("hidden_states_151_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495934720)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495937344)))]; + tensor hidden_states_151_cast_fp16 = layer_norm(axes = hidden_states_151_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("hidden_states_151_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(495939968))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497168832))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497169024)))]; + tensor linear_150_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_150_cast_fp16")]; + tensor var_2323_to_fp16 = const()[name = tensor("op_2323_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_155_cast_fp16 = mul(x = linear_150_cast_fp16, y = var_2323_to_fp16)[name = tensor("tensor_155_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(497171648))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498400512))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498400704)))]; + tensor linear_151_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor var_2328 = const()[name = tensor("op_2328"), val = tensor([1, -1, 20, 64])]; + tensor var_2329_cast_fp16 = reshape(shape = var_2328, x = linear_151_cast_fp16)[name = tensor("op_2329_cast_fp16")]; + tensor var_2330_perm_0 = const()[name = tensor("op_2330_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498403328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499632192))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499632384)))]; + tensor linear_152_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_151_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor var_2335 = const()[name = tensor("op_2335"), val = tensor([1, -1, 20, 64])]; + tensor var_2336_cast_fp16 = reshape(shape = var_2335, x = linear_152_cast_fp16)[name = tensor("op_2336_cast_fp16")]; + tensor var_2337_perm_0 = const()[name = tensor("op_2337_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2344 = const()[name = tensor("op_2344"), val = tensor([1, 77, 20, 64])]; + tensor var_2345_cast_fp16 = reshape(shape = var_2344, x = tensor_155_cast_fp16)[name = tensor("op_2345_cast_fp16")]; + tensor var_2346_perm_0 = const()[name = tensor("op_2346_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2348 = const()[name = tensor("op_2348"), val = tensor([20, -1, 64])]; + tensor transpose_27 = transpose(perm = var_2346_perm_0, x = var_2345_cast_fp16)[name = tensor("transpose_27")]; + tensor query_states_51_cast_fp16 = reshape(shape = var_2348, x = transpose_27)[name = tensor("query_states_51_cast_fp16")]; + tensor var_2350 = const()[name = tensor("op_2350"), val = tensor([20, -1, 64])]; + tensor transpose_26 = transpose(perm = var_2330_perm_0, x = var_2329_cast_fp16)[name = tensor("transpose_26")]; + tensor key_states_103_cast_fp16 = reshape(shape = var_2350, x = transpose_26)[name = tensor("key_states_103_cast_fp16")]; + tensor var_2352 = const()[name = tensor("op_2352"), val = tensor([20, -1, 64])]; + tensor transpose_25 = transpose(perm = var_2337_perm_0, x = var_2336_cast_fp16)[name = tensor("transpose_25")]; + tensor value_states_103_cast_fp16 = reshape(shape = var_2352, x = transpose_25)[name = tensor("value_states_103_cast_fp16")]; + tensor attn_weights_151_transpose_x_1 = const()[name = tensor("attn_weights_151_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_151_transpose_y_1 = const()[name = tensor("attn_weights_151_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_151_cast_fp16 = matmul(transpose_x = attn_weights_151_transpose_x_1, transpose_y = attn_weights_151_transpose_y_1, x = query_states_51_cast_fp16, y = key_states_103_cast_fp16)[name = tensor("attn_weights_151_cast_fp16")]; + tensor var_2357 = const()[name = tensor("op_2357"), val = tensor([1, 20, 77, 77])]; + tensor var_2358_cast_fp16 = reshape(shape = var_2357, x = attn_weights_151_cast_fp16)[name = tensor("op_2358_cast_fp16")]; + tensor attn_weights_153_cast_fp16 = add(x = var_2358_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_153_cast_fp16")]; + tensor var_2363 = const()[name = tensor("op_2363"), val = tensor([20, 77, 77])]; + tensor input_405_cast_fp16 = reshape(shape = var_2363, x = attn_weights_153_cast_fp16)[name = tensor("input_405_cast_fp16")]; + tensor input_407_cast_fp16 = softmax(axis = var_5, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor attn_output_151_transpose_x_0 = const()[name = tensor("attn_output_151_transpose_x_0"), val = tensor(false)]; + tensor attn_output_151_transpose_y_0 = const()[name = tensor("attn_output_151_transpose_y_0"), val = tensor(false)]; + tensor attn_output_151_cast_fp16 = matmul(transpose_x = attn_output_151_transpose_x_0, transpose_y = attn_output_151_transpose_y_0, x = input_407_cast_fp16, y = value_states_103_cast_fp16)[name = tensor("attn_output_151_cast_fp16")]; + tensor var_2368 = const()[name = tensor("op_2368"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_153_cast_fp16 = reshape(shape = var_2368, x = attn_output_151_cast_fp16)[name = tensor("attn_output_153_cast_fp16")]; + tensor attn_output_155_perm_0 = const()[name = tensor("attn_output_155_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2371 = const()[name = tensor("op_2371"), val = tensor([1, 77, 1280])]; + tensor transpose_24 = transpose(perm = attn_output_155_perm_0, x = attn_output_153_cast_fp16)[name = tensor("transpose_24")]; + tensor input_409_cast_fp16 = reshape(shape = var_2371, x = transpose_24)[name = tensor("input_409_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499635008))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500863872))), name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500864064)))]; + tensor linear_153_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16_palettized, x = input_409_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_403_cast_fp16, y = linear_153_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor input_413_axes_0 = const()[name = tensor("input_413_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500866688)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500869312)))]; + tensor input_413_cast_fp16 = layer_norm(axes = input_413_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("input_413_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500871936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505787200))), name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505787392))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505791296))), name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_154_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16_palettized, x = input_413_cast_fp16)[name = tensor("linear_154_cast_fp16")]; + tensor input_417_mode_0 = const()[name = tensor("input_417_mode_0"), val = tensor("EXACT")]; + tensor input_417_cast_fp16 = gelu(mode = input_417_mode_0, x = linear_154_cast_fp16)[name = tensor("input_417_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(505791488))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510706752))), name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510706944)))]; + tensor linear_155_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16_palettized, x = input_417_cast_fp16)[name = tensor("linear_155_cast_fp16")]; + tensor input_419_cast_fp16 = add(x = input_411_cast_fp16, y = linear_155_cast_fp16)[name = tensor("input_419_cast_fp16")]; + tensor hidden_states_157_axes_0 = const()[name = tensor("hidden_states_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510709568)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510712192)))]; + tensor hidden_states_157_cast_fp16 = layer_norm(axes = hidden_states_157_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("hidden_states_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510714816))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511943680))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511943872)))]; + tensor linear_156_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_156_cast_fp16")]; + tensor var_2410_to_fp16 = const()[name = tensor("op_2410_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_161_cast_fp16 = mul(x = linear_156_cast_fp16, y = var_2410_to_fp16)[name = tensor("tensor_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(511946496))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513175360))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513175552)))]; + tensor linear_157_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_157_cast_fp16")]; + tensor var_2415 = const()[name = tensor("op_2415"), val = tensor([1, -1, 20, 64])]; + tensor var_2416_cast_fp16 = reshape(shape = var_2415, x = linear_157_cast_fp16)[name = tensor("op_2416_cast_fp16")]; + tensor var_2417_perm_0 = const()[name = tensor("op_2417_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(513178176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514407040))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514407232)))]; + tensor linear_158_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_157_cast_fp16)[name = tensor("linear_158_cast_fp16")]; + tensor var_2422 = const()[name = tensor("op_2422"), val = tensor([1, -1, 20, 64])]; + tensor var_2423_cast_fp16 = reshape(shape = var_2422, x = linear_158_cast_fp16)[name = tensor("op_2423_cast_fp16")]; + tensor var_2424_perm_0 = const()[name = tensor("op_2424_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2431 = const()[name = tensor("op_2431"), val = tensor([1, 77, 20, 64])]; + tensor var_2432_cast_fp16 = reshape(shape = var_2431, x = tensor_161_cast_fp16)[name = tensor("op_2432_cast_fp16")]; + tensor var_2433_perm_0 = const()[name = tensor("op_2433_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2435 = const()[name = tensor("op_2435"), val = tensor([20, -1, 64])]; + tensor transpose_23 = transpose(perm = var_2433_perm_0, x = var_2432_cast_fp16)[name = tensor("transpose_23")]; + tensor query_states_53_cast_fp16 = reshape(shape = var_2435, x = transpose_23)[name = tensor("query_states_53_cast_fp16")]; + tensor var_2437 = const()[name = tensor("op_2437"), val = tensor([20, -1, 64])]; + tensor transpose_22 = transpose(perm = var_2417_perm_0, x = var_2416_cast_fp16)[name = tensor("transpose_22")]; + tensor key_states_107_cast_fp16 = reshape(shape = var_2437, x = transpose_22)[name = tensor("key_states_107_cast_fp16")]; + tensor var_2439 = const()[name = tensor("op_2439"), val = tensor([20, -1, 64])]; + tensor transpose_21 = transpose(perm = var_2424_perm_0, x = var_2423_cast_fp16)[name = tensor("transpose_21")]; + tensor value_states_107_cast_fp16 = reshape(shape = var_2439, x = transpose_21)[name = tensor("value_states_107_cast_fp16")]; + tensor attn_weights_157_transpose_x_1 = const()[name = tensor("attn_weights_157_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_157_transpose_y_1 = const()[name = tensor("attn_weights_157_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_157_cast_fp16 = matmul(transpose_x = attn_weights_157_transpose_x_1, transpose_y = attn_weights_157_transpose_y_1, x = query_states_53_cast_fp16, y = key_states_107_cast_fp16)[name = tensor("attn_weights_157_cast_fp16")]; + tensor var_2444 = const()[name = tensor("op_2444"), val = tensor([1, 20, 77, 77])]; + tensor var_2445_cast_fp16 = reshape(shape = var_2444, x = attn_weights_157_cast_fp16)[name = tensor("op_2445_cast_fp16")]; + tensor attn_weights_159_cast_fp16 = add(x = var_2445_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_159_cast_fp16")]; + tensor var_2450 = const()[name = tensor("op_2450"), val = tensor([20, 77, 77])]; + tensor input_421_cast_fp16 = reshape(shape = var_2450, x = attn_weights_159_cast_fp16)[name = tensor("input_421_cast_fp16")]; + tensor input_423_cast_fp16 = softmax(axis = var_5, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor attn_output_157_transpose_x_0 = const()[name = tensor("attn_output_157_transpose_x_0"), val = tensor(false)]; + tensor attn_output_157_transpose_y_0 = const()[name = tensor("attn_output_157_transpose_y_0"), val = tensor(false)]; + tensor attn_output_157_cast_fp16 = matmul(transpose_x = attn_output_157_transpose_x_0, transpose_y = attn_output_157_transpose_y_0, x = input_423_cast_fp16, y = value_states_107_cast_fp16)[name = tensor("attn_output_157_cast_fp16")]; + tensor var_2455 = const()[name = tensor("op_2455"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_159_cast_fp16 = reshape(shape = var_2455, x = attn_output_157_cast_fp16)[name = tensor("attn_output_159_cast_fp16")]; + tensor attn_output_161_perm_0 = const()[name = tensor("attn_output_161_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, 77, 1280])]; + tensor transpose_20 = transpose(perm = attn_output_161_perm_0, x = attn_output_159_cast_fp16)[name = tensor("transpose_20")]; + tensor input_425_cast_fp16 = reshape(shape = var_2458, x = transpose_20)[name = tensor("input_425_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514409856))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515638720))), name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515638912)))]; + tensor linear_159_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16_palettized, x = input_425_cast_fp16)[name = tensor("linear_159_cast_fp16")]; + tensor input_427_cast_fp16 = add(x = input_419_cast_fp16, y = linear_159_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515641536)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515644160)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(515646784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520562048))), name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520562240))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520566144))), name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_160_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16_palettized, x = input_429_cast_fp16)[name = tensor("linear_160_cast_fp16")]; + tensor input_433_mode_0 = const()[name = tensor("input_433_mode_0"), val = tensor("EXACT")]; + tensor input_433_cast_fp16 = gelu(mode = input_433_mode_0, x = linear_160_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520566336))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525481600))), name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525481792)))]; + tensor linear_161_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16_palettized, x = input_433_cast_fp16)[name = tensor("linear_161_cast_fp16")]; + tensor input_435_cast_fp16 = add(x = input_427_cast_fp16, y = linear_161_cast_fp16)[name = tensor("input_435_cast_fp16")]; + tensor hidden_states_163_axes_0 = const()[name = tensor("hidden_states_163_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525484416)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525487040)))]; + tensor hidden_states_163_cast_fp16 = layer_norm(axes = hidden_states_163_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("hidden_states_163_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525489664))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526718528))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526718720)))]; + tensor linear_162_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_162_cast_fp16")]; + tensor var_2497_to_fp16 = const()[name = tensor("op_2497_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_167_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_2497_to_fp16)[name = tensor("tensor_167_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526721344))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527950208))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527950400)))]; + tensor linear_163_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_163_cast_fp16")]; + tensor var_2502 = const()[name = tensor("op_2502"), val = tensor([1, -1, 20, 64])]; + tensor var_2503_cast_fp16 = reshape(shape = var_2502, x = linear_163_cast_fp16)[name = tensor("op_2503_cast_fp16")]; + tensor var_2504_perm_0 = const()[name = tensor("op_2504_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527953024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529181888))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529182080)))]; + tensor linear_164_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_163_cast_fp16)[name = tensor("linear_164_cast_fp16")]; + tensor var_2509 = const()[name = tensor("op_2509"), val = tensor([1, -1, 20, 64])]; + tensor var_2510_cast_fp16 = reshape(shape = var_2509, x = linear_164_cast_fp16)[name = tensor("op_2510_cast_fp16")]; + tensor var_2511_perm_0 = const()[name = tensor("op_2511_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2518 = const()[name = tensor("op_2518"), val = tensor([1, 77, 20, 64])]; + tensor var_2519_cast_fp16 = reshape(shape = var_2518, x = tensor_167_cast_fp16)[name = tensor("op_2519_cast_fp16")]; + tensor var_2520_perm_0 = const()[name = tensor("op_2520_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2522 = const()[name = tensor("op_2522"), val = tensor([20, -1, 64])]; + tensor transpose_19 = transpose(perm = var_2520_perm_0, x = var_2519_cast_fp16)[name = tensor("transpose_19")]; + tensor query_states_55_cast_fp16 = reshape(shape = var_2522, x = transpose_19)[name = tensor("query_states_55_cast_fp16")]; + tensor var_2524 = const()[name = tensor("op_2524"), val = tensor([20, -1, 64])]; + tensor transpose_18 = transpose(perm = var_2504_perm_0, x = var_2503_cast_fp16)[name = tensor("transpose_18")]; + tensor key_states_111_cast_fp16 = reshape(shape = var_2524, x = transpose_18)[name = tensor("key_states_111_cast_fp16")]; + tensor var_2526 = const()[name = tensor("op_2526"), val = tensor([20, -1, 64])]; + tensor transpose_17 = transpose(perm = var_2511_perm_0, x = var_2510_cast_fp16)[name = tensor("transpose_17")]; + tensor value_states_111_cast_fp16 = reshape(shape = var_2526, x = transpose_17)[name = tensor("value_states_111_cast_fp16")]; + tensor attn_weights_163_transpose_x_1 = const()[name = tensor("attn_weights_163_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_163_transpose_y_1 = const()[name = tensor("attn_weights_163_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_163_cast_fp16 = matmul(transpose_x = attn_weights_163_transpose_x_1, transpose_y = attn_weights_163_transpose_y_1, x = query_states_55_cast_fp16, y = key_states_111_cast_fp16)[name = tensor("attn_weights_163_cast_fp16")]; + tensor var_2531 = const()[name = tensor("op_2531"), val = tensor([1, 20, 77, 77])]; + tensor var_2532_cast_fp16 = reshape(shape = var_2531, x = attn_weights_163_cast_fp16)[name = tensor("op_2532_cast_fp16")]; + tensor attn_weights_165_cast_fp16 = add(x = var_2532_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_165_cast_fp16")]; + tensor var_2537 = const()[name = tensor("op_2537"), val = tensor([20, 77, 77])]; + tensor input_437_cast_fp16 = reshape(shape = var_2537, x = attn_weights_165_cast_fp16)[name = tensor("input_437_cast_fp16")]; + tensor input_439_cast_fp16 = softmax(axis = var_5, x = input_437_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor attn_output_163_transpose_x_0 = const()[name = tensor("attn_output_163_transpose_x_0"), val = tensor(false)]; + tensor attn_output_163_transpose_y_0 = const()[name = tensor("attn_output_163_transpose_y_0"), val = tensor(false)]; + tensor attn_output_163_cast_fp16 = matmul(transpose_x = attn_output_163_transpose_x_0, transpose_y = attn_output_163_transpose_y_0, x = input_439_cast_fp16, y = value_states_111_cast_fp16)[name = tensor("attn_output_163_cast_fp16")]; + tensor var_2542 = const()[name = tensor("op_2542"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_165_cast_fp16 = reshape(shape = var_2542, x = attn_output_163_cast_fp16)[name = tensor("attn_output_165_cast_fp16")]; + tensor attn_output_167_perm_0 = const()[name = tensor("attn_output_167_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2545 = const()[name = tensor("op_2545"), val = tensor([1, 77, 1280])]; + tensor transpose_16 = transpose(perm = attn_output_167_perm_0, x = attn_output_165_cast_fp16)[name = tensor("transpose_16")]; + tensor input_441_cast_fp16 = reshape(shape = var_2545, x = transpose_16)[name = tensor("input_441_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529184704))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530413568))), name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530413760)))]; + tensor linear_165_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16_palettized, x = input_441_cast_fp16)[name = tensor("linear_165_cast_fp16")]; + tensor input_443_cast_fp16 = add(x = input_435_cast_fp16, y = linear_165_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor input_445_axes_0 = const()[name = tensor("input_445_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530416384)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530419008)))]; + tensor input_445_cast_fp16 = layer_norm(axes = input_445_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("input_445_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530421632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535336896))), name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535337088))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535340992))), name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_166_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16_palettized, x = input_445_cast_fp16)[name = tensor("linear_166_cast_fp16")]; + tensor input_449_mode_0 = const()[name = tensor("input_449_mode_0"), val = tensor("EXACT")]; + tensor input_449_cast_fp16 = gelu(mode = input_449_mode_0, x = linear_166_cast_fp16)[name = tensor("input_449_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(535341184))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540256448))), name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540256640)))]; + tensor linear_167_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16_palettized, x = input_449_cast_fp16)[name = tensor("linear_167_cast_fp16")]; + tensor input_451_cast_fp16 = add(x = input_443_cast_fp16, y = linear_167_cast_fp16)[name = tensor("input_451_cast_fp16")]; + tensor hidden_states_169_axes_0 = const()[name = tensor("hidden_states_169_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540259264)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540261888)))]; + tensor hidden_states_169_cast_fp16 = layer_norm(axes = hidden_states_169_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("hidden_states_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(540264512))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541493376))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541493568)))]; + tensor linear_168_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_168_cast_fp16")]; + tensor var_2584_to_fp16 = const()[name = tensor("op_2584_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_173_cast_fp16 = mul(x = linear_168_cast_fp16, y = var_2584_to_fp16)[name = tensor("tensor_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(541496192))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542725056))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542725248)))]; + tensor linear_169_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_169_cast_fp16")]; + tensor var_2589 = const()[name = tensor("op_2589"), val = tensor([1, -1, 20, 64])]; + tensor var_2590_cast_fp16 = reshape(shape = var_2589, x = linear_169_cast_fp16)[name = tensor("op_2590_cast_fp16")]; + tensor var_2591_perm_0 = const()[name = tensor("op_2591_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542727872))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543956736))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543956928)))]; + tensor linear_170_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_169_cast_fp16)[name = tensor("linear_170_cast_fp16")]; + tensor var_2596 = const()[name = tensor("op_2596"), val = tensor([1, -1, 20, 64])]; + tensor var_2597_cast_fp16 = reshape(shape = var_2596, x = linear_170_cast_fp16)[name = tensor("op_2597_cast_fp16")]; + tensor var_2598_perm_0 = const()[name = tensor("op_2598_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2605 = const()[name = tensor("op_2605"), val = tensor([1, 77, 20, 64])]; + tensor var_2606_cast_fp16 = reshape(shape = var_2605, x = tensor_173_cast_fp16)[name = tensor("op_2606_cast_fp16")]; + tensor var_2607_perm_0 = const()[name = tensor("op_2607_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2609 = const()[name = tensor("op_2609"), val = tensor([20, -1, 64])]; + tensor transpose_15 = transpose(perm = var_2607_perm_0, x = var_2606_cast_fp16)[name = tensor("transpose_15")]; + tensor query_states_57_cast_fp16 = reshape(shape = var_2609, x = transpose_15)[name = tensor("query_states_57_cast_fp16")]; + tensor var_2611 = const()[name = tensor("op_2611"), val = tensor([20, -1, 64])]; + tensor transpose_14 = transpose(perm = var_2591_perm_0, x = var_2590_cast_fp16)[name = tensor("transpose_14")]; + tensor key_states_115_cast_fp16 = reshape(shape = var_2611, x = transpose_14)[name = tensor("key_states_115_cast_fp16")]; + tensor var_2613 = const()[name = tensor("op_2613"), val = tensor([20, -1, 64])]; + tensor transpose_13 = transpose(perm = var_2598_perm_0, x = var_2597_cast_fp16)[name = tensor("transpose_13")]; + tensor value_states_115_cast_fp16 = reshape(shape = var_2613, x = transpose_13)[name = tensor("value_states_115_cast_fp16")]; + tensor attn_weights_169_transpose_x_1 = const()[name = tensor("attn_weights_169_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_169_transpose_y_1 = const()[name = tensor("attn_weights_169_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_169_cast_fp16 = matmul(transpose_x = attn_weights_169_transpose_x_1, transpose_y = attn_weights_169_transpose_y_1, x = query_states_57_cast_fp16, y = key_states_115_cast_fp16)[name = tensor("attn_weights_169_cast_fp16")]; + tensor var_2618 = const()[name = tensor("op_2618"), val = tensor([1, 20, 77, 77])]; + tensor var_2619_cast_fp16 = reshape(shape = var_2618, x = attn_weights_169_cast_fp16)[name = tensor("op_2619_cast_fp16")]; + tensor attn_weights_171_cast_fp16 = add(x = var_2619_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_171_cast_fp16")]; + tensor var_2624 = const()[name = tensor("op_2624"), val = tensor([20, 77, 77])]; + tensor input_453_cast_fp16 = reshape(shape = var_2624, x = attn_weights_171_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor input_455_cast_fp16 = softmax(axis = var_5, x = input_453_cast_fp16)[name = tensor("input_455_cast_fp16")]; + tensor attn_output_169_transpose_x_0 = const()[name = tensor("attn_output_169_transpose_x_0"), val = tensor(false)]; + tensor attn_output_169_transpose_y_0 = const()[name = tensor("attn_output_169_transpose_y_0"), val = tensor(false)]; + tensor attn_output_169_cast_fp16 = matmul(transpose_x = attn_output_169_transpose_x_0, transpose_y = attn_output_169_transpose_y_0, x = input_455_cast_fp16, y = value_states_115_cast_fp16)[name = tensor("attn_output_169_cast_fp16")]; + tensor var_2629 = const()[name = tensor("op_2629"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_171_cast_fp16 = reshape(shape = var_2629, x = attn_output_169_cast_fp16)[name = tensor("attn_output_171_cast_fp16")]; + tensor attn_output_173_perm_0 = const()[name = tensor("attn_output_173_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, 77, 1280])]; + tensor transpose_12 = transpose(perm = attn_output_173_perm_0, x = attn_output_171_cast_fp16)[name = tensor("transpose_12")]; + tensor input_457_cast_fp16 = reshape(shape = var_2632, x = transpose_12)[name = tensor("input_457_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(543959552))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545188416))), name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545188608)))]; + tensor linear_171_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16_palettized, x = input_457_cast_fp16)[name = tensor("linear_171_cast_fp16")]; + tensor input_459_cast_fp16 = add(x = input_451_cast_fp16, y = linear_171_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor input_461_axes_0 = const()[name = tensor("input_461_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545191232)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545193856)))]; + tensor input_461_cast_fp16 = layer_norm(axes = input_461_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("input_461_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(545196480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550111744))), name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550111936))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550115840))), name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_172_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16_palettized, x = input_461_cast_fp16)[name = tensor("linear_172_cast_fp16")]; + tensor input_465_mode_0 = const()[name = tensor("input_465_mode_0"), val = tensor("EXACT")]; + tensor input_465_cast_fp16 = gelu(mode = input_465_mode_0, x = linear_172_cast_fp16)[name = tensor("input_465_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550116032))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555031296))), name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555031488)))]; + tensor linear_173_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16_palettized, x = input_465_cast_fp16)[name = tensor("linear_173_cast_fp16")]; + tensor input_467_cast_fp16 = add(x = input_459_cast_fp16, y = linear_173_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor hidden_states_175_axes_0 = const()[name = tensor("hidden_states_175_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555034112)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555036736)))]; + tensor hidden_states_175_cast_fp16 = layer_norm(axes = hidden_states_175_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16, x = input_467_cast_fp16)[name = tensor("hidden_states_175_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(555039360))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556268224))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556268416)))]; + tensor linear_174_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_174_cast_fp16")]; + tensor var_2671_to_fp16 = const()[name = tensor("op_2671_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_179_cast_fp16 = mul(x = linear_174_cast_fp16, y = var_2671_to_fp16)[name = tensor("tensor_179_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(556271040))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557499904))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557500096)))]; + tensor linear_175_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_175_cast_fp16")]; + tensor var_2676 = const()[name = tensor("op_2676"), val = tensor([1, -1, 20, 64])]; + tensor var_2677_cast_fp16 = reshape(shape = var_2676, x = linear_175_cast_fp16)[name = tensor("op_2677_cast_fp16")]; + tensor var_2678_perm_0 = const()[name = tensor("op_2678_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(557502720))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558731584))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558731776)))]; + tensor linear_176_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_175_cast_fp16)[name = tensor("linear_176_cast_fp16")]; + tensor var_2683 = const()[name = tensor("op_2683"), val = tensor([1, -1, 20, 64])]; + tensor var_2684_cast_fp16 = reshape(shape = var_2683, x = linear_176_cast_fp16)[name = tensor("op_2684_cast_fp16")]; + tensor var_2685_perm_0 = const()[name = tensor("op_2685_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2692 = const()[name = tensor("op_2692"), val = tensor([1, 77, 20, 64])]; + tensor var_2693_cast_fp16 = reshape(shape = var_2692, x = tensor_179_cast_fp16)[name = tensor("op_2693_cast_fp16")]; + tensor var_2694_perm_0 = const()[name = tensor("op_2694_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2696 = const()[name = tensor("op_2696"), val = tensor([20, -1, 64])]; + tensor transpose_11 = transpose(perm = var_2694_perm_0, x = var_2693_cast_fp16)[name = tensor("transpose_11")]; + tensor query_states_59_cast_fp16 = reshape(shape = var_2696, x = transpose_11)[name = tensor("query_states_59_cast_fp16")]; + tensor var_2698 = const()[name = tensor("op_2698"), val = tensor([20, -1, 64])]; + tensor transpose_10 = transpose(perm = var_2678_perm_0, x = var_2677_cast_fp16)[name = tensor("transpose_10")]; + tensor key_states_119_cast_fp16 = reshape(shape = var_2698, x = transpose_10)[name = tensor("key_states_119_cast_fp16")]; + tensor var_2700 = const()[name = tensor("op_2700"), val = tensor([20, -1, 64])]; + tensor transpose_9 = transpose(perm = var_2685_perm_0, x = var_2684_cast_fp16)[name = tensor("transpose_9")]; + tensor value_states_119_cast_fp16 = reshape(shape = var_2700, x = transpose_9)[name = tensor("value_states_119_cast_fp16")]; + tensor attn_weights_175_transpose_x_1 = const()[name = tensor("attn_weights_175_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_175_transpose_y_1 = const()[name = tensor("attn_weights_175_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_175_cast_fp16 = matmul(transpose_x = attn_weights_175_transpose_x_1, transpose_y = attn_weights_175_transpose_y_1, x = query_states_59_cast_fp16, y = key_states_119_cast_fp16)[name = tensor("attn_weights_175_cast_fp16")]; + tensor var_2705 = const()[name = tensor("op_2705"), val = tensor([1, 20, 77, 77])]; + tensor var_2706_cast_fp16 = reshape(shape = var_2705, x = attn_weights_175_cast_fp16)[name = tensor("op_2706_cast_fp16")]; + tensor attn_weights_177_cast_fp16 = add(x = var_2706_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_177_cast_fp16")]; + tensor var_2711 = const()[name = tensor("op_2711"), val = tensor([20, 77, 77])]; + tensor input_469_cast_fp16 = reshape(shape = var_2711, x = attn_weights_177_cast_fp16)[name = tensor("input_469_cast_fp16")]; + tensor input_471_cast_fp16 = softmax(axis = var_5, x = input_469_cast_fp16)[name = tensor("input_471_cast_fp16")]; + tensor attn_output_175_transpose_x_0 = const()[name = tensor("attn_output_175_transpose_x_0"), val = tensor(false)]; + tensor attn_output_175_transpose_y_0 = const()[name = tensor("attn_output_175_transpose_y_0"), val = tensor(false)]; + tensor attn_output_175_cast_fp16 = matmul(transpose_x = attn_output_175_transpose_x_0, transpose_y = attn_output_175_transpose_y_0, x = input_471_cast_fp16, y = value_states_119_cast_fp16)[name = tensor("attn_output_175_cast_fp16")]; + tensor var_2716 = const()[name = tensor("op_2716"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_177_cast_fp16 = reshape(shape = var_2716, x = attn_output_175_cast_fp16)[name = tensor("attn_output_177_cast_fp16")]; + tensor attn_output_179_perm_0 = const()[name = tensor("attn_output_179_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2719 = const()[name = tensor("op_2719"), val = tensor([1, 77, 1280])]; + tensor transpose_8 = transpose(perm = attn_output_179_perm_0, x = attn_output_177_cast_fp16)[name = tensor("transpose_8")]; + tensor input_473_cast_fp16 = reshape(shape = var_2719, x = transpose_8)[name = tensor("input_473_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558734400))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559963264))), name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559963456)))]; + tensor linear_177_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16_palettized, x = input_473_cast_fp16)[name = tensor("linear_177_cast_fp16")]; + tensor input_475_cast_fp16 = add(x = input_467_cast_fp16, y = linear_177_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor input_477_axes_0 = const()[name = tensor("input_477_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559966080)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559968704)))]; + tensor input_477_cast_fp16 = layer_norm(axes = input_477_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("input_477_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559971328))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(564886592))), name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(564886784))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(564890688))), name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_178_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16_palettized, x = input_477_cast_fp16)[name = tensor("linear_178_cast_fp16")]; + tensor input_481_mode_0 = const()[name = tensor("input_481_mode_0"), val = tensor("EXACT")]; + tensor input_481_cast_fp16 = gelu(mode = input_481_mode_0, x = linear_178_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(564890880))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569806144))), name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569806336)))]; + tensor linear_179_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16_palettized, x = input_481_cast_fp16)[name = tensor("linear_179_cast_fp16")]; + tensor input_483_cast_fp16 = add(x = input_475_cast_fp16, y = linear_179_cast_fp16)[name = tensor("input_483_cast_fp16")]; + tensor hidden_states_181_axes_0 = const()[name = tensor("hidden_states_181_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569808960)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569811584)))]; + tensor hidden_states_181_cast_fp16 = layer_norm(axes = hidden_states_181_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("hidden_states_181_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569814208))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571043072))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571043264)))]; + tensor linear_180_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_180_cast_fp16")]; + tensor var_2758_to_fp16 = const()[name = tensor("op_2758_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_185_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_2758_to_fp16)[name = tensor("tensor_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571045888))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572274752))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572274944)))]; + tensor linear_181_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_181_cast_fp16")]; + tensor var_2763 = const()[name = tensor("op_2763"), val = tensor([1, -1, 20, 64])]; + tensor var_2764_cast_fp16 = reshape(shape = var_2763, x = linear_181_cast_fp16)[name = tensor("op_2764_cast_fp16")]; + tensor var_2765_perm_0 = const()[name = tensor("op_2765_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572277568))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573506432))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573506624)))]; + tensor linear_182_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_181_cast_fp16)[name = tensor("linear_182_cast_fp16")]; + tensor var_2770 = const()[name = tensor("op_2770"), val = tensor([1, -1, 20, 64])]; + tensor var_2771_cast_fp16 = reshape(shape = var_2770, x = linear_182_cast_fp16)[name = tensor("op_2771_cast_fp16")]; + tensor var_2772_perm_0 = const()[name = tensor("op_2772_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2779 = const()[name = tensor("op_2779"), val = tensor([1, 77, 20, 64])]; + tensor var_2780_cast_fp16 = reshape(shape = var_2779, x = tensor_185_cast_fp16)[name = tensor("op_2780_cast_fp16")]; + tensor var_2781_perm_0 = const()[name = tensor("op_2781_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2783 = const()[name = tensor("op_2783"), val = tensor([20, -1, 64])]; + tensor transpose_7 = transpose(perm = var_2781_perm_0, x = var_2780_cast_fp16)[name = tensor("transpose_7")]; + tensor query_states_61_cast_fp16 = reshape(shape = var_2783, x = transpose_7)[name = tensor("query_states_61_cast_fp16")]; + tensor var_2785 = const()[name = tensor("op_2785"), val = tensor([20, -1, 64])]; + tensor transpose_6 = transpose(perm = var_2765_perm_0, x = var_2764_cast_fp16)[name = tensor("transpose_6")]; + tensor key_states_123_cast_fp16 = reshape(shape = var_2785, x = transpose_6)[name = tensor("key_states_123_cast_fp16")]; + tensor var_2787 = const()[name = tensor("op_2787"), val = tensor([20, -1, 64])]; + tensor transpose_5 = transpose(perm = var_2772_perm_0, x = var_2771_cast_fp16)[name = tensor("transpose_5")]; + tensor value_states_123_cast_fp16 = reshape(shape = var_2787, x = transpose_5)[name = tensor("value_states_123_cast_fp16")]; + tensor attn_weights_181_transpose_x_1 = const()[name = tensor("attn_weights_181_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_181_transpose_y_1 = const()[name = tensor("attn_weights_181_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_181_cast_fp16 = matmul(transpose_x = attn_weights_181_transpose_x_1, transpose_y = attn_weights_181_transpose_y_1, x = query_states_61_cast_fp16, y = key_states_123_cast_fp16)[name = tensor("attn_weights_181_cast_fp16")]; + tensor var_2792 = const()[name = tensor("op_2792"), val = tensor([1, 20, 77, 77])]; + tensor var_2793_cast_fp16 = reshape(shape = var_2792, x = attn_weights_181_cast_fp16)[name = tensor("op_2793_cast_fp16")]; + tensor attn_weights_183_cast_fp16 = add(x = var_2793_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_183_cast_fp16")]; + tensor var_2798 = const()[name = tensor("op_2798"), val = tensor([20, 77, 77])]; + tensor input_485_cast_fp16 = reshape(shape = var_2798, x = attn_weights_183_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor input_487_cast_fp16 = softmax(axis = var_5, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; + tensor attn_output_181_transpose_x_0 = const()[name = tensor("attn_output_181_transpose_x_0"), val = tensor(false)]; + tensor attn_output_181_transpose_y_0 = const()[name = tensor("attn_output_181_transpose_y_0"), val = tensor(false)]; + tensor attn_output_181_cast_fp16 = matmul(transpose_x = attn_output_181_transpose_x_0, transpose_y = attn_output_181_transpose_y_0, x = input_487_cast_fp16, y = value_states_123_cast_fp16)[name = tensor("attn_output_181_cast_fp16")]; + tensor var_2803 = const()[name = tensor("op_2803"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_183_cast_fp16 = reshape(shape = var_2803, x = attn_output_181_cast_fp16)[name = tensor("attn_output_183_cast_fp16")]; + tensor attn_output_185_perm_0 = const()[name = tensor("attn_output_185_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, 77, 1280])]; + tensor transpose_4 = transpose(perm = attn_output_185_perm_0, x = attn_output_183_cast_fp16)[name = tensor("transpose_4")]; + tensor input_489_cast_fp16 = reshape(shape = var_2806, x = transpose_4)[name = tensor("input_489_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573509248))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574738112))), name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574738304)))]; + tensor linear_183_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16_palettized, x = input_489_cast_fp16)[name = tensor("linear_183_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_483_cast_fp16, y = linear_183_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574740928)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574743552)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574746176))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579661440))), name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579661632))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579665536))), name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_184_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16_palettized, x = input_493_cast_fp16)[name = tensor("linear_184_cast_fp16")]; + tensor input_497_mode_0 = const()[name = tensor("input_497_mode_0"), val = tensor("EXACT")]; + tensor input_497_cast_fp16 = gelu(mode = input_497_mode_0, x = linear_184_cast_fp16)[name = tensor("input_497_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(579665728))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584580992))), name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584581184)))]; + tensor linear_185_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16_palettized, x = input_497_cast_fp16)[name = tensor("linear_185_cast_fp16")]; + tensor input_499_cast_fp16 = add(x = input_491_cast_fp16, y = linear_185_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor input_499_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("input_499_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor hidden_states_187_axes_0 = const()[name = tensor("hidden_states_187_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584583808)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584586432)))]; + tensor hidden_states_187_cast_fp16 = layer_norm(axes = hidden_states_187_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("hidden_states_187_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(584589056))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585817920))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585818112)))]; + tensor linear_186_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_186_cast_fp16")]; + tensor var_2845_to_fp16 = const()[name = tensor("op_2845_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_cast_fp16 = mul(x = linear_186_cast_fp16, y = var_2845_to_fp16)[name = tensor("tensor_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585820736))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587049600))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587049792)))]; + tensor linear_187_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_187_cast_fp16")]; + tensor var_2850 = const()[name = tensor("op_2850"), val = tensor([1, -1, 20, 64])]; + tensor var_2851_cast_fp16 = reshape(shape = var_2850, x = linear_187_cast_fp16)[name = tensor("op_2851_cast_fp16")]; + tensor var_2852_perm_0 = const()[name = tensor("op_2852_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587052416))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588281280))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588281472)))]; + tensor linear_188_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16_palettized, x = hidden_states_187_cast_fp16)[name = tensor("linear_188_cast_fp16")]; + tensor var_2857 = const()[name = tensor("op_2857"), val = tensor([1, -1, 20, 64])]; + tensor var_2858_cast_fp16 = reshape(shape = var_2857, x = linear_188_cast_fp16)[name = tensor("op_2858_cast_fp16")]; + tensor var_2859_perm_0 = const()[name = tensor("op_2859_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2866 = const()[name = tensor("op_2866"), val = tensor([1, 77, 20, 64])]; + tensor var_2867_cast_fp16 = reshape(shape = var_2866, x = tensor_cast_fp16)[name = tensor("op_2867_cast_fp16")]; + tensor var_2868_perm_0 = const()[name = tensor("op_2868_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2870 = const()[name = tensor("op_2870"), val = tensor([20, -1, 64])]; + tensor transpose_3 = transpose(perm = var_2868_perm_0, x = var_2867_cast_fp16)[name = tensor("transpose_3")]; + tensor query_states_cast_fp16 = reshape(shape = var_2870, x = transpose_3)[name = tensor("query_states_cast_fp16")]; + tensor var_2872 = const()[name = tensor("op_2872"), val = tensor([20, -1, 64])]; + tensor transpose_2 = transpose(perm = var_2852_perm_0, x = var_2851_cast_fp16)[name = tensor("transpose_2")]; + tensor key_states_cast_fp16 = reshape(shape = var_2872, x = transpose_2)[name = tensor("key_states_cast_fp16")]; + tensor var_2874 = const()[name = tensor("op_2874"), val = tensor([20, -1, 64])]; + tensor transpose_1 = transpose(perm = var_2859_perm_0, x = var_2858_cast_fp16)[name = tensor("transpose_1")]; + tensor value_states_cast_fp16 = reshape(shape = var_2874, x = transpose_1)[name = tensor("value_states_cast_fp16")]; + tensor attn_weights_187_transpose_x_1 = const()[name = tensor("attn_weights_187_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_187_transpose_y_1 = const()[name = tensor("attn_weights_187_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_187_cast_fp16 = matmul(transpose_x = attn_weights_187_transpose_x_1, transpose_y = attn_weights_187_transpose_y_1, x = query_states_cast_fp16, y = key_states_cast_fp16)[name = tensor("attn_weights_187_cast_fp16")]; + tensor var_2879 = const()[name = tensor("op_2879"), val = tensor([1, 20, 77, 77])]; + tensor var_2880_cast_fp16 = reshape(shape = var_2879, x = attn_weights_187_cast_fp16)[name = tensor("op_2880_cast_fp16")]; + tensor attn_weights_189_cast_fp16 = add(x = var_2880_cast_fp16, y = op_58_to_fp16_palettized)[name = tensor("attn_weights_189_cast_fp16")]; + tensor var_2885 = const()[name = tensor("op_2885"), val = tensor([20, 77, 77])]; + tensor input_501_cast_fp16 = reshape(shape = var_2885, x = attn_weights_189_cast_fp16)[name = tensor("input_501_cast_fp16")]; + tensor input_503_cast_fp16 = softmax(axis = var_5, x = input_501_cast_fp16)[name = tensor("input_503_cast_fp16")]; + tensor attn_output_187_transpose_x_0 = const()[name = tensor("attn_output_187_transpose_x_0"), val = tensor(false)]; + tensor attn_output_187_transpose_y_0 = const()[name = tensor("attn_output_187_transpose_y_0"), val = tensor(false)]; + tensor attn_output_187_cast_fp16 = matmul(transpose_x = attn_output_187_transpose_x_0, transpose_y = attn_output_187_transpose_y_0, x = input_503_cast_fp16, y = value_states_cast_fp16)[name = tensor("attn_output_187_cast_fp16")]; + tensor var_2890 = const()[name = tensor("op_2890"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_189_cast_fp16 = reshape(shape = var_2890, x = attn_output_187_cast_fp16)[name = tensor("attn_output_189_cast_fp16")]; + tensor attn_output_perm_0 = const()[name = tensor("attn_output_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2893 = const()[name = tensor("op_2893"), val = tensor([1, 77, 1280])]; + tensor transpose_0 = transpose(perm = attn_output_perm_0, x = attn_output_189_cast_fp16)[name = tensor("transpose_0")]; + tensor input_505_cast_fp16 = reshape(shape = var_2893, x = transpose_0)[name = tensor("input_505_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588284096))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589512960))), name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589513152)))]; + tensor linear_189_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16_palettized, x = input_505_cast_fp16)[name = tensor("linear_189_cast_fp16")]; + tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_189_cast_fp16)[name = tensor("input_507_cast_fp16")]; + tensor input_509_axes_0 = const()[name = tensor("input_509_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589515776)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589518400)))]; + tensor input_509_cast_fp16 = layer_norm(axes = input_509_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("input_509_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589521024))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594436288))), name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized"), shape = tensor([5120, 1280])]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594436480))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594440384))), name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16_palettized"), shape = tensor([5120])]; + tensor linear_190_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16_palettized, weight = text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16_palettized, x = input_509_cast_fp16)[name = tensor("linear_190_cast_fp16")]; + tensor input_513_mode_0 = const()[name = tensor("input_513_mode_0"), val = tensor("EXACT")]; + tensor input_513_cast_fp16 = gelu(mode = input_513_mode_0, x = linear_190_cast_fp16)[name = tensor("input_513_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(594440576))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599355840))), name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized"), shape = tensor([1280, 5120])]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599356032)))]; + tensor linear_191_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16_palettized, x = input_513_cast_fp16)[name = tensor("linear_191_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_507_cast_fp16, y = linear_191_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor last_hidden_state_axes_0 = const()[name = tensor("last_hidden_state_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_final_layer_norm_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599358656)))]; + tensor text_encoder_text_model_final_layer_norm_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599361280)))]; + tensor last_hidden_state_cast_fp16 = layer_norm(axes = last_hidden_state_axes_0, beta = text_encoder_text_model_final_layer_norm_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_final_layer_norm_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("last_hidden_state_cast_fp16")]; + tensor var_2921 = const()[name = tensor("op_2921"), val = tensor([0])]; + tensor var_2923 = reduce_argmax(axis = var_5, keep_dims = var_6, x = cast_2)[name = tensor("op_2923")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (var_2921, var_2923))[name = tensor("stack_0")]; + tensor input_transpose_batch_dims_0 = const()[name = tensor("input_transpose_batch_dims_0"), val = tensor(0)]; + tensor input_transpose_cast_fp16 = gather_nd(batch_dims = input_transpose_batch_dims_0, indices = stack_0, x = last_hidden_state_cast_fp16)[name = tensor("input_transpose_cast_fp16")]; + tensor text_encoder_text_projection_weight_to_fp16_palettized = constexpr_lut_to_dense()[indices = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(599363904))), lut = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(600592768))), name = tensor("text_encoder_text_projection_weight_to_fp16_palettized"), shape = tensor([1280, 1280])]; + tensor linear_192_bias_0_to_fp16 = const()[name = tensor("linear_192_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(600592960)))]; + tensor linear_192_cast_fp16 = linear(bias = linear_192_bias_0_to_fp16, weight = text_encoder_text_projection_weight_to_fp16_palettized, x = input_transpose_cast_fp16)[name = tensor("linear_192_cast_fp16")]; + tensor linear_192_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("linear_192_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor hidden_embeds = cast(dtype = input_499_cast_fp16_to_fp32_dtype_0, x = input_499_cast_fp16)[name = tensor("cast_0")]; + tensor pooled_outputs = cast(dtype = linear_192_cast_fp16_to_fp32_dtype_0, x = linear_192_cast_fp16)[name = tensor("cast_1")]; + } -> (hidden_embeds, pooled_outputs); +} \ No newline at end of file