program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3520.2.1"}, {"coremlc-version", "3520.2.1"}, {"mldb_token", "mldb-ndg4j6nglq"}})] { func main(tensor input_wav) { tensor feats_axes_0 = const()[name = tensor("feats_axes_0"), val = tensor([0])]; tensor feats_cast_fp16 = expand_dims(axes = feats_axes_0, x = input_wav)[name = tensor("feats_cast_fp16")]; tensor var_5 = const()[name = tensor("op_5"), val = tensor(-1)]; tensor var_13 = const()[name = tensor("op_13"), val = tensor(144)]; tensor var_15 = const()[name = tensor("op_15"), val = tensor(1)]; tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([1])]; tensor input_1_cast_fp16 = expand_dims(axes = input_1_axes_0, x = feats_cast_fp16)[name = tensor("input_1_cast_fp16")]; tensor var_41 = const()[name = tensor("op_41"), val = tensor([2, 2])]; tensor var_43 = const()[name = tensor("op_43"), val = tensor([1, 1])]; tensor input_3_pad_type_0 = const()[name = tensor("input_3_pad_type_0"), val = tensor("custom")]; tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0])]; tensor encoder_conv_subsample_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_conv_subsample_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor encoder_conv_subsample_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_conv_subsample_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2752)))]; tensor input_3_cast_fp16 = conv(bias = encoder_conv_subsample_sequential_0_bias_to_fp16, dilations = var_43, groups = var_15, pad = input_3_pad_0, pad_type = input_3_pad_type_0, strides = var_41, weight = encoder_conv_subsample_sequential_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; tensor input_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor var_49 = const()[name = tensor("op_49"), val = tensor([2, 2])]; tensor var_51 = const()[name = tensor("op_51"), val = tensor([1, 1])]; tensor input_7_pad_type_0 = const()[name = tensor("input_7_pad_type_0"), val = tensor("custom")]; tensor input_7_pad_0 = const()[name = tensor("input_7_pad_0"), val = tensor([0, 0, 0, 0])]; tensor encoder_conv_subsample_sequential_2_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_conv_subsample_sequential_2_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190080))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189824)))]; tensor encoder_conv_subsample_sequential_2_bias_to_fp16 = const()[name = tensor("encoder_conv_subsample_sequential_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190464)))]; tensor input_7_cast_fp16 = conv(bias = encoder_conv_subsample_sequential_2_bias_to_fp16, dilations = var_51, groups = var_15, pad = input_7_pad_0, pad_type = input_7_pad_type_0, strides = var_49, weight = encoder_conv_subsample_sequential_2_weight_to_fp16_affine_quantized, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor outputs_1_cast_fp16 = relu(x = input_7_cast_fp16)[name = tensor("outputs_1_cast_fp16")]; tensor var_61 = const()[name = tensor("op_61"), val = tensor([0, 2, 1, 3])]; tensor var_66 = const()[name = tensor("op_66"), val = tensor([1, 48, 1296])]; tensor transpose_78 = transpose(perm = var_61, x = outputs_1_cast_fp16)[name = tensor("transpose_78")]; tensor input_9_cast_fp16 = reshape(shape = var_66, x = transpose_78)[name = tensor("input_9_cast_fp16")]; tensor encoder_input_projection_0_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_input_projection_0_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377792))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377536)))]; tensor encoder_input_projection_0_linear_bias_to_fp16 = const()[name = tensor("encoder_input_projection_0_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378176)))]; tensor linear_0_cast_fp16 = linear(bias = encoder_input_projection_0_linear_bias_to_fp16, weight = encoder_input_projection_0_linear_weight_to_fp16_affine_quantized, x = input_9_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor input_15_axes_0 = const()[name = tensor("input_15_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378560)))]; tensor encoder_layers_0_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378944)))]; tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(0x1.5p-17)]; tensor input_15_cast_fp16 = layer_norm(axes = input_15_axes_0, beta = encoder_layers_0_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_0_sequential_0_module_sequential_0_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor encoder_layers_0_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(462336)))]; tensor encoder_layers_0_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(464192)))]; tensor linear_1_cast_fp16 = linear(bias = encoder_layers_0_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_0_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_15_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_17_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor encoder_layers_0_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548416)))]; tensor encoder_layers_0_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549056)))]; tensor linear_2_cast_fp16 = linear(bias = encoder_layers_0_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_0_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_17_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor var_101_to_fp16 = const()[name = tensor("op_101_to_fp16"), val = tensor(0x1p-1)]; tensor var_102_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_101_to_fp16)[name = tensor("op_102_cast_fp16")]; tensor inputs_3_cast_fp16 = add(x = var_102_cast_fp16, y = linear_0_cast_fp16)[name = tensor("inputs_3_cast_fp16")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549440)))]; tensor encoder_layers_0_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549824)))]; tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = encoder_layers_0_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_0_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_3_cast_fp16)[name = tensor("query_1_cast_fp16")]; tensor encoder_layers_0_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550208))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571264))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571008)))]; tensor encoder_layers_0_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571648)))]; tensor linear_3_cast_fp16 = linear(bias = encoder_layers_0_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_0_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor var_133 = const()[name = tensor("op_133"), val = tensor([1, -1, 4, 36])]; tensor query_3_cast_fp16 = reshape(shape = var_133, x = linear_3_cast_fp16)[name = tensor("query_3_cast_fp16")]; tensor encoder_layers_0_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572032))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(592832)))]; tensor encoder_layers_0_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593472)))]; tensor linear_4_cast_fp16 = linear(bias = encoder_layers_0_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_0_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; tensor var_139 = const()[name = tensor("op_139"), val = tensor([1, -1, 4, 36])]; tensor var_140_cast_fp16 = reshape(shape = var_139, x = linear_4_cast_fp16)[name = tensor("op_140_cast_fp16")]; tensor encoder_layers_0_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(593856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(614912))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(614656)))]; tensor encoder_layers_0_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(615296)))]; tensor linear_5_cast_fp16 = linear(bias = encoder_layers_0_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_0_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; tensor var_147 = const()[name = tensor("op_147"), val = tensor([1, -1, 4, 36])]; tensor var_148_cast_fp16 = reshape(shape = var_147, x = linear_5_cast_fp16)[name = tensor("op_148_cast_fp16")]; tensor var_149 = const()[name = tensor("op_149"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_0_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(615680)))]; tensor var_156_cast_fp16 = add(x = query_3_cast_fp16, y = encoder_layers_0_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_156_cast_fp16")]; tensor content_score_1_transpose_x_0 = const()[name = tensor("content_score_1_transpose_x_0"), val = tensor(false)]; tensor content_score_1_transpose_y_0 = const()[name = tensor("content_score_1_transpose_y_0"), val = tensor(false)]; tensor transpose_24_perm_0 = const()[name = tensor("transpose_24_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_25_perm_0 = const()[name = tensor("transpose_25_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_76 = transpose(perm = transpose_25_perm_0, x = var_140_cast_fp16)[name = tensor("transpose_76")]; tensor transpose_77 = transpose(perm = transpose_24_perm_0, x = var_156_cast_fp16)[name = tensor("transpose_77")]; tensor content_score_1_cast_fp16 = matmul(transpose_x = content_score_1_transpose_x_0, transpose_y = content_score_1_transpose_y_0, x = transpose_77, y = transpose_76)[name = tensor("content_score_1_cast_fp16")]; tensor encoder_layers_0_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(616064)))]; tensor var_160_cast_fp16 = add(x = query_3_cast_fp16, y = encoder_layers_0_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_160_cast_fp16")]; tensor var_161_perm_0 = const()[name = tensor("op_161_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_1_transpose_x_0 = const()[name = tensor("pos_score_1_transpose_x_0"), val = tensor(false)]; tensor pos_score_1_transpose_y_0 = const()[name = tensor("pos_score_1_transpose_y_0"), val = tensor(false)]; tensor op_163_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_163_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(616448))), scale = tensor(0x1.698p-6), zero_point = tensor(12)]; tensor transpose_75 = transpose(perm = var_161_perm_0, x = var_160_cast_fp16)[name = tensor("transpose_75")]; tensor pos_score_1_cast_fp16 = matmul(transpose_x = pos_score_1_transpose_x_0, transpose_y = pos_score_1_transpose_y_0, x = transpose_75, y = op_163_to_fp16_affine_quantized)[name = tensor("pos_score_1_cast_fp16")]; tensor padded_pos_score_1_interleave_0 = const()[name = tensor("padded_pos_score_1_interleave_0"), val = tensor(false)]; tensor zeros_1_to_fp16 = const()[name = tensor("zeros_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(623424)))]; tensor padded_pos_score_1_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_1_interleave_0, values = (zeros_1_to_fp16, pos_score_1_cast_fp16))[name = tensor("padded_pos_score_1_cast_fp16")]; tensor var_177 = const()[name = tensor("op_177"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_3_cast_fp16 = reshape(shape = var_177, x = padded_pos_score_1_cast_fp16)[name = tensor("padded_pos_score_3_cast_fp16")]; tensor var_181_begin_0 = const()[name = tensor("op_181_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_181_end_0 = const()[name = tensor("op_181_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_181_end_mask_0 = const()[name = tensor("op_181_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_181_cast_fp16 = slice_by_index(begin = var_181_begin_0, end = var_181_end_0, end_mask = var_181_end_mask_0, x = padded_pos_score_3_cast_fp16)[name = tensor("op_181_cast_fp16")]; tensor var_188_cast_fp16 = add(x = content_score_1_cast_fp16, y = var_181_cast_fp16)[name = tensor("op_188_cast_fp16")]; tensor _inversed_input_25_y_0_to_fp16 = const()[name = tensor("_inversed_input_25_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_25_cast_fp16 = mul(x = var_188_cast_fp16, y = _inversed_input_25_y_0_to_fp16)[name = tensor("_inversed_input_25_cast_fp16")]; tensor input_27_cast_fp16 = softmax(axis = var_5, x = _inversed_input_25_cast_fp16)[name = tensor("input_27_cast_fp16")]; tensor var_193_transpose_x_0 = const()[name = tensor("op_193_transpose_x_0"), val = tensor(false)]; tensor var_193_transpose_y_0 = const()[name = tensor("op_193_transpose_y_0"), val = tensor(false)]; tensor transpose_74 = transpose(perm = var_149, x = var_148_cast_fp16)[name = tensor("transpose_74")]; tensor var_193_cast_fp16 = matmul(transpose_x = var_193_transpose_x_0, transpose_y = var_193_transpose_y_0, x = input_27_cast_fp16, y = transpose_74)[name = tensor("op_193_cast_fp16")]; tensor context_1_perm_0 = const()[name = tensor("context_1_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_196 = const()[name = tensor("op_196"), val = tensor([1, -1, 144])]; tensor transpose_73 = transpose(perm = context_1_perm_0, x = var_193_cast_fp16)[name = tensor("transpose_73")]; tensor input_29_cast_fp16 = reshape(shape = var_196, x = transpose_73)[name = tensor("input_29_cast_fp16")]; tensor encoder_layers_0_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(623872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(644928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(644672)))]; tensor encoder_layers_0_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645312)))]; tensor linear_7_cast_fp16 = linear(bias = encoder_layers_0_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_0_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_29_cast_fp16)[name = tensor("linear_7_cast_fp16")]; tensor input_33_cast_fp16 = add(x = linear_7_cast_fp16, y = inputs_3_cast_fp16)[name = tensor("input_33_cast_fp16")]; tensor x_1_axes_0 = const()[name = tensor("x_1_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(645696)))]; tensor encoder_layers_0_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646080)))]; tensor x_1_cast_fp16 = layer_norm(axes = x_1_axes_0, beta = encoder_layers_0_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_0_sequential_2_module_sequential_0_weight_to_fp16, x = input_33_cast_fp16)[name = tensor("x_1_cast_fp16")]; tensor input_35_perm_0 = const()[name = tensor("input_35_perm_0"), val = tensor([0, 2, 1])]; tensor var_223 = const()[name = tensor("op_223"), val = tensor([1])]; tensor var_225 = const()[name = tensor("op_225"), val = tensor([1])]; tensor inputs_5_pad_type_0 = const()[name = tensor("inputs_5_pad_type_0"), val = tensor("custom")]; tensor inputs_5_pad_0 = const()[name = tensor("inputs_5_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_0_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(646464))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(688384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(688000)))]; tensor encoder_layers_0_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(689024)))]; tensor transpose_72 = transpose(perm = input_35_perm_0, x = x_1_cast_fp16)[name = tensor("transpose_72")]; tensor inputs_5_cast_fp16 = conv(bias = encoder_layers_0_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_225, groups = var_15, pad = inputs_5_pad_0, pad_type = inputs_5_pad_type_0, strides = var_223, weight = encoder_layers_0_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_72)[name = tensor("inputs_5_cast_fp16")]; tensor var_228_split_sizes_0 = const()[name = tensor("op_228_split_sizes_0"), val = tensor([144, 144])]; tensor var_228_axis_0 = const()[name = tensor("op_228_axis_0"), val = tensor(1)]; tensor var_228_cast_fp16_0, tensor var_228_cast_fp16_1 = split(axis = var_228_axis_0, split_sizes = var_228_split_sizes_0, x = inputs_5_cast_fp16)[name = tensor("op_228_cast_fp16")]; tensor var_230_cast_fp16 = sigmoid(x = var_228_cast_fp16_1)[name = tensor("op_230_cast_fp16")]; tensor input_37_cast_fp16 = mul(x = var_228_cast_fp16_0, y = var_230_cast_fp16)[name = tensor("input_37_cast_fp16")]; tensor var_234 = const()[name = tensor("op_234"), val = tensor([1])]; tensor var_236 = const()[name = tensor("op_236"), val = tensor([1])]; tensor input_39_pad_type_0 = const()[name = tensor("input_39_pad_type_0"), val = tensor("custom")]; tensor input_39_pad_0 = const()[name = tensor("input_39_pad_0"), val = tensor([1, 1])]; tensor inputs_7_weight_0_to_fp16 = const()[name = tensor("inputs_7_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(689664)))]; tensor inputs_7_bias_0_to_fp16 = const()[name = tensor("inputs_7_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(690624)))]; tensor inputs_7_cast_fp16 = conv(bias = inputs_7_bias_0_to_fp16, dilations = var_236, groups = var_13, pad = input_39_pad_0, pad_type = input_39_pad_type_0, strides = var_234, weight = inputs_7_weight_0_to_fp16, x = input_37_cast_fp16)[name = tensor("inputs_7_cast_fp16")]; tensor input_41_cast_fp16 = silu(x = inputs_7_cast_fp16)[name = tensor("input_41_cast_fp16")]; tensor var_249 = const()[name = tensor("op_249"), val = tensor([1])]; tensor var_251 = const()[name = tensor("op_251"), val = tensor([1])]; tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("custom")]; tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_0_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(691008))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(711808)))]; tensor encoder_layers_0_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712448)))]; tensor input_43_cast_fp16 = conv(bias = encoder_layers_0_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_251, groups = var_15, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = var_249, weight = encoder_layers_0_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; tensor var_255_perm_0 = const()[name = tensor("op_255_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_71 = transpose(perm = var_255_perm_0, x = input_43_cast_fp16)[name = tensor("transpose_71")]; tensor input_45_cast_fp16 = add(x = transpose_71, y = input_33_cast_fp16)[name = tensor("input_45_cast_fp16")]; tensor input_47_axes_0 = const()[name = tensor("input_47_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(712832)))]; tensor encoder_layers_0_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(713216)))]; tensor input_47_cast_fp16 = layer_norm(axes = input_47_axes_0, beta = encoder_layers_0_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_0_sequential_3_module_sequential_0_weight_to_fp16, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; tensor encoder_layers_0_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(713600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(797248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(796608)))]; tensor encoder_layers_0_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(798464)))]; tensor linear_8_cast_fp16 = linear(bias = encoder_layers_0_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_0_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_47_cast_fp16)[name = tensor("linear_8_cast_fp16")]; tensor input_49_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_49_cast_fp16")]; tensor encoder_layers_0_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_0_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(799680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(882944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(882688)))]; tensor encoder_layers_0_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(883328)))]; tensor linear_9_cast_fp16 = linear(bias = encoder_layers_0_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_0_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_49_cast_fp16)[name = tensor("linear_9_cast_fp16")]; tensor var_282_to_fp16 = const()[name = tensor("op_282_to_fp16"), val = tensor(0x1p-1)]; tensor var_283_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_282_to_fp16)[name = tensor("op_283_cast_fp16")]; tensor input_55_cast_fp16 = add(x = var_283_cast_fp16, y = input_45_cast_fp16)[name = tensor("input_55_cast_fp16")]; tensor input_57_axes_0 = const()[name = tensor("input_57_axes_0"), val = tensor([-1])]; tensor encoder_layers_0_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(883712)))]; tensor encoder_layers_0_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_0_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884096)))]; tensor input_57_cast_fp16 = layer_norm(axes = input_57_axes_0, beta = encoder_layers_0_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_0_sequential_4_weight_to_fp16, x = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884480)))]; tensor encoder_layers_1_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884864)))]; tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = encoder_layers_1_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_1_sequential_0_module_sequential_0_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; tensor encoder_layers_1_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(885248))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(968896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(968256)))]; tensor encoder_layers_1_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(970112)))]; tensor linear_10_cast_fp16 = linear(bias = encoder_layers_1_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_1_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_59_cast_fp16)[name = tensor("linear_10_cast_fp16")]; tensor input_61_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_61_cast_fp16")]; tensor encoder_layers_1_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(971328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1054592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1054336)))]; tensor encoder_layers_1_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1054976)))]; tensor linear_11_cast_fp16 = linear(bias = encoder_layers_1_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_1_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_61_cast_fp16)[name = tensor("linear_11_cast_fp16")]; tensor var_318_to_fp16 = const()[name = tensor("op_318_to_fp16"), val = tensor(0x1p-1)]; tensor var_319_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_318_to_fp16)[name = tensor("op_319_cast_fp16")]; tensor inputs_13_cast_fp16 = add(x = var_319_cast_fp16, y = input_57_cast_fp16)[name = tensor("inputs_13_cast_fp16")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055360)))]; tensor encoder_layers_1_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1055744)))]; tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = encoder_layers_1_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_1_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_13_cast_fp16)[name = tensor("query_5_cast_fp16")]; tensor encoder_layers_1_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1056128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1076928)))]; tensor encoder_layers_1_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077568)))]; tensor linear_12_cast_fp16 = linear(bias = encoder_layers_1_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_1_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_5_cast_fp16)[name = tensor("linear_12_cast_fp16")]; tensor var_350 = const()[name = tensor("op_350"), val = tensor([1, -1, 4, 36])]; tensor query_7_cast_fp16 = reshape(shape = var_350, x = linear_12_cast_fp16)[name = tensor("query_7_cast_fp16")]; tensor encoder_layers_1_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1099008))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1098752)))]; tensor encoder_layers_1_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1099392)))]; tensor linear_13_cast_fp16 = linear(bias = encoder_layers_1_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_1_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_5_cast_fp16)[name = tensor("linear_13_cast_fp16")]; tensor var_356 = const()[name = tensor("op_356"), val = tensor([1, -1, 4, 36])]; tensor var_357_cast_fp16 = reshape(shape = var_356, x = linear_13_cast_fp16)[name = tensor("op_357_cast_fp16")]; tensor encoder_layers_1_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1099776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120576)))]; tensor encoder_layers_1_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1121216)))]; tensor linear_14_cast_fp16 = linear(bias = encoder_layers_1_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_1_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_5_cast_fp16)[name = tensor("linear_14_cast_fp16")]; tensor var_364 = const()[name = tensor("op_364"), val = tensor([1, -1, 4, 36])]; tensor var_365_cast_fp16 = reshape(shape = var_364, x = linear_14_cast_fp16)[name = tensor("op_365_cast_fp16")]; tensor var_366 = const()[name = tensor("op_366"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_1_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1121600)))]; tensor var_373_cast_fp16 = add(x = query_7_cast_fp16, y = encoder_layers_1_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_373_cast_fp16")]; tensor content_score_3_transpose_x_0 = const()[name = tensor("content_score_3_transpose_x_0"), val = tensor(false)]; tensor content_score_3_transpose_y_0 = const()[name = tensor("content_score_3_transpose_y_0"), val = tensor(false)]; tensor transpose_26_perm_0 = const()[name = tensor("transpose_26_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_27_perm_0 = const()[name = tensor("transpose_27_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_69 = transpose(perm = transpose_27_perm_0, x = var_357_cast_fp16)[name = tensor("transpose_69")]; tensor transpose_70 = transpose(perm = transpose_26_perm_0, x = var_373_cast_fp16)[name = tensor("transpose_70")]; tensor content_score_3_cast_fp16 = matmul(transpose_x = content_score_3_transpose_x_0, transpose_y = content_score_3_transpose_y_0, x = transpose_70, y = transpose_69)[name = tensor("content_score_3_cast_fp16")]; tensor encoder_layers_1_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1121984)))]; tensor var_377_cast_fp16 = add(x = query_7_cast_fp16, y = encoder_layers_1_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_377_cast_fp16")]; tensor var_378_perm_0 = const()[name = tensor("op_378_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_5_transpose_x_0 = const()[name = tensor("pos_score_5_transpose_x_0"), val = tensor(false)]; tensor pos_score_5_transpose_y_0 = const()[name = tensor("pos_score_5_transpose_y_0"), val = tensor(false)]; tensor op_380_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_380_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1122368))), scale = tensor(0x1.014p-4), zero_point = tensor(-9)]; tensor transpose_68 = transpose(perm = var_378_perm_0, x = var_377_cast_fp16)[name = tensor("transpose_68")]; tensor pos_score_5_cast_fp16 = matmul(transpose_x = pos_score_5_transpose_x_0, transpose_y = pos_score_5_transpose_y_0, x = transpose_68, y = op_380_to_fp16_affine_quantized)[name = tensor("pos_score_5_cast_fp16")]; tensor padded_pos_score_5_interleave_0 = const()[name = tensor("padded_pos_score_5_interleave_0"), val = tensor(false)]; tensor padded_pos_score_5_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_5_interleave_0, values = (zeros_1_to_fp16, pos_score_5_cast_fp16))[name = tensor("padded_pos_score_5_cast_fp16")]; tensor var_394 = const()[name = tensor("op_394"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_7_cast_fp16 = reshape(shape = var_394, x = padded_pos_score_5_cast_fp16)[name = tensor("padded_pos_score_7_cast_fp16")]; tensor var_398_begin_0 = const()[name = tensor("op_398_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_398_end_0 = const()[name = tensor("op_398_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_398_end_mask_0 = const()[name = tensor("op_398_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_398_cast_fp16 = slice_by_index(begin = var_398_begin_0, end = var_398_end_0, end_mask = var_398_end_mask_0, x = padded_pos_score_7_cast_fp16)[name = tensor("op_398_cast_fp16")]; tensor var_405_cast_fp16 = add(x = content_score_3_cast_fp16, y = var_398_cast_fp16)[name = tensor("op_405_cast_fp16")]; tensor _inversed_input_69_y_0_to_fp16 = const()[name = tensor("_inversed_input_69_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_69_cast_fp16 = mul(x = var_405_cast_fp16, y = _inversed_input_69_y_0_to_fp16)[name = tensor("_inversed_input_69_cast_fp16")]; tensor input_71_cast_fp16 = softmax(axis = var_5, x = _inversed_input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; tensor var_410_transpose_x_0 = const()[name = tensor("op_410_transpose_x_0"), val = tensor(false)]; tensor var_410_transpose_y_0 = const()[name = tensor("op_410_transpose_y_0"), val = tensor(false)]; tensor transpose_67 = transpose(perm = var_366, x = var_365_cast_fp16)[name = tensor("transpose_67")]; tensor var_410_cast_fp16 = matmul(transpose_x = var_410_transpose_x_0, transpose_y = var_410_transpose_y_0, x = input_71_cast_fp16, y = transpose_67)[name = tensor("op_410_cast_fp16")]; tensor context_3_perm_0 = const()[name = tensor("context_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_413 = const()[name = tensor("op_413"), val = tensor([1, -1, 144])]; tensor transpose_66 = transpose(perm = context_3_perm_0, x = var_410_cast_fp16)[name = tensor("transpose_66")]; tensor input_73_cast_fp16 = reshape(shape = var_413, x = transpose_66)[name = tensor("input_73_cast_fp16")]; tensor encoder_layers_1_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1129344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1150400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1150144)))]; tensor encoder_layers_1_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1150784)))]; tensor linear_16_cast_fp16 = linear(bias = encoder_layers_1_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_1_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_73_cast_fp16)[name = tensor("linear_16_cast_fp16")]; tensor input_77_cast_fp16 = add(x = linear_16_cast_fp16, y = inputs_13_cast_fp16)[name = tensor("input_77_cast_fp16")]; tensor x_3_axes_0 = const()[name = tensor("x_3_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1151168)))]; tensor encoder_layers_1_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1151552)))]; tensor x_3_cast_fp16 = layer_norm(axes = x_3_axes_0, beta = encoder_layers_1_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_1_sequential_2_module_sequential_0_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("x_3_cast_fp16")]; tensor input_79_perm_0 = const()[name = tensor("input_79_perm_0"), val = tensor([0, 2, 1])]; tensor var_440 = const()[name = tensor("op_440"), val = tensor([1])]; tensor var_442 = const()[name = tensor("op_442"), val = tensor([1])]; tensor inputs_15_pad_type_0 = const()[name = tensor("inputs_15_pad_type_0"), val = tensor("custom")]; tensor inputs_15_pad_0 = const()[name = tensor("inputs_15_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_1_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1151936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1193856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1193472)))]; tensor encoder_layers_1_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1194496)))]; tensor transpose_65 = transpose(perm = input_79_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_65")]; tensor inputs_15_cast_fp16 = conv(bias = encoder_layers_1_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_442, groups = var_15, pad = inputs_15_pad_0, pad_type = inputs_15_pad_type_0, strides = var_440, weight = encoder_layers_1_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_65)[name = tensor("inputs_15_cast_fp16")]; tensor var_445_split_sizes_0 = const()[name = tensor("op_445_split_sizes_0"), val = tensor([144, 144])]; tensor var_445_axis_0 = const()[name = tensor("op_445_axis_0"), val = tensor(1)]; tensor var_445_cast_fp16_0, tensor var_445_cast_fp16_1 = split(axis = var_445_axis_0, split_sizes = var_445_split_sizes_0, x = inputs_15_cast_fp16)[name = tensor("op_445_cast_fp16")]; tensor var_447_cast_fp16 = sigmoid(x = var_445_cast_fp16_1)[name = tensor("op_447_cast_fp16")]; tensor input_81_cast_fp16 = mul(x = var_445_cast_fp16_0, y = var_447_cast_fp16)[name = tensor("input_81_cast_fp16")]; tensor var_451 = const()[name = tensor("op_451"), val = tensor([1])]; tensor var_453 = const()[name = tensor("op_453"), val = tensor([1])]; tensor input_83_pad_type_0 = const()[name = tensor("input_83_pad_type_0"), val = tensor("custom")]; tensor input_83_pad_0 = const()[name = tensor("input_83_pad_0"), val = tensor([1, 1])]; tensor inputs_17_weight_0_to_fp16 = const()[name = tensor("inputs_17_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1195136)))]; tensor inputs_17_bias_0_to_fp16 = const()[name = tensor("inputs_17_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1196096)))]; tensor inputs_17_cast_fp16 = conv(bias = inputs_17_bias_0_to_fp16, dilations = var_453, groups = var_13, pad = input_83_pad_0, pad_type = input_83_pad_type_0, strides = var_451, weight = inputs_17_weight_0_to_fp16, x = input_81_cast_fp16)[name = tensor("inputs_17_cast_fp16")]; tensor input_85_cast_fp16 = silu(x = inputs_17_cast_fp16)[name = tensor("input_85_cast_fp16")]; tensor var_466 = const()[name = tensor("op_466"), val = tensor([1])]; tensor var_468 = const()[name = tensor("op_468"), val = tensor([1])]; tensor input_87_pad_type_0 = const()[name = tensor("input_87_pad_type_0"), val = tensor("custom")]; tensor input_87_pad_0 = const()[name = tensor("input_87_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_1_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1196480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1217536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1217280)))]; tensor encoder_layers_1_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1217920)))]; tensor input_87_cast_fp16 = conv(bias = encoder_layers_1_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_468, groups = var_15, pad = input_87_pad_0, pad_type = input_87_pad_type_0, strides = var_466, weight = encoder_layers_1_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; tensor var_472_perm_0 = const()[name = tensor("op_472_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_64 = transpose(perm = var_472_perm_0, x = input_87_cast_fp16)[name = tensor("transpose_64")]; tensor input_89_cast_fp16 = add(x = transpose_64, y = input_77_cast_fp16)[name = tensor("input_89_cast_fp16")]; tensor input_91_axes_0 = const()[name = tensor("input_91_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1218304)))]; tensor encoder_layers_1_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1218688)))]; tensor input_91_cast_fp16 = layer_norm(axes = input_91_axes_0, beta = encoder_layers_1_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_1_sequential_3_module_sequential_0_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("input_91_cast_fp16")]; tensor encoder_layers_1_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1219072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1302720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1302080)))]; tensor encoder_layers_1_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1303936)))]; tensor linear_17_cast_fp16 = linear(bias = encoder_layers_1_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_1_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_91_cast_fp16)[name = tensor("linear_17_cast_fp16")]; tensor input_93_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_93_cast_fp16")]; tensor encoder_layers_1_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_1_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1305152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1388416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1388160)))]; tensor encoder_layers_1_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1388800)))]; tensor linear_18_cast_fp16 = linear(bias = encoder_layers_1_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_1_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_93_cast_fp16)[name = tensor("linear_18_cast_fp16")]; tensor var_499_to_fp16 = const()[name = tensor("op_499_to_fp16"), val = tensor(0x1p-1)]; tensor var_500_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_499_to_fp16)[name = tensor("op_500_cast_fp16")]; tensor input_99_cast_fp16 = add(x = var_500_cast_fp16, y = input_89_cast_fp16)[name = tensor("input_99_cast_fp16")]; tensor input_101_axes_0 = const()[name = tensor("input_101_axes_0"), val = tensor([-1])]; tensor encoder_layers_1_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1389184)))]; tensor encoder_layers_1_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_1_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1389568)))]; tensor input_101_cast_fp16 = layer_norm(axes = input_101_axes_0, beta = encoder_layers_1_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_1_sequential_4_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("input_101_cast_fp16")]; tensor input_103_axes_0 = const()[name = tensor("input_103_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1389952)))]; tensor encoder_layers_2_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1390336)))]; tensor input_103_cast_fp16 = layer_norm(axes = input_103_axes_0, beta = encoder_layers_2_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_2_sequential_0_module_sequential_0_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; tensor encoder_layers_2_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1390720))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1474368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1473728)))]; tensor encoder_layers_2_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1475584)))]; tensor linear_19_cast_fp16 = linear(bias = encoder_layers_2_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_2_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_103_cast_fp16)[name = tensor("linear_19_cast_fp16")]; tensor input_105_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_105_cast_fp16")]; tensor encoder_layers_2_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1476800))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1559808)))]; tensor encoder_layers_2_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560448)))]; tensor linear_20_cast_fp16 = linear(bias = encoder_layers_2_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_2_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_105_cast_fp16)[name = tensor("linear_20_cast_fp16")]; tensor var_535_to_fp16 = const()[name = tensor("op_535_to_fp16"), val = tensor(0x1p-1)]; tensor var_536_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_535_to_fp16)[name = tensor("op_536_cast_fp16")]; tensor inputs_23_cast_fp16 = add(x = var_536_cast_fp16, y = input_101_cast_fp16)[name = tensor("inputs_23_cast_fp16")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1560832)))]; tensor encoder_layers_2_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1561216)))]; tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = encoder_layers_2_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_2_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_23_cast_fp16)[name = tensor("query_9_cast_fp16")]; tensor encoder_layers_2_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1561600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1582656))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1582400)))]; tensor encoder_layers_2_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1583040)))]; tensor linear_21_cast_fp16 = linear(bias = encoder_layers_2_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_2_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_9_cast_fp16)[name = tensor("linear_21_cast_fp16")]; tensor var_567 = const()[name = tensor("op_567"), val = tensor([1, -1, 4, 36])]; tensor query_11_cast_fp16 = reshape(shape = var_567, x = linear_21_cast_fp16)[name = tensor("query_11_cast_fp16")]; tensor encoder_layers_2_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1583424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1604480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1604224)))]; tensor encoder_layers_2_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1604864)))]; tensor linear_22_cast_fp16 = linear(bias = encoder_layers_2_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_2_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_9_cast_fp16)[name = tensor("linear_22_cast_fp16")]; tensor var_573 = const()[name = tensor("op_573"), val = tensor([1, -1, 4, 36])]; tensor var_574_cast_fp16 = reshape(shape = var_573, x = linear_22_cast_fp16)[name = tensor("op_574_cast_fp16")]; tensor encoder_layers_2_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1605248))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1626304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1626048)))]; tensor encoder_layers_2_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1626688)))]; tensor linear_23_cast_fp16 = linear(bias = encoder_layers_2_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_2_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_9_cast_fp16)[name = tensor("linear_23_cast_fp16")]; tensor var_581 = const()[name = tensor("op_581"), val = tensor([1, -1, 4, 36])]; tensor var_582_cast_fp16 = reshape(shape = var_581, x = linear_23_cast_fp16)[name = tensor("op_582_cast_fp16")]; tensor var_583 = const()[name = tensor("op_583"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_2_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1627072)))]; tensor var_590_cast_fp16 = add(x = query_11_cast_fp16, y = encoder_layers_2_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_590_cast_fp16")]; tensor content_score_5_transpose_x_0 = const()[name = tensor("content_score_5_transpose_x_0"), val = tensor(false)]; tensor content_score_5_transpose_y_0 = const()[name = tensor("content_score_5_transpose_y_0"), val = tensor(false)]; tensor transpose_28_perm_0 = const()[name = tensor("transpose_28_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_29_perm_0 = const()[name = tensor("transpose_29_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_62 = transpose(perm = transpose_29_perm_0, x = var_574_cast_fp16)[name = tensor("transpose_62")]; tensor transpose_63 = transpose(perm = transpose_28_perm_0, x = var_590_cast_fp16)[name = tensor("transpose_63")]; tensor content_score_5_cast_fp16 = matmul(transpose_x = content_score_5_transpose_x_0, transpose_y = content_score_5_transpose_y_0, x = transpose_63, y = transpose_62)[name = tensor("content_score_5_cast_fp16")]; tensor encoder_layers_2_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1627456)))]; tensor var_594_cast_fp16 = add(x = query_11_cast_fp16, y = encoder_layers_2_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_594_cast_fp16")]; tensor var_595_perm_0 = const()[name = tensor("op_595_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_9_transpose_x_0 = const()[name = tensor("pos_score_9_transpose_x_0"), val = tensor(false)]; tensor pos_score_9_transpose_y_0 = const()[name = tensor("pos_score_9_transpose_y_0"), val = tensor(false)]; tensor op_597_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_597_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1627840))), scale = tensor(0x1.b34p-5), zero_point = tensor(-15)]; tensor transpose_61 = transpose(perm = var_595_perm_0, x = var_594_cast_fp16)[name = tensor("transpose_61")]; tensor pos_score_9_cast_fp16 = matmul(transpose_x = pos_score_9_transpose_x_0, transpose_y = pos_score_9_transpose_y_0, x = transpose_61, y = op_597_to_fp16_affine_quantized)[name = tensor("pos_score_9_cast_fp16")]; tensor padded_pos_score_9_interleave_0 = const()[name = tensor("padded_pos_score_9_interleave_0"), val = tensor(false)]; tensor padded_pos_score_9_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_9_interleave_0, values = (zeros_1_to_fp16, pos_score_9_cast_fp16))[name = tensor("padded_pos_score_9_cast_fp16")]; tensor var_611 = const()[name = tensor("op_611"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_11_cast_fp16 = reshape(shape = var_611, x = padded_pos_score_9_cast_fp16)[name = tensor("padded_pos_score_11_cast_fp16")]; tensor var_615_begin_0 = const()[name = tensor("op_615_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_615_end_0 = const()[name = tensor("op_615_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_615_end_mask_0 = const()[name = tensor("op_615_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_615_cast_fp16 = slice_by_index(begin = var_615_begin_0, end = var_615_end_0, end_mask = var_615_end_mask_0, x = padded_pos_score_11_cast_fp16)[name = tensor("op_615_cast_fp16")]; tensor var_622_cast_fp16 = add(x = content_score_5_cast_fp16, y = var_615_cast_fp16)[name = tensor("op_622_cast_fp16")]; tensor _inversed_input_113_y_0_to_fp16 = const()[name = tensor("_inversed_input_113_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_113_cast_fp16 = mul(x = var_622_cast_fp16, y = _inversed_input_113_y_0_to_fp16)[name = tensor("_inversed_input_113_cast_fp16")]; tensor input_115_cast_fp16 = softmax(axis = var_5, x = _inversed_input_113_cast_fp16)[name = tensor("input_115_cast_fp16")]; tensor var_627_transpose_x_0 = const()[name = tensor("op_627_transpose_x_0"), val = tensor(false)]; tensor var_627_transpose_y_0 = const()[name = tensor("op_627_transpose_y_0"), val = tensor(false)]; tensor transpose_60 = transpose(perm = var_583, x = var_582_cast_fp16)[name = tensor("transpose_60")]; tensor var_627_cast_fp16 = matmul(transpose_x = var_627_transpose_x_0, transpose_y = var_627_transpose_y_0, x = input_115_cast_fp16, y = transpose_60)[name = tensor("op_627_cast_fp16")]; tensor context_5_perm_0 = const()[name = tensor("context_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_630 = const()[name = tensor("op_630"), val = tensor([1, -1, 144])]; tensor transpose_59 = transpose(perm = context_5_perm_0, x = var_627_cast_fp16)[name = tensor("transpose_59")]; tensor input_117_cast_fp16 = reshape(shape = var_630, x = transpose_59)[name = tensor("input_117_cast_fp16")]; tensor encoder_layers_2_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1634816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1655872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1655616)))]; tensor encoder_layers_2_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1656256)))]; tensor linear_25_cast_fp16 = linear(bias = encoder_layers_2_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_2_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_117_cast_fp16)[name = tensor("linear_25_cast_fp16")]; tensor input_121_cast_fp16 = add(x = linear_25_cast_fp16, y = inputs_23_cast_fp16)[name = tensor("input_121_cast_fp16")]; tensor x_5_axes_0 = const()[name = tensor("x_5_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1656640)))]; tensor encoder_layers_2_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657024)))]; tensor x_5_cast_fp16 = layer_norm(axes = x_5_axes_0, beta = encoder_layers_2_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_2_sequential_2_module_sequential_0_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("x_5_cast_fp16")]; tensor input_123_perm_0 = const()[name = tensor("input_123_perm_0"), val = tensor([0, 2, 1])]; tensor var_657 = const()[name = tensor("op_657"), val = tensor([1])]; tensor var_659 = const()[name = tensor("op_659"), val = tensor([1])]; tensor inputs_25_pad_type_0 = const()[name = tensor("inputs_25_pad_type_0"), val = tensor("custom")]; tensor inputs_25_pad_0 = const()[name = tensor("inputs_25_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_2_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1657408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1699328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1698944)))]; tensor encoder_layers_2_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1699968)))]; tensor transpose_58 = transpose(perm = input_123_perm_0, x = x_5_cast_fp16)[name = tensor("transpose_58")]; tensor inputs_25_cast_fp16 = conv(bias = encoder_layers_2_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_659, groups = var_15, pad = inputs_25_pad_0, pad_type = inputs_25_pad_type_0, strides = var_657, weight = encoder_layers_2_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_58)[name = tensor("inputs_25_cast_fp16")]; tensor var_662_split_sizes_0 = const()[name = tensor("op_662_split_sizes_0"), val = tensor([144, 144])]; tensor var_662_axis_0 = const()[name = tensor("op_662_axis_0"), val = tensor(1)]; tensor var_662_cast_fp16_0, tensor var_662_cast_fp16_1 = split(axis = var_662_axis_0, split_sizes = var_662_split_sizes_0, x = inputs_25_cast_fp16)[name = tensor("op_662_cast_fp16")]; tensor var_664_cast_fp16 = sigmoid(x = var_662_cast_fp16_1)[name = tensor("op_664_cast_fp16")]; tensor input_125_cast_fp16 = mul(x = var_662_cast_fp16_0, y = var_664_cast_fp16)[name = tensor("input_125_cast_fp16")]; tensor var_668 = const()[name = tensor("op_668"), val = tensor([1])]; tensor var_670 = const()[name = tensor("op_670"), val = tensor([1])]; tensor input_127_pad_type_0 = const()[name = tensor("input_127_pad_type_0"), val = tensor("custom")]; tensor input_127_pad_0 = const()[name = tensor("input_127_pad_0"), val = tensor([1, 1])]; tensor inputs_27_weight_0_to_fp16 = const()[name = tensor("inputs_27_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1700608)))]; tensor inputs_27_bias_0_to_fp16 = const()[name = tensor("inputs_27_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1701568)))]; tensor inputs_27_cast_fp16 = conv(bias = inputs_27_bias_0_to_fp16, dilations = var_670, groups = var_13, pad = input_127_pad_0, pad_type = input_127_pad_type_0, strides = var_668, weight = inputs_27_weight_0_to_fp16, x = input_125_cast_fp16)[name = tensor("inputs_27_cast_fp16")]; tensor input_129_cast_fp16 = silu(x = inputs_27_cast_fp16)[name = tensor("input_129_cast_fp16")]; tensor var_683 = const()[name = tensor("op_683"), val = tensor([1])]; tensor var_685 = const()[name = tensor("op_685"), val = tensor([1])]; tensor input_131_pad_type_0 = const()[name = tensor("input_131_pad_type_0"), val = tensor("custom")]; tensor input_131_pad_0 = const()[name = tensor("input_131_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_2_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1701952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1723008))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1722752)))]; tensor encoder_layers_2_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1723392)))]; tensor input_131_cast_fp16 = conv(bias = encoder_layers_2_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_685, groups = var_15, pad = input_131_pad_0, pad_type = input_131_pad_type_0, strides = var_683, weight = encoder_layers_2_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; tensor var_689_perm_0 = const()[name = tensor("op_689_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_57 = transpose(perm = var_689_perm_0, x = input_131_cast_fp16)[name = tensor("transpose_57")]; tensor input_133_cast_fp16 = add(x = transpose_57, y = input_121_cast_fp16)[name = tensor("input_133_cast_fp16")]; tensor input_135_axes_0 = const()[name = tensor("input_135_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1723776)))]; tensor encoder_layers_2_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1724160)))]; tensor input_135_cast_fp16 = layer_norm(axes = input_135_axes_0, beta = encoder_layers_2_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_2_sequential_3_module_sequential_0_weight_to_fp16, x = input_133_cast_fp16)[name = tensor("input_135_cast_fp16")]; tensor encoder_layers_2_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1724544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1808192))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1807552)))]; tensor encoder_layers_2_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1809408)))]; tensor linear_26_cast_fp16 = linear(bias = encoder_layers_2_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_2_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_135_cast_fp16)[name = tensor("linear_26_cast_fp16")]; tensor input_137_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_137_cast_fp16")]; tensor encoder_layers_2_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_2_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1810624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1893888))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1893632)))]; tensor encoder_layers_2_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1894272)))]; tensor linear_27_cast_fp16 = linear(bias = encoder_layers_2_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_2_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_137_cast_fp16)[name = tensor("linear_27_cast_fp16")]; tensor var_716_to_fp16 = const()[name = tensor("op_716_to_fp16"), val = tensor(0x1p-1)]; tensor var_717_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_716_to_fp16)[name = tensor("op_717_cast_fp16")]; tensor input_143_cast_fp16 = add(x = var_717_cast_fp16, y = input_133_cast_fp16)[name = tensor("input_143_cast_fp16")]; tensor input_145_axes_0 = const()[name = tensor("input_145_axes_0"), val = tensor([-1])]; tensor encoder_layers_2_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1894656)))]; tensor encoder_layers_2_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_2_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1895040)))]; tensor input_145_cast_fp16 = layer_norm(axes = input_145_axes_0, beta = encoder_layers_2_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_2_sequential_4_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("input_145_cast_fp16")]; tensor input_147_axes_0 = const()[name = tensor("input_147_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1895424)))]; tensor encoder_layers_3_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1895808)))]; tensor input_147_cast_fp16 = layer_norm(axes = input_147_axes_0, beta = encoder_layers_3_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_3_sequential_0_module_sequential_0_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; tensor encoder_layers_3_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1896192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1979840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1979200)))]; tensor encoder_layers_3_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1981056)))]; tensor linear_28_cast_fp16 = linear(bias = encoder_layers_3_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_3_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_147_cast_fp16)[name = tensor("linear_28_cast_fp16")]; tensor input_149_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_149_cast_fp16")]; tensor encoder_layers_3_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1982272))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2065536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2065280)))]; tensor encoder_layers_3_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2065920)))]; tensor linear_29_cast_fp16 = linear(bias = encoder_layers_3_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_3_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_149_cast_fp16)[name = tensor("linear_29_cast_fp16")]; tensor var_752_to_fp16 = const()[name = tensor("op_752_to_fp16"), val = tensor(0x1p-1)]; tensor var_753_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_752_to_fp16)[name = tensor("op_753_cast_fp16")]; tensor inputs_33_cast_fp16 = add(x = var_753_cast_fp16, y = input_145_cast_fp16)[name = tensor("inputs_33_cast_fp16")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2066304)))]; tensor encoder_layers_3_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2066688)))]; tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = encoder_layers_3_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_3_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_33_cast_fp16)[name = tensor("query_13_cast_fp16")]; tensor encoder_layers_3_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2067072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2088128))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2087872)))]; tensor encoder_layers_3_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2088512)))]; tensor linear_30_cast_fp16 = linear(bias = encoder_layers_3_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_3_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_13_cast_fp16)[name = tensor("linear_30_cast_fp16")]; tensor var_784 = const()[name = tensor("op_784"), val = tensor([1, -1, 4, 36])]; tensor query_15_cast_fp16 = reshape(shape = var_784, x = linear_30_cast_fp16)[name = tensor("query_15_cast_fp16")]; tensor encoder_layers_3_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2088896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2109952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2109696)))]; tensor encoder_layers_3_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2110336)))]; tensor linear_31_cast_fp16 = linear(bias = encoder_layers_3_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_3_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_13_cast_fp16)[name = tensor("linear_31_cast_fp16")]; tensor var_790 = const()[name = tensor("op_790"), val = tensor([1, -1, 4, 36])]; tensor var_791_cast_fp16 = reshape(shape = var_790, x = linear_31_cast_fp16)[name = tensor("op_791_cast_fp16")]; tensor encoder_layers_3_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2110720))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2131776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2131520)))]; tensor encoder_layers_3_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2132160)))]; tensor linear_32_cast_fp16 = linear(bias = encoder_layers_3_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_3_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_13_cast_fp16)[name = tensor("linear_32_cast_fp16")]; tensor var_798 = const()[name = tensor("op_798"), val = tensor([1, -1, 4, 36])]; tensor var_799_cast_fp16 = reshape(shape = var_798, x = linear_32_cast_fp16)[name = tensor("op_799_cast_fp16")]; tensor var_800 = const()[name = tensor("op_800"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_3_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2132544)))]; tensor var_807_cast_fp16 = add(x = query_15_cast_fp16, y = encoder_layers_3_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_807_cast_fp16")]; tensor content_score_7_transpose_x_0 = const()[name = tensor("content_score_7_transpose_x_0"), val = tensor(false)]; tensor content_score_7_transpose_y_0 = const()[name = tensor("content_score_7_transpose_y_0"), val = tensor(false)]; tensor transpose_30_perm_0 = const()[name = tensor("transpose_30_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_31_perm_0 = const()[name = tensor("transpose_31_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_55 = transpose(perm = transpose_31_perm_0, x = var_791_cast_fp16)[name = tensor("transpose_55")]; tensor transpose_56 = transpose(perm = transpose_30_perm_0, x = var_807_cast_fp16)[name = tensor("transpose_56")]; tensor content_score_7_cast_fp16 = matmul(transpose_x = content_score_7_transpose_x_0, transpose_y = content_score_7_transpose_y_0, x = transpose_56, y = transpose_55)[name = tensor("content_score_7_cast_fp16")]; tensor encoder_layers_3_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2132928)))]; tensor var_811_cast_fp16 = add(x = query_15_cast_fp16, y = encoder_layers_3_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_811_cast_fp16")]; tensor var_812_perm_0 = const()[name = tensor("op_812_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_13_transpose_x_0 = const()[name = tensor("pos_score_13_transpose_x_0"), val = tensor(false)]; tensor pos_score_13_transpose_y_0 = const()[name = tensor("pos_score_13_transpose_y_0"), val = tensor(false)]; tensor op_814_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_814_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2133312))), scale = tensor(0x1.d1cp-5), zero_point = tensor(-22)]; tensor transpose_54 = transpose(perm = var_812_perm_0, x = var_811_cast_fp16)[name = tensor("transpose_54")]; tensor pos_score_13_cast_fp16 = matmul(transpose_x = pos_score_13_transpose_x_0, transpose_y = pos_score_13_transpose_y_0, x = transpose_54, y = op_814_to_fp16_affine_quantized)[name = tensor("pos_score_13_cast_fp16")]; tensor padded_pos_score_13_interleave_0 = const()[name = tensor("padded_pos_score_13_interleave_0"), val = tensor(false)]; tensor padded_pos_score_13_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_13_interleave_0, values = (zeros_1_to_fp16, pos_score_13_cast_fp16))[name = tensor("padded_pos_score_13_cast_fp16")]; tensor var_828 = const()[name = tensor("op_828"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_15_cast_fp16 = reshape(shape = var_828, x = padded_pos_score_13_cast_fp16)[name = tensor("padded_pos_score_15_cast_fp16")]; tensor var_832_begin_0 = const()[name = tensor("op_832_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_832_end_0 = const()[name = tensor("op_832_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_832_end_mask_0 = const()[name = tensor("op_832_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_832_cast_fp16 = slice_by_index(begin = var_832_begin_0, end = var_832_end_0, end_mask = var_832_end_mask_0, x = padded_pos_score_15_cast_fp16)[name = tensor("op_832_cast_fp16")]; tensor var_839_cast_fp16 = add(x = content_score_7_cast_fp16, y = var_832_cast_fp16)[name = tensor("op_839_cast_fp16")]; tensor _inversed_input_157_y_0_to_fp16 = const()[name = tensor("_inversed_input_157_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_157_cast_fp16 = mul(x = var_839_cast_fp16, y = _inversed_input_157_y_0_to_fp16)[name = tensor("_inversed_input_157_cast_fp16")]; tensor input_159_cast_fp16 = softmax(axis = var_5, x = _inversed_input_157_cast_fp16)[name = tensor("input_159_cast_fp16")]; tensor var_844_transpose_x_0 = const()[name = tensor("op_844_transpose_x_0"), val = tensor(false)]; tensor var_844_transpose_y_0 = const()[name = tensor("op_844_transpose_y_0"), val = tensor(false)]; tensor transpose_53 = transpose(perm = var_800, x = var_799_cast_fp16)[name = tensor("transpose_53")]; tensor var_844_cast_fp16 = matmul(transpose_x = var_844_transpose_x_0, transpose_y = var_844_transpose_y_0, x = input_159_cast_fp16, y = transpose_53)[name = tensor("op_844_cast_fp16")]; tensor context_7_perm_0 = const()[name = tensor("context_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_847 = const()[name = tensor("op_847"), val = tensor([1, -1, 144])]; tensor transpose_52 = transpose(perm = context_7_perm_0, x = var_844_cast_fp16)[name = tensor("transpose_52")]; tensor input_161_cast_fp16 = reshape(shape = var_847, x = transpose_52)[name = tensor("input_161_cast_fp16")]; tensor encoder_layers_3_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2140288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2161344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2161088)))]; tensor encoder_layers_3_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2161728)))]; tensor linear_34_cast_fp16 = linear(bias = encoder_layers_3_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_3_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_161_cast_fp16)[name = tensor("linear_34_cast_fp16")]; tensor input_165_cast_fp16 = add(x = linear_34_cast_fp16, y = inputs_33_cast_fp16)[name = tensor("input_165_cast_fp16")]; tensor x_7_axes_0 = const()[name = tensor("x_7_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2162112)))]; tensor encoder_layers_3_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2162496)))]; tensor x_7_cast_fp16 = layer_norm(axes = x_7_axes_0, beta = encoder_layers_3_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_3_sequential_2_module_sequential_0_weight_to_fp16, x = input_165_cast_fp16)[name = tensor("x_7_cast_fp16")]; tensor input_167_perm_0 = const()[name = tensor("input_167_perm_0"), val = tensor([0, 2, 1])]; tensor var_874 = const()[name = tensor("op_874"), val = tensor([1])]; tensor var_876 = const()[name = tensor("op_876"), val = tensor([1])]; tensor inputs_35_pad_type_0 = const()[name = tensor("inputs_35_pad_type_0"), val = tensor("custom")]; tensor inputs_35_pad_0 = const()[name = tensor("inputs_35_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_3_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2162880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2204800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2204416)))]; tensor encoder_layers_3_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2205440)))]; tensor transpose_51 = transpose(perm = input_167_perm_0, x = x_7_cast_fp16)[name = tensor("transpose_51")]; tensor inputs_35_cast_fp16 = conv(bias = encoder_layers_3_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_876, groups = var_15, pad = inputs_35_pad_0, pad_type = inputs_35_pad_type_0, strides = var_874, weight = encoder_layers_3_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_51)[name = tensor("inputs_35_cast_fp16")]; tensor var_879_split_sizes_0 = const()[name = tensor("op_879_split_sizes_0"), val = tensor([144, 144])]; tensor var_879_axis_0 = const()[name = tensor("op_879_axis_0"), val = tensor(1)]; tensor var_879_cast_fp16_0, tensor var_879_cast_fp16_1 = split(axis = var_879_axis_0, split_sizes = var_879_split_sizes_0, x = inputs_35_cast_fp16)[name = tensor("op_879_cast_fp16")]; tensor var_881_cast_fp16 = sigmoid(x = var_879_cast_fp16_1)[name = tensor("op_881_cast_fp16")]; tensor input_169_cast_fp16 = mul(x = var_879_cast_fp16_0, y = var_881_cast_fp16)[name = tensor("input_169_cast_fp16")]; tensor var_885 = const()[name = tensor("op_885"), val = tensor([1])]; tensor var_887 = const()[name = tensor("op_887"), val = tensor([1])]; tensor input_171_pad_type_0 = const()[name = tensor("input_171_pad_type_0"), val = tensor("custom")]; tensor input_171_pad_0 = const()[name = tensor("input_171_pad_0"), val = tensor([1, 1])]; tensor inputs_37_weight_0_to_fp16 = const()[name = tensor("inputs_37_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2206080)))]; tensor inputs_37_bias_0_to_fp16 = const()[name = tensor("inputs_37_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2207040)))]; tensor inputs_37_cast_fp16 = conv(bias = inputs_37_bias_0_to_fp16, dilations = var_887, groups = var_13, pad = input_171_pad_0, pad_type = input_171_pad_type_0, strides = var_885, weight = inputs_37_weight_0_to_fp16, x = input_169_cast_fp16)[name = tensor("inputs_37_cast_fp16")]; tensor input_173_cast_fp16 = silu(x = inputs_37_cast_fp16)[name = tensor("input_173_cast_fp16")]; tensor var_900 = const()[name = tensor("op_900"), val = tensor([1])]; tensor var_902 = const()[name = tensor("op_902"), val = tensor([1])]; tensor input_175_pad_type_0 = const()[name = tensor("input_175_pad_type_0"), val = tensor("custom")]; tensor input_175_pad_0 = const()[name = tensor("input_175_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_3_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2207424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2228480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2228224)))]; tensor encoder_layers_3_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2228864)))]; tensor input_175_cast_fp16 = conv(bias = encoder_layers_3_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_902, groups = var_15, pad = input_175_pad_0, pad_type = input_175_pad_type_0, strides = var_900, weight = encoder_layers_3_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; tensor var_906_perm_0 = const()[name = tensor("op_906_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_50 = transpose(perm = var_906_perm_0, x = input_175_cast_fp16)[name = tensor("transpose_50")]; tensor input_177_cast_fp16 = add(x = transpose_50, y = input_165_cast_fp16)[name = tensor("input_177_cast_fp16")]; tensor input_179_axes_0 = const()[name = tensor("input_179_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2229248)))]; tensor encoder_layers_3_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2229632)))]; tensor input_179_cast_fp16 = layer_norm(axes = input_179_axes_0, beta = encoder_layers_3_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_3_sequential_3_module_sequential_0_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("input_179_cast_fp16")]; tensor encoder_layers_3_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2230016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2313664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2313024)))]; tensor encoder_layers_3_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2314880)))]; tensor linear_35_cast_fp16 = linear(bias = encoder_layers_3_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_3_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_179_cast_fp16)[name = tensor("linear_35_cast_fp16")]; tensor input_181_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_181_cast_fp16")]; tensor encoder_layers_3_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_3_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2316096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2399360))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2399104)))]; tensor encoder_layers_3_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2399744)))]; tensor linear_36_cast_fp16 = linear(bias = encoder_layers_3_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_3_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_181_cast_fp16)[name = tensor("linear_36_cast_fp16")]; tensor var_933_to_fp16 = const()[name = tensor("op_933_to_fp16"), val = tensor(0x1p-1)]; tensor var_934_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_933_to_fp16)[name = tensor("op_934_cast_fp16")]; tensor input_187_cast_fp16 = add(x = var_934_cast_fp16, y = input_177_cast_fp16)[name = tensor("input_187_cast_fp16")]; tensor input_189_axes_0 = const()[name = tensor("input_189_axes_0"), val = tensor([-1])]; tensor encoder_layers_3_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2400128)))]; tensor encoder_layers_3_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_3_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2400512)))]; tensor input_189_cast_fp16 = layer_norm(axes = input_189_axes_0, beta = encoder_layers_3_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_3_sequential_4_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("input_189_cast_fp16")]; tensor input_191_axes_0 = const()[name = tensor("input_191_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2400896)))]; tensor encoder_layers_4_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2401280)))]; tensor input_191_cast_fp16 = layer_norm(axes = input_191_axes_0, beta = encoder_layers_4_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_4_sequential_0_module_sequential_0_weight_to_fp16, x = input_189_cast_fp16)[name = tensor("input_191_cast_fp16")]; tensor encoder_layers_4_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2401664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2485312))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2484672)))]; tensor encoder_layers_4_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2486528)))]; tensor linear_37_cast_fp16 = linear(bias = encoder_layers_4_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_4_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_191_cast_fp16)[name = tensor("linear_37_cast_fp16")]; tensor input_193_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_193_cast_fp16")]; tensor encoder_layers_4_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2487744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2571008))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2570752)))]; tensor encoder_layers_4_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2571392)))]; tensor linear_38_cast_fp16 = linear(bias = encoder_layers_4_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_4_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_193_cast_fp16)[name = tensor("linear_38_cast_fp16")]; tensor var_969_to_fp16 = const()[name = tensor("op_969_to_fp16"), val = tensor(0x1p-1)]; tensor var_970_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_969_to_fp16)[name = tensor("op_970_cast_fp16")]; tensor inputs_43_cast_fp16 = add(x = var_970_cast_fp16, y = input_189_cast_fp16)[name = tensor("inputs_43_cast_fp16")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2571776)))]; tensor encoder_layers_4_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2572160)))]; tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = encoder_layers_4_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_4_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_43_cast_fp16)[name = tensor("query_17_cast_fp16")]; tensor encoder_layers_4_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2572544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2593600))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2593344)))]; tensor encoder_layers_4_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2593984)))]; tensor linear_39_cast_fp16 = linear(bias = encoder_layers_4_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_4_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_17_cast_fp16)[name = tensor("linear_39_cast_fp16")]; tensor var_1001 = const()[name = tensor("op_1001"), val = tensor([1, -1, 4, 36])]; tensor query_19_cast_fp16 = reshape(shape = var_1001, x = linear_39_cast_fp16)[name = tensor("query_19_cast_fp16")]; tensor encoder_layers_4_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2594368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2615424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2615168)))]; tensor encoder_layers_4_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2615808)))]; tensor linear_40_cast_fp16 = linear(bias = encoder_layers_4_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_4_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_17_cast_fp16)[name = tensor("linear_40_cast_fp16")]; tensor var_1007 = const()[name = tensor("op_1007"), val = tensor([1, -1, 4, 36])]; tensor var_1008_cast_fp16 = reshape(shape = var_1007, x = linear_40_cast_fp16)[name = tensor("op_1008_cast_fp16")]; tensor encoder_layers_4_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2616192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2637248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2636992)))]; tensor encoder_layers_4_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2637632)))]; tensor linear_41_cast_fp16 = linear(bias = encoder_layers_4_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_4_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_17_cast_fp16)[name = tensor("linear_41_cast_fp16")]; tensor var_1015 = const()[name = tensor("op_1015"), val = tensor([1, -1, 4, 36])]; tensor var_1016_cast_fp16 = reshape(shape = var_1015, x = linear_41_cast_fp16)[name = tensor("op_1016_cast_fp16")]; tensor var_1017 = const()[name = tensor("op_1017"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_4_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2638016)))]; tensor var_1024_cast_fp16 = add(x = query_19_cast_fp16, y = encoder_layers_4_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_1024_cast_fp16")]; tensor content_score_9_transpose_x_0 = const()[name = tensor("content_score_9_transpose_x_0"), val = tensor(false)]; tensor content_score_9_transpose_y_0 = const()[name = tensor("content_score_9_transpose_y_0"), val = tensor(false)]; tensor transpose_32_perm_0 = const()[name = tensor("transpose_32_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_33_perm_0 = const()[name = tensor("transpose_33_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_48 = transpose(perm = transpose_33_perm_0, x = var_1008_cast_fp16)[name = tensor("transpose_48")]; tensor transpose_49 = transpose(perm = transpose_32_perm_0, x = var_1024_cast_fp16)[name = tensor("transpose_49")]; tensor content_score_9_cast_fp16 = matmul(transpose_x = content_score_9_transpose_x_0, transpose_y = content_score_9_transpose_y_0, x = transpose_49, y = transpose_48)[name = tensor("content_score_9_cast_fp16")]; tensor encoder_layers_4_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2638400)))]; tensor var_1028_cast_fp16 = add(x = query_19_cast_fp16, y = encoder_layers_4_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_1028_cast_fp16")]; tensor var_1029_perm_0 = const()[name = tensor("op_1029_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_17_transpose_x_0 = const()[name = tensor("pos_score_17_transpose_x_0"), val = tensor(false)]; tensor pos_score_17_transpose_y_0 = const()[name = tensor("pos_score_17_transpose_y_0"), val = tensor(false)]; tensor op_1031_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_1031_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2638784))), scale = tensor(0x1.66p-5), zero_point = tensor(-13)]; tensor transpose_47 = transpose(perm = var_1029_perm_0, x = var_1028_cast_fp16)[name = tensor("transpose_47")]; tensor pos_score_17_cast_fp16 = matmul(transpose_x = pos_score_17_transpose_x_0, transpose_y = pos_score_17_transpose_y_0, x = transpose_47, y = op_1031_to_fp16_affine_quantized)[name = tensor("pos_score_17_cast_fp16")]; tensor padded_pos_score_17_interleave_0 = const()[name = tensor("padded_pos_score_17_interleave_0"), val = tensor(false)]; tensor padded_pos_score_17_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_17_interleave_0, values = (zeros_1_to_fp16, pos_score_17_cast_fp16))[name = tensor("padded_pos_score_17_cast_fp16")]; tensor var_1045 = const()[name = tensor("op_1045"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_19_cast_fp16 = reshape(shape = var_1045, x = padded_pos_score_17_cast_fp16)[name = tensor("padded_pos_score_19_cast_fp16")]; tensor var_1049_begin_0 = const()[name = tensor("op_1049_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1049_end_0 = const()[name = tensor("op_1049_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_1049_end_mask_0 = const()[name = tensor("op_1049_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1049_cast_fp16 = slice_by_index(begin = var_1049_begin_0, end = var_1049_end_0, end_mask = var_1049_end_mask_0, x = padded_pos_score_19_cast_fp16)[name = tensor("op_1049_cast_fp16")]; tensor var_1056_cast_fp16 = add(x = content_score_9_cast_fp16, y = var_1049_cast_fp16)[name = tensor("op_1056_cast_fp16")]; tensor _inversed_input_201_y_0_to_fp16 = const()[name = tensor("_inversed_input_201_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_201_cast_fp16 = mul(x = var_1056_cast_fp16, y = _inversed_input_201_y_0_to_fp16)[name = tensor("_inversed_input_201_cast_fp16")]; tensor input_203_cast_fp16 = softmax(axis = var_5, x = _inversed_input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; tensor var_1061_transpose_x_0 = const()[name = tensor("op_1061_transpose_x_0"), val = tensor(false)]; tensor var_1061_transpose_y_0 = const()[name = tensor("op_1061_transpose_y_0"), val = tensor(false)]; tensor transpose_46 = transpose(perm = var_1017, x = var_1016_cast_fp16)[name = tensor("transpose_46")]; tensor var_1061_cast_fp16 = matmul(transpose_x = var_1061_transpose_x_0, transpose_y = var_1061_transpose_y_0, x = input_203_cast_fp16, y = transpose_46)[name = tensor("op_1061_cast_fp16")]; tensor context_9_perm_0 = const()[name = tensor("context_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, -1, 144])]; tensor transpose_45 = transpose(perm = context_9_perm_0, x = var_1061_cast_fp16)[name = tensor("transpose_45")]; tensor input_205_cast_fp16 = reshape(shape = var_1064, x = transpose_45)[name = tensor("input_205_cast_fp16")]; tensor encoder_layers_4_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2645760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2666816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2666560)))]; tensor encoder_layers_4_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2667200)))]; tensor linear_43_cast_fp16 = linear(bias = encoder_layers_4_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_4_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_205_cast_fp16)[name = tensor("linear_43_cast_fp16")]; tensor input_209_cast_fp16 = add(x = linear_43_cast_fp16, y = inputs_43_cast_fp16)[name = tensor("input_209_cast_fp16")]; tensor x_9_axes_0 = const()[name = tensor("x_9_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2667584)))]; tensor encoder_layers_4_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2667968)))]; tensor x_9_cast_fp16 = layer_norm(axes = x_9_axes_0, beta = encoder_layers_4_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_4_sequential_2_module_sequential_0_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_9_cast_fp16")]; tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; tensor var_1091 = const()[name = tensor("op_1091"), val = tensor([1])]; tensor var_1093 = const()[name = tensor("op_1093"), val = tensor([1])]; tensor inputs_45_pad_type_0 = const()[name = tensor("inputs_45_pad_type_0"), val = tensor("custom")]; tensor inputs_45_pad_0 = const()[name = tensor("inputs_45_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_4_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2668352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2710272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2709888)))]; tensor encoder_layers_4_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2710912)))]; tensor transpose_44 = transpose(perm = input_211_perm_0, x = x_9_cast_fp16)[name = tensor("transpose_44")]; tensor inputs_45_cast_fp16 = conv(bias = encoder_layers_4_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_1093, groups = var_15, pad = inputs_45_pad_0, pad_type = inputs_45_pad_type_0, strides = var_1091, weight = encoder_layers_4_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_44)[name = tensor("inputs_45_cast_fp16")]; tensor var_1096_split_sizes_0 = const()[name = tensor("op_1096_split_sizes_0"), val = tensor([144, 144])]; tensor var_1096_axis_0 = const()[name = tensor("op_1096_axis_0"), val = tensor(1)]; tensor var_1096_cast_fp16_0, tensor var_1096_cast_fp16_1 = split(axis = var_1096_axis_0, split_sizes = var_1096_split_sizes_0, x = inputs_45_cast_fp16)[name = tensor("op_1096_cast_fp16")]; tensor var_1098_cast_fp16 = sigmoid(x = var_1096_cast_fp16_1)[name = tensor("op_1098_cast_fp16")]; tensor input_213_cast_fp16 = mul(x = var_1096_cast_fp16_0, y = var_1098_cast_fp16)[name = tensor("input_213_cast_fp16")]; tensor var_1102 = const()[name = tensor("op_1102"), val = tensor([1])]; tensor var_1104 = const()[name = tensor("op_1104"), val = tensor([1])]; tensor input_215_pad_type_0 = const()[name = tensor("input_215_pad_type_0"), val = tensor("custom")]; tensor input_215_pad_0 = const()[name = tensor("input_215_pad_0"), val = tensor([1, 1])]; tensor inputs_47_weight_0_to_fp16 = const()[name = tensor("inputs_47_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2711552)))]; tensor inputs_47_bias_0_to_fp16 = const()[name = tensor("inputs_47_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2712512)))]; tensor inputs_47_cast_fp16 = conv(bias = inputs_47_bias_0_to_fp16, dilations = var_1104, groups = var_13, pad = input_215_pad_0, pad_type = input_215_pad_type_0, strides = var_1102, weight = inputs_47_weight_0_to_fp16, x = input_213_cast_fp16)[name = tensor("inputs_47_cast_fp16")]; tensor input_217_cast_fp16 = silu(x = inputs_47_cast_fp16)[name = tensor("input_217_cast_fp16")]; tensor var_1117 = const()[name = tensor("op_1117"), val = tensor([1])]; tensor var_1119 = const()[name = tensor("op_1119"), val = tensor([1])]; tensor input_219_pad_type_0 = const()[name = tensor("input_219_pad_type_0"), val = tensor("custom")]; tensor input_219_pad_0 = const()[name = tensor("input_219_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_4_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2712896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2733952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2733696)))]; tensor encoder_layers_4_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2734336)))]; tensor input_219_cast_fp16 = conv(bias = encoder_layers_4_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_1119, groups = var_15, pad = input_219_pad_0, pad_type = input_219_pad_type_0, strides = var_1117, weight = encoder_layers_4_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_217_cast_fp16)[name = tensor("input_219_cast_fp16")]; tensor var_1123_perm_0 = const()[name = tensor("op_1123_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_43 = transpose(perm = var_1123_perm_0, x = input_219_cast_fp16)[name = tensor("transpose_43")]; tensor input_221_cast_fp16 = add(x = transpose_43, y = input_209_cast_fp16)[name = tensor("input_221_cast_fp16")]; tensor input_223_axes_0 = const()[name = tensor("input_223_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2734720)))]; tensor encoder_layers_4_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2735104)))]; tensor input_223_cast_fp16 = layer_norm(axes = input_223_axes_0, beta = encoder_layers_4_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_4_sequential_3_module_sequential_0_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("input_223_cast_fp16")]; tensor encoder_layers_4_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2735488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2819136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2818496)))]; tensor encoder_layers_4_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2820352)))]; tensor linear_44_cast_fp16 = linear(bias = encoder_layers_4_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_4_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_223_cast_fp16)[name = tensor("linear_44_cast_fp16")]; tensor input_225_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_225_cast_fp16")]; tensor encoder_layers_4_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_4_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2821568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904576)))]; tensor encoder_layers_4_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2905216)))]; tensor linear_45_cast_fp16 = linear(bias = encoder_layers_4_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_4_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_225_cast_fp16)[name = tensor("linear_45_cast_fp16")]; tensor var_1150_to_fp16 = const()[name = tensor("op_1150_to_fp16"), val = tensor(0x1p-1)]; tensor var_1151_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1150_to_fp16)[name = tensor("op_1151_cast_fp16")]; tensor input_231_cast_fp16 = add(x = var_1151_cast_fp16, y = input_221_cast_fp16)[name = tensor("input_231_cast_fp16")]; tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; tensor encoder_layers_4_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2905600)))]; tensor encoder_layers_4_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_4_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2905984)))]; tensor input_233_cast_fp16 = layer_norm(axes = input_233_axes_0, beta = encoder_layers_4_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_4_sequential_4_weight_to_fp16, x = input_231_cast_fp16)[name = tensor("input_233_cast_fp16")]; tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_sequential_0_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_0_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2906368)))]; tensor encoder_layers_5_sequential_0_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_0_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2906752)))]; tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_layers_5_sequential_0_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_5_sequential_0_module_sequential_0_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; tensor encoder_layers_5_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2907136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2990784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2990144)))]; tensor encoder_layers_5_sequential_0_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_0_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2992000)))]; tensor linear_46_cast_fp16 = linear(bias = encoder_layers_5_sequential_0_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_5_sequential_0_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_235_cast_fp16)[name = tensor("linear_46_cast_fp16")]; tensor input_237_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_237_cast_fp16")]; tensor encoder_layers_5_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2993216))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3076480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3076224)))]; tensor encoder_layers_5_sequential_0_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_0_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3076864)))]; tensor linear_47_cast_fp16 = linear(bias = encoder_layers_5_sequential_0_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_5_sequential_0_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_237_cast_fp16)[name = tensor("linear_47_cast_fp16")]; tensor var_1186_to_fp16 = const()[name = tensor("op_1186_to_fp16"), val = tensor(0x1p-1)]; tensor var_1187_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1186_to_fp16)[name = tensor("op_1187_cast_fp16")]; tensor inputs_53_cast_fp16 = add(x = var_1187_cast_fp16, y = input_233_cast_fp16)[name = tensor("inputs_53_cast_fp16")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_sequential_1_module_layer_norm_weight_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3077248)))]; tensor encoder_layers_5_sequential_1_module_layer_norm_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3077632)))]; tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = encoder_layers_5_sequential_1_module_layer_norm_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_5_sequential_1_module_layer_norm_weight_to_fp16, x = inputs_53_cast_fp16)[name = tensor("query_21_cast_fp16")]; tensor encoder_layers_5_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3078016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3099072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3098816)))]; tensor encoder_layers_5_sequential_1_module_attention_query_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_query_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3099456)))]; tensor linear_48_cast_fp16 = linear(bias = encoder_layers_5_sequential_1_module_attention_query_proj_linear_bias_to_fp16, weight = encoder_layers_5_sequential_1_module_attention_query_proj_linear_weight_to_fp16_affine_quantized, x = query_21_cast_fp16)[name = tensor("linear_48_cast_fp16")]; tensor var_1218 = const()[name = tensor("op_1218"), val = tensor([1, -1, 4, 36])]; tensor query_cast_fp16 = reshape(shape = var_1218, x = linear_48_cast_fp16)[name = tensor("query_cast_fp16")]; tensor encoder_layers_5_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3099840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3120896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3120640)))]; tensor encoder_layers_5_sequential_1_module_attention_key_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_key_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3121280)))]; tensor linear_49_cast_fp16 = linear(bias = encoder_layers_5_sequential_1_module_attention_key_proj_linear_bias_to_fp16, weight = encoder_layers_5_sequential_1_module_attention_key_proj_linear_weight_to_fp16_affine_quantized, x = query_21_cast_fp16)[name = tensor("linear_49_cast_fp16")]; tensor var_1224 = const()[name = tensor("op_1224"), val = tensor([1, -1, 4, 36])]; tensor var_1225_cast_fp16 = reshape(shape = var_1224, x = linear_49_cast_fp16)[name = tensor("op_1225_cast_fp16")]; tensor encoder_layers_5_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3121664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3142720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3142464)))]; tensor encoder_layers_5_sequential_1_module_attention_value_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_value_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3143104)))]; tensor linear_50_cast_fp16 = linear(bias = encoder_layers_5_sequential_1_module_attention_value_proj_linear_bias_to_fp16, weight = encoder_layers_5_sequential_1_module_attention_value_proj_linear_weight_to_fp16_affine_quantized, x = query_21_cast_fp16)[name = tensor("linear_50_cast_fp16")]; tensor var_1232 = const()[name = tensor("op_1232"), val = tensor([1, -1, 4, 36])]; tensor var_1233_cast_fp16 = reshape(shape = var_1232, x = linear_50_cast_fp16)[name = tensor("op_1233_cast_fp16")]; tensor var_1234 = const()[name = tensor("op_1234"), val = tensor([0, 2, -3, -1])]; tensor encoder_layers_5_sequential_1_module_attention_u_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_u_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3143488)))]; tensor var_1241_cast_fp16 = add(x = query_cast_fp16, y = encoder_layers_5_sequential_1_module_attention_u_bias_to_fp16)[name = tensor("op_1241_cast_fp16")]; tensor content_score_transpose_x_0 = const()[name = tensor("content_score_transpose_x_0"), val = tensor(false)]; tensor content_score_transpose_y_0 = const()[name = tensor("content_score_transpose_y_0"), val = tensor(false)]; tensor transpose_34_perm_0 = const()[name = tensor("transpose_34_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_35_perm_0 = const()[name = tensor("transpose_35_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_41 = transpose(perm = transpose_35_perm_0, x = var_1225_cast_fp16)[name = tensor("transpose_41")]; tensor transpose_42 = transpose(perm = transpose_34_perm_0, x = var_1241_cast_fp16)[name = tensor("transpose_42")]; tensor content_score_cast_fp16 = matmul(transpose_x = content_score_transpose_x_0, transpose_y = content_score_transpose_y_0, x = transpose_42, y = transpose_41)[name = tensor("content_score_cast_fp16")]; tensor encoder_layers_5_sequential_1_module_attention_v_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3143872)))]; tensor var_1245_cast_fp16 = add(x = query_cast_fp16, y = encoder_layers_5_sequential_1_module_attention_v_bias_to_fp16)[name = tensor("op_1245_cast_fp16")]; tensor var_1246_perm_0 = const()[name = tensor("op_1246_perm_0"), val = tensor([0, 2, -3, -1])]; tensor pos_score_21_transpose_x_0 = const()[name = tensor("pos_score_21_transpose_x_0"), val = tensor(false)]; tensor pos_score_21_transpose_y_0 = const()[name = tensor("pos_score_21_transpose_y_0"), val = tensor(false)]; tensor op_1248_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(-1), name = tensor("op_1248_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3144256))), scale = tensor(0x1.76cp-5), zero_point = tensor(9)]; tensor transpose_40 = transpose(perm = var_1246_perm_0, x = var_1245_cast_fp16)[name = tensor("transpose_40")]; tensor pos_score_21_cast_fp16 = matmul(transpose_x = pos_score_21_transpose_x_0, transpose_y = pos_score_21_transpose_y_0, x = transpose_40, y = op_1248_to_fp16_affine_quantized)[name = tensor("pos_score_21_cast_fp16")]; tensor padded_pos_score_21_interleave_0 = const()[name = tensor("padded_pos_score_21_interleave_0"), val = tensor(false)]; tensor padded_pos_score_21_cast_fp16 = concat(axis = var_5, interleave = padded_pos_score_21_interleave_0, values = (zeros_1_to_fp16, pos_score_21_cast_fp16))[name = tensor("padded_pos_score_21_cast_fp16")]; tensor var_1262 = const()[name = tensor("op_1262"), val = tensor([1, 4, 49, 48])]; tensor padded_pos_score_cast_fp16 = reshape(shape = var_1262, x = padded_pos_score_21_cast_fp16)[name = tensor("padded_pos_score_cast_fp16")]; tensor var_1266_begin_0 = const()[name = tensor("op_1266_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1266_end_0 = const()[name = tensor("op_1266_end_0"), val = tensor([1, 4, 49, 48])]; tensor var_1266_end_mask_0 = const()[name = tensor("op_1266_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1266_cast_fp16 = slice_by_index(begin = var_1266_begin_0, end = var_1266_end_0, end_mask = var_1266_end_mask_0, x = padded_pos_score_cast_fp16)[name = tensor("op_1266_cast_fp16")]; tensor var_1273_cast_fp16 = add(x = content_score_cast_fp16, y = var_1266_cast_fp16)[name = tensor("op_1273_cast_fp16")]; tensor _inversed_input_245_y_0_to_fp16 = const()[name = tensor("_inversed_input_245_y_0_to_fp16"), val = tensor(0x1.554p-4)]; tensor _inversed_input_245_cast_fp16 = mul(x = var_1273_cast_fp16, y = _inversed_input_245_y_0_to_fp16)[name = tensor("_inversed_input_245_cast_fp16")]; tensor input_247_cast_fp16 = softmax(axis = var_5, x = _inversed_input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; tensor var_1278_transpose_x_0 = const()[name = tensor("op_1278_transpose_x_0"), val = tensor(false)]; tensor var_1278_transpose_y_0 = const()[name = tensor("op_1278_transpose_y_0"), val = tensor(false)]; tensor transpose_39 = transpose(perm = var_1234, x = var_1233_cast_fp16)[name = tensor("transpose_39")]; tensor var_1278_cast_fp16 = matmul(transpose_x = var_1278_transpose_x_0, transpose_y = var_1278_transpose_y_0, x = input_247_cast_fp16, y = transpose_39)[name = tensor("op_1278_cast_fp16")]; tensor context_perm_0 = const()[name = tensor("context_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1281 = const()[name = tensor("op_1281"), val = tensor([1, -1, 144])]; tensor transpose_38 = transpose(perm = context_perm_0, x = var_1278_cast_fp16)[name = tensor("transpose_38")]; tensor input_249_cast_fp16 = reshape(shape = var_1281, x = transpose_38)[name = tensor("input_249_cast_fp16")]; tensor encoder_layers_5_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3151232))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3172288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3172032)))]; tensor encoder_layers_5_sequential_1_module_attention_out_proj_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_1_module_attention_out_proj_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3172672)))]; tensor linear_52_cast_fp16 = linear(bias = encoder_layers_5_sequential_1_module_attention_out_proj_linear_bias_to_fp16, weight = encoder_layers_5_sequential_1_module_attention_out_proj_linear_weight_to_fp16_affine_quantized, x = input_249_cast_fp16)[name = tensor("linear_52_cast_fp16")]; tensor input_253_cast_fp16 = add(x = linear_52_cast_fp16, y = inputs_53_cast_fp16)[name = tensor("input_253_cast_fp16")]; tensor x_axes_0 = const()[name = tensor("x_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_sequential_2_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_2_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3173056)))]; tensor encoder_layers_5_sequential_2_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_2_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3173440)))]; tensor x_cast_fp16 = layer_norm(axes = x_axes_0, beta = encoder_layers_5_sequential_2_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_5_sequential_2_module_sequential_0_weight_to_fp16, x = input_253_cast_fp16)[name = tensor("x_cast_fp16")]; tensor input_255_perm_0 = const()[name = tensor("input_255_perm_0"), val = tensor([0, 2, 1])]; tensor var_1308 = const()[name = tensor("op_1308"), val = tensor([1])]; tensor var_1310 = const()[name = tensor("op_1310"), val = tensor([1])]; tensor inputs_55_pad_type_0 = const()[name = tensor("inputs_55_pad_type_0"), val = tensor("custom")]; tensor inputs_55_pad_0 = const()[name = tensor("inputs_55_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_5_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3173824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3215744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3215360)))]; tensor encoder_layers_5_sequential_2_module_sequential_2_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_2_module_sequential_2_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3216384)))]; tensor transpose_37 = transpose(perm = input_255_perm_0, x = x_cast_fp16)[name = tensor("transpose_37")]; tensor inputs_55_cast_fp16 = conv(bias = encoder_layers_5_sequential_2_module_sequential_2_conv_bias_to_fp16, dilations = var_1310, groups = var_15, pad = inputs_55_pad_0, pad_type = inputs_55_pad_type_0, strides = var_1308, weight = encoder_layers_5_sequential_2_module_sequential_2_conv_weight_to_fp16_affine_quantized, x = transpose_37)[name = tensor("inputs_55_cast_fp16")]; tensor var_1313_split_sizes_0 = const()[name = tensor("op_1313_split_sizes_0"), val = tensor([144, 144])]; tensor var_1313_axis_0 = const()[name = tensor("op_1313_axis_0"), val = tensor(1)]; tensor var_1313_cast_fp16_0, tensor var_1313_cast_fp16_1 = split(axis = var_1313_axis_0, split_sizes = var_1313_split_sizes_0, x = inputs_55_cast_fp16)[name = tensor("op_1313_cast_fp16")]; tensor var_1315_cast_fp16 = sigmoid(x = var_1313_cast_fp16_1)[name = tensor("op_1315_cast_fp16")]; tensor input_257_cast_fp16 = mul(x = var_1313_cast_fp16_0, y = var_1315_cast_fp16)[name = tensor("input_257_cast_fp16")]; tensor var_1319 = const()[name = tensor("op_1319"), val = tensor([1])]; tensor var_1321 = const()[name = tensor("op_1321"), val = tensor([1])]; tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("custom")]; tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([1, 1])]; tensor inputs_57_weight_0_to_fp16 = const()[name = tensor("inputs_57_weight_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3217024)))]; tensor inputs_57_bias_0_to_fp16 = const()[name = tensor("inputs_57_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3217984)))]; tensor inputs_57_cast_fp16 = conv(bias = inputs_57_bias_0_to_fp16, dilations = var_1321, groups = var_13, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = var_1319, weight = inputs_57_weight_0_to_fp16, x = input_257_cast_fp16)[name = tensor("inputs_57_cast_fp16")]; tensor input_261_cast_fp16 = silu(x = inputs_57_cast_fp16)[name = tensor("input_261_cast_fp16")]; tensor var_1334 = const()[name = tensor("op_1334"), val = tensor([1])]; tensor var_1336 = const()[name = tensor("op_1336"), val = tensor([1])]; tensor input_263_pad_type_0 = const()[name = tensor("input_263_pad_type_0"), val = tensor("custom")]; tensor input_263_pad_0 = const()[name = tensor("input_263_pad_0"), val = tensor([0, 0])]; tensor encoder_layers_5_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3218368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3239424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3239168)))]; tensor encoder_layers_5_sequential_2_module_sequential_7_conv_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_2_module_sequential_7_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3239808)))]; tensor input_263_cast_fp16 = conv(bias = encoder_layers_5_sequential_2_module_sequential_7_conv_bias_to_fp16, dilations = var_1336, groups = var_15, pad = input_263_pad_0, pad_type = input_263_pad_type_0, strides = var_1334, weight = encoder_layers_5_sequential_2_module_sequential_7_conv_weight_to_fp16_affine_quantized, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; tensor var_1340_perm_0 = const()[name = tensor("op_1340_perm_0"), val = tensor([0, 2, 1])]; tensor transpose_36 = transpose(perm = var_1340_perm_0, x = input_263_cast_fp16)[name = tensor("transpose_36")]; tensor input_265_cast_fp16 = add(x = transpose_36, y = input_253_cast_fp16)[name = tensor("input_265_cast_fp16")]; tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_sequential_3_module_sequential_0_weight_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_3_module_sequential_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3240192)))]; tensor encoder_layers_5_sequential_3_module_sequential_0_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_3_module_sequential_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3240576)))]; tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = encoder_layers_5_sequential_3_module_sequential_0_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_5_sequential_3_module_sequential_0_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; tensor encoder_layers_5_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3240960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3324608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3323968)))]; tensor encoder_layers_5_sequential_3_module_sequential_1_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_3_module_sequential_1_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3325824)))]; tensor linear_53_cast_fp16 = linear(bias = encoder_layers_5_sequential_3_module_sequential_1_linear_bias_to_fp16, weight = encoder_layers_5_sequential_3_module_sequential_1_linear_weight_to_fp16_affine_quantized, x = input_267_cast_fp16)[name = tensor("linear_53_cast_fp16")]; tensor input_269_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_269_cast_fp16")]; tensor encoder_layers_5_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("encoder_layers_5_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3327040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3410304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3410048)))]; tensor encoder_layers_5_sequential_3_module_sequential_4_linear_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_3_module_sequential_4_linear_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3410688)))]; tensor linear_54_cast_fp16 = linear(bias = encoder_layers_5_sequential_3_module_sequential_4_linear_bias_to_fp16, weight = encoder_layers_5_sequential_3_module_sequential_4_linear_weight_to_fp16_affine_quantized, x = input_269_cast_fp16)[name = tensor("linear_54_cast_fp16")]; tensor var_1367_to_fp16 = const()[name = tensor("op_1367_to_fp16"), val = tensor(0x1p-1)]; tensor var_1368_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1367_to_fp16)[name = tensor("op_1368_cast_fp16")]; tensor input_cast_fp16 = add(x = var_1368_cast_fp16, y = input_265_cast_fp16)[name = tensor("input_cast_fp16")]; tensor var_1375_axes_0 = const()[name = tensor("op_1375_axes_0"), val = tensor([-1])]; tensor encoder_layers_5_sequential_4_weight_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_4_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3411072)))]; tensor encoder_layers_5_sequential_4_bias_to_fp16 = const()[name = tensor("encoder_layers_5_sequential_4_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3411456)))]; tensor enc_3d_output = layer_norm(axes = var_1375_axes_0, beta = encoder_layers_5_sequential_4_bias_to_fp16, epsilon = var_12_to_fp16, gamma = encoder_layers_5_sequential_4_weight_to_fp16, x = input_cast_fp16)[name = tensor("op_1375_cast_fp16")]; } -> (enc_3d_output); }