program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3520.2.1"}, {"coremlc-version", "3520.2.1"}, {"mldb_token", "mldb-rdg2nkxg5z"}})] { func main(tensor inp) { tensor input_1_axes_0 = const()[name = tensor("input_1_axes_0"), val = tensor([2])]; tensor inp_to_fp16_dtype_0 = const()[name = tensor("inp_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cast_1 = cast(dtype = inp_to_fp16_dtype_0, x = inp)[name = tensor("cast_1")]; tensor input_1_cast_fp16 = squeeze(axes = input_1_axes_0, x = cast_1)[name = tensor("input_1_cast_fp16")]; tensor var_14 = const()[name = tensor("op_14"), val = tensor(1)]; tensor var_17 = const()[name = tensor("op_17"), val = tensor([1])]; tensor var_19 = const()[name = tensor("op_19"), val = tensor([1])]; tensor input_5_pad_type_0 = const()[name = tensor("input_5_pad_type_0"), val = tensor("custom")]; tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0])]; tensor l_conv_1_cross_1_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("l_conv_1_cross_1_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131200)))]; tensor l_conv_1_cross_1_bias_to_fp16 = const()[name = tensor("l_conv_1_cross_1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132096)))]; tensor input_5_cast_fp16 = conv(bias = l_conv_1_cross_1_bias_to_fp16, dilations = var_19, groups = var_14, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = var_17, weight = l_conv_1_cross_1_weight_to_fp16_affine_quantized, x = input_1_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor var_26 = const()[name = tensor("op_26"), val = tensor(1)]; tensor var_29 = const()[name = tensor("op_29"), val = tensor([1])]; tensor var_31 = const()[name = tensor("op_31"), val = tensor([1])]; tensor input_9_pad_type_0 = const()[name = tensor("input_9_pad_type_0"), val = tensor("custom")]; tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0])]; tensor l_conv_1_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("l_conv_1_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132672))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(296768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(296576)))]; tensor l_conv_1_bias_to_fp16 = const()[name = tensor("l_conv_1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297088)))]; tensor input_9_cast_fp16 = conv(bias = l_conv_1_bias_to_fp16, dilations = var_31, groups = var_26, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = var_29, weight = l_conv_1_weight_to_fp16_affine_quantized, x = input_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; tensor var_39 = const()[name = tensor("op_39"), val = tensor([5])]; tensor var_40 = const()[name = tensor("op_40"), val = tensor([5])]; tensor input_13_pad_type_0 = const()[name = tensor("input_13_pad_type_0"), val = tensor("custom")]; tensor input_13_pad_0 = const()[name = tensor("input_13_pad_0"), val = tensor([0, 0])]; tensor input_13_ceil_mode_0 = const()[name = tensor("input_13_ceil_mode_0"), val = tensor(false)]; tensor input_13_cast_fp16 = max_pool(ceil_mode = input_13_ceil_mode_0, kernel_sizes = var_39, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = var_40, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; tensor var_47 = const()[name = tensor("op_47"), val = tensor(1)]; tensor var_50 = const()[name = tensor("op_50"), val = tensor([1])]; tensor var_52 = const()[name = tensor("op_52"), val = tensor([1])]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("custom")]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([0, 0])]; tensor l_conv_2_weight_to_fp16_affine_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("l_conv_2_weight_to_fp16_affine_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338432)))]; tensor l_conv_2_bias_to_fp16 = const()[name = tensor("l_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338752)))]; tensor input_15_cast_fp16 = conv(bias = l_conv_2_bias_to_fp16, dilations = var_52, groups = var_47, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = var_50, weight = l_conv_2_weight_to_fp16_affine_quantized, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor input_17_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor var_60 = const()[name = tensor("op_60"), val = tensor([5])]; tensor var_61 = const()[name = tensor("op_61"), val = tensor([5])]; tensor x_pad_type_0 = const()[name = tensor("x_pad_type_0"), val = tensor("custom")]; tensor x_pad_0 = const()[name = tensor("x_pad_0"), val = tensor([0, 0])]; tensor x_ceil_mode_0 = const()[name = tensor("x_ceil_mode_0"), val = tensor(false)]; tensor x_cast_fp16 = max_pool(ceil_mode = x_ceil_mode_0, kernel_sizes = var_60, pad = x_pad_0, pad_type = x_pad_type_0, strides = var_61, x = input_17_cast_fp16)[name = tensor("x_cast_fp16")]; tensor var_67 = const()[name = tensor("op_67"), val = tensor([-1, 576])]; tensor input_19_cast_fp16 = reshape(shape = var_67, x = x_cast_fp16)[name = tensor("input_19_cast_fp16")]; tensor l_fc_1_weight_to_fp16 = const()[name = tensor("l_fc_1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(338944)))]; tensor l_fc_1_bias_to_fp16 = const()[name = tensor("l_fc_1_bias_to_fp16"), val = tensor([-0x1.d0cp-5, 0x1.528p-5])]; tensor linear_0_cast_fp16 = linear(bias = l_fc_1_bias_to_fp16, weight = l_fc_1_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor linear_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("linear_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor linear_0 = cast(dtype = linear_0_cast_fp16_to_fp32_dtype_0, x = linear_0_cast_fp16)[name = tensor("cast_0")]; } -> (linear_0); }