program(1.3) [buildInfo = dict({{"coremlc-component-MIL", "3402.3.2"}, {"coremlc-version", "3402.5.1"}, {"mldb_token", "mldb-7mar5v7y6j"}})] { func main(pixel_buffer image) { tensor img_tensor_raw = pixel_buffer_to_tensor(input=image); tensor img_tensor_sliced = slice_by_index(x = img_tensor_raw, begin = tensor([0, 0, 0]), end = tensor([3, 0, 0]), end_mask = tensor([false, true, true])); tensor img_bgr_batch = expand_dims(axes = tensor([0]), x = img_tensor_sliced); string image_to_fp16_dtype_0 = const()[name = string("image_to_fp16_dtype_0"), val = string("fp16")]; tensor image_to_fp16 = cast(dtype = image_to_fp16_dtype_0, x = img_bgr_batch)[name = string("cast_0")]; fp16 image__scaled___y_0 = const()[name = string("image__scaled___y_0"), val = fp16(0x1.01p-8)]; tensor image__scaled__ = mul(x = image_to_fp16, y = image__scaled___y_0)[name = string("image__scaled__")]; tensor stage0_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(64))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1472))))[name = string("stage0_reparam_conv_weight_cast_fp16")]; tensor stage1_0_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1664))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(2176))))[name = string("stage1_0_reparam_conv_weight_cast_fp16")]; tensor stage1_1_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(2368))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(4736))))[name = string("stage1_1_reparam_conv_weight_cast_fp16")]; tensor stage1_2_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(4928))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(5440))))[name = string("stage1_2_reparam_conv_weight_cast_fp16")]; tensor stage1_3_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(5632))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(8000))))[name = string("stage1_3_reparam_conv_weight_cast_fp16")]; tensor stage2_0_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(8192))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(8704))))[name = string("stage2_0_reparam_conv_weight_cast_fp16")]; tensor stage2_1_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(8896))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(15104))))[name = string("stage2_1_reparam_conv_weight_cast_fp16")]; tensor stage2_2_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(15424))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(16640))))[name = string("stage2_2_reparam_conv_weight_cast_fp16")]; tensor stage2_3_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(16960))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(33408))))[name = string("stage2_3_reparam_conv_weight_cast_fp16")]; tensor stage2_4_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(33728))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(34944))))[name = string("stage2_4_reparam_conv_weight_cast_fp16")]; tensor stage2_5_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(35264))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(51712))))[name = string("stage2_5_reparam_conv_weight_cast_fp16")]; tensor stage2_6_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(52032))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(53248))))[name = string("stage2_6_reparam_conv_weight_cast_fp16")]; tensor stage2_7_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(53568))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(70016))))[name = string("stage2_7_reparam_conv_weight_cast_fp16")]; tensor stage2_8_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(70336))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(71552))))[name = string("stage2_8_reparam_conv_weight_cast_fp16")]; tensor stage2_9_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(71872))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(88320))))[name = string("stage2_9_reparam_conv_weight_cast_fp16")]; tensor stage2_10_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(88640))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(89856))))[name = string("stage2_10_reparam_conv_weight_cast_fp16")]; tensor stage2_11_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(90176))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(106624))))[name = string("stage2_11_reparam_conv_weight_cast_fp16")]; tensor stage2_12_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(106944))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(108160))))[name = string("stage2_12_reparam_conv_weight_cast_fp16")]; tensor stage2_13_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(108480))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(124928))))[name = string("stage2_13_reparam_conv_weight_cast_fp16")]; tensor stage2_14_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(125248))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(126464))))[name = string("stage2_14_reparam_conv_weight_cast_fp16")]; tensor stage2_15_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(126784))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(143232))))[name = string("stage2_15_reparam_conv_weight_cast_fp16")]; tensor stage3_0_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(143552))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(144768))))[name = string("stage3_0_reparam_conv_weight_cast_fp16")]; tensor stage3_1_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(145088))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(177920))))[name = string("stage3_1_reparam_conv_weight_cast_fp16")]; tensor stage3_2_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(178496))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(180864))))[name = string("stage3_2_reparam_conv_weight_cast_fp16")]; tensor stage3_3_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(181440))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(247040))))[name = string("stage3_3_reparam_conv_weight_cast_fp16")]; tensor stage3_4_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(247616))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(249984))))[name = string("stage3_4_reparam_conv_weight_cast_fp16")]; tensor stage3_5_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(250560))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(316160))))[name = string("stage3_5_reparam_conv_weight_cast_fp16")]; tensor stage3_6_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(316736))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(319104))))[name = string("stage3_6_reparam_conv_weight_cast_fp16")]; tensor stage3_7_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(319680))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(385280))))[name = string("stage3_7_reparam_conv_weight_cast_fp16")]; tensor stage3_8_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(385856))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(388224))))[name = string("stage3_8_reparam_conv_weight_cast_fp16")]; tensor stage3_9_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(388800))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(454400))))[name = string("stage3_9_reparam_conv_weight_cast_fp16")]; tensor stage3_10_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(454976))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(457344))))[name = string("stage3_10_reparam_conv_weight_cast_fp16")]; tensor stage3_11_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(457920))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(523520))))[name = string("stage3_11_reparam_conv_weight_cast_fp16")]; tensor stage3_12_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(524096))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(526464))))[name = string("stage3_12_reparam_conv_weight_cast_fp16")]; tensor stage3_13_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(527040))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(592640))))[name = string("stage3_13_reparam_conv_weight_cast_fp16")]; tensor stage3_14_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(593216))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(595584))))[name = string("stage3_14_reparam_conv_weight_cast_fp16")]; tensor stage3_15_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(596160))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(661760))))[name = string("stage3_15_reparam_conv_weight_cast_fp16")]; tensor stage3_16_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(662336))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(664704))))[name = string("stage3_16_reparam_conv_weight_cast_fp16")]; tensor stage3_17_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(665280))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(730880))))[name = string("stage3_17_reparam_conv_weight_cast_fp16")]; tensor stage3_18_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(731456))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(733824))))[name = string("stage3_18_reparam_conv_weight_cast_fp16")]; tensor stage3_19_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(734400))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(800000))))[name = string("stage3_19_reparam_conv_weight_cast_fp16")]; tensor stage4_0_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(800576))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(802944))))[name = string("stage4_0_reparam_conv_weight_cast_fp16")]; tensor stage4_1_reparam_conv_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(803520))), scale = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1065728))))[name = string("stage4_1_reparam_conv_weight_cast_fp16")]; tensor linear_weight_cast_fp16 = constexpr_blockwise_shift_scale(data = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1067840))), scale = tensor([[0x1.6ecp-10]]))[name = string("linear_weight_cast_fp16")]; string input_1_pad_type_0 = const()[name = string("input_1_pad_type_0"), val = string("custom")]; tensor input_1_pad_0 = const()[name = string("input_1_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_1_strides_0 = const()[name = string("input_1_strides_0"), val = tensor([2, 2])]; tensor input_1_dilations_0 = const()[name = string("input_1_dilations_0"), val = tensor([1, 1])]; int32 input_1_groups_0 = const()[name = string("input_1_groups_0"), val = int32(1)]; tensor stage0_reparam_conv_bias_to_fp16 = const()[name = string("stage0_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1068928)))]; fp16 quantize_0_scale_1 = const()[name = string("quantize_0_scale_1"), val = fp16(0x1.02p-7)]; string quantize_0_output_dtype_1 = const()[name = string("quantize_0_output_dtype_1"), val = string("int8")]; tensor quantize_0 = quantize(input = image__scaled__, output_dtype = quantize_0_output_dtype_1, scale = quantize_0_scale_1)[name = string("quantize_0")]; fp16 dequantize_0_scale_1 = const()[name = string("dequantize_0_scale_1"), val = fp16(0x1.02p-7)]; tensor dequantize_0 = dequantize(input = quantize_0, scale = dequantize_0_scale_1)[name = string("dequantize_0")]; tensor input_1_cast_fp16 = conv(bias = stage0_reparam_conv_bias_to_fp16, dilations = input_1_dilations_0, groups = input_1_groups_0, pad = input_1_pad_0, pad_type = input_1_pad_type_0, strides = input_1_strides_0, weight = stage0_reparam_conv_weight_cast_fp16, x = dequantize_0)[name = string("input_1_cast_fp16")]; tensor input_3_cast_fp16 = relu(x = input_1_cast_fp16)[name = string("input_3_cast_fp16")]; string input_5_pad_type_0 = const()[name = string("input_5_pad_type_0"), val = string("custom")]; tensor input_5_pad_0 = const()[name = string("input_5_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_5_strides_0 = const()[name = string("input_5_strides_0"), val = tensor([2, 2])]; int32 input_5_groups_0 = const()[name = string("input_5_groups_0"), val = int32(48)]; tensor input_5_dilations_0 = const()[name = string("input_5_dilations_0"), val = tensor([1, 1])]; tensor stage1_0_reparam_conv_bias_to_fp16 = const()[name = string("stage1_0_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1069120)))]; fp16 quantize_1_scale_1 = const()[name = string("quantize_1_scale_1"), val = fp16(0x1.538p-8)]; string quantize_1_output_dtype_1 = const()[name = string("quantize_1_output_dtype_1"), val = string("int8")]; tensor quantize_1 = quantize(input = input_3_cast_fp16, output_dtype = quantize_1_output_dtype_1, scale = quantize_1_scale_1)[name = string("quantize_1")]; fp16 dequantize_1_scale_1 = const()[name = string("dequantize_1_scale_1"), val = fp16(0x1.538p-8)]; tensor dequantize_1 = dequantize(input = quantize_1, scale = dequantize_1_scale_1)[name = string("dequantize_1")]; tensor input_5_cast_fp16 = conv(bias = stage1_0_reparam_conv_bias_to_fp16, dilations = input_5_dilations_0, groups = input_5_groups_0, pad = input_5_pad_0, pad_type = input_5_pad_type_0, strides = input_5_strides_0, weight = stage1_0_reparam_conv_weight_cast_fp16, x = dequantize_1)[name = string("input_5_cast_fp16")]; tensor input_7_cast_fp16 = relu(x = input_5_cast_fp16)[name = string("input_7_cast_fp16")]; string input_9_pad_type_0 = const()[name = string("input_9_pad_type_0"), val = string("valid")]; tensor input_9_strides_0 = const()[name = string("input_9_strides_0"), val = tensor([1, 1])]; tensor input_9_pad_0 = const()[name = string("input_9_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_9_dilations_0 = const()[name = string("input_9_dilations_0"), val = tensor([1, 1])]; int32 input_9_groups_0 = const()[name = string("input_9_groups_0"), val = int32(1)]; tensor stage1_1_reparam_conv_bias_to_fp16 = const()[name = string("stage1_1_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1069312)))]; fp16 quantize_2_scale_1 = const()[name = string("quantize_2_scale_1"), val = fp16(0x1.23cp-8)]; string quantize_2_output_dtype_1 = const()[name = string("quantize_2_output_dtype_1"), val = string("int8")]; tensor quantize_2 = quantize(input = input_7_cast_fp16, output_dtype = quantize_2_output_dtype_1, scale = quantize_2_scale_1)[name = string("quantize_2")]; fp16 dequantize_2_scale_1 = const()[name = string("dequantize_2_scale_1"), val = fp16(0x1.23cp-8)]; tensor dequantize_2 = dequantize(input = quantize_2, scale = dequantize_2_scale_1)[name = string("dequantize_2")]; tensor input_9_cast_fp16 = conv(bias = stage1_1_reparam_conv_bias_to_fp16, dilations = input_9_dilations_0, groups = input_9_groups_0, pad = input_9_pad_0, pad_type = input_9_pad_type_0, strides = input_9_strides_0, weight = stage1_1_reparam_conv_weight_cast_fp16, x = dequantize_2)[name = string("input_9_cast_fp16")]; tensor input_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = string("input_11_cast_fp16")]; string input_13_pad_type_0 = const()[name = string("input_13_pad_type_0"), val = string("custom")]; tensor input_13_pad_0 = const()[name = string("input_13_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_13_groups_0 = const()[name = string("input_13_groups_0"), val = int32(48)]; tensor input_13_strides_0 = const()[name = string("input_13_strides_0"), val = tensor([1, 1])]; tensor input_13_dilations_0 = const()[name = string("input_13_dilations_0"), val = tensor([1, 1])]; tensor stage1_2_reparam_conv_bias_to_fp16 = const()[name = string("stage1_2_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1069504)))]; fp16 quantize_3_scale_1 = const()[name = string("quantize_3_scale_1"), val = fp16(0x1.758p-8)]; string quantize_3_output_dtype_1 = const()[name = string("quantize_3_output_dtype_1"), val = string("int8")]; tensor quantize_3 = quantize(input = input_11_cast_fp16, output_dtype = quantize_3_output_dtype_1, scale = quantize_3_scale_1)[name = string("quantize_3")]; fp16 dequantize_3_scale_1 = const()[name = string("dequantize_3_scale_1"), val = fp16(0x1.758p-8)]; tensor dequantize_3 = dequantize(input = quantize_3, scale = dequantize_3_scale_1)[name = string("dequantize_3")]; tensor input_13_cast_fp16 = conv(bias = stage1_2_reparam_conv_bias_to_fp16, dilations = input_13_dilations_0, groups = input_13_groups_0, pad = input_13_pad_0, pad_type = input_13_pad_type_0, strides = input_13_strides_0, weight = stage1_2_reparam_conv_weight_cast_fp16, x = dequantize_3)[name = string("input_13_cast_fp16")]; tensor input_15_cast_fp16 = relu(x = input_13_cast_fp16)[name = string("input_15_cast_fp16")]; string input_17_pad_type_0 = const()[name = string("input_17_pad_type_0"), val = string("valid")]; tensor input_17_strides_0 = const()[name = string("input_17_strides_0"), val = tensor([1, 1])]; tensor input_17_pad_0 = const()[name = string("input_17_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_17_dilations_0 = const()[name = string("input_17_dilations_0"), val = tensor([1, 1])]; int32 input_17_groups_0 = const()[name = string("input_17_groups_0"), val = int32(1)]; tensor stage1_3_reparam_conv_bias_to_fp16 = const()[name = string("stage1_3_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1069696)))]; fp16 quantize_4_scale_1 = const()[name = string("quantize_4_scale_1"), val = fp16(0x1.45cp-8)]; string quantize_4_output_dtype_1 = const()[name = string("quantize_4_output_dtype_1"), val = string("int8")]; tensor quantize_4 = quantize(input = input_15_cast_fp16, output_dtype = quantize_4_output_dtype_1, scale = quantize_4_scale_1)[name = string("quantize_4")]; fp16 dequantize_4_scale_1 = const()[name = string("dequantize_4_scale_1"), val = fp16(0x1.45cp-8)]; tensor dequantize_4 = dequantize(input = quantize_4, scale = dequantize_4_scale_1)[name = string("dequantize_4")]; tensor input_17_cast_fp16 = conv(bias = stage1_3_reparam_conv_bias_to_fp16, dilations = input_17_dilations_0, groups = input_17_groups_0, pad = input_17_pad_0, pad_type = input_17_pad_type_0, strides = input_17_strides_0, weight = stage1_3_reparam_conv_weight_cast_fp16, x = dequantize_4)[name = string("input_17_cast_fp16")]; tensor input_19_cast_fp16 = relu(x = input_17_cast_fp16)[name = string("input_19_cast_fp16")]; string input_21_pad_type_0 = const()[name = string("input_21_pad_type_0"), val = string("custom")]; tensor input_21_pad_0 = const()[name = string("input_21_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_21_strides_0 = const()[name = string("input_21_strides_0"), val = tensor([2, 2])]; int32 input_21_groups_0 = const()[name = string("input_21_groups_0"), val = int32(48)]; tensor input_21_dilations_0 = const()[name = string("input_21_dilations_0"), val = tensor([1, 1])]; tensor stage2_0_reparam_conv_bias_to_fp16 = const()[name = string("stage2_0_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1069888)))]; fp16 quantize_5_scale_1 = const()[name = string("quantize_5_scale_1"), val = fp16(0x1.104p-8)]; string quantize_5_output_dtype_1 = const()[name = string("quantize_5_output_dtype_1"), val = string("int8")]; tensor quantize_5 = quantize(input = input_19_cast_fp16, output_dtype = quantize_5_output_dtype_1, scale = quantize_5_scale_1)[name = string("quantize_5")]; fp16 dequantize_5_scale_1 = const()[name = string("dequantize_5_scale_1"), val = fp16(0x1.104p-8)]; tensor dequantize_5 = dequantize(input = quantize_5, scale = dequantize_5_scale_1)[name = string("dequantize_5")]; tensor input_21_cast_fp16 = conv(bias = stage2_0_reparam_conv_bias_to_fp16, dilations = input_21_dilations_0, groups = input_21_groups_0, pad = input_21_pad_0, pad_type = input_21_pad_type_0, strides = input_21_strides_0, weight = stage2_0_reparam_conv_weight_cast_fp16, x = dequantize_5)[name = string("input_21_cast_fp16")]; tensor input_23_cast_fp16 = relu(x = input_21_cast_fp16)[name = string("input_23_cast_fp16")]; string input_25_pad_type_0 = const()[name = string("input_25_pad_type_0"), val = string("valid")]; tensor input_25_strides_0 = const()[name = string("input_25_strides_0"), val = tensor([1, 1])]; tensor input_25_pad_0 = const()[name = string("input_25_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_25_dilations_0 = const()[name = string("input_25_dilations_0"), val = tensor([1, 1])]; int32 input_25_groups_0 = const()[name = string("input_25_groups_0"), val = int32(1)]; tensor stage2_1_reparam_conv_bias_to_fp16 = const()[name = string("stage2_1_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1070080)))]; fp16 quantize_6_scale_1 = const()[name = string("quantize_6_scale_1"), val = fp16(0x1.094p-8)]; string quantize_6_output_dtype_1 = const()[name = string("quantize_6_output_dtype_1"), val = string("int8")]; tensor quantize_6 = quantize(input = input_23_cast_fp16, output_dtype = quantize_6_output_dtype_1, scale = quantize_6_scale_1)[name = string("quantize_6")]; fp16 dequantize_6_scale_1 = const()[name = string("dequantize_6_scale_1"), val = fp16(0x1.094p-8)]; tensor dequantize_6 = dequantize(input = quantize_6, scale = dequantize_6_scale_1)[name = string("dequantize_6")]; tensor input_25_cast_fp16 = conv(bias = stage2_1_reparam_conv_bias_to_fp16, dilations = input_25_dilations_0, groups = input_25_groups_0, pad = input_25_pad_0, pad_type = input_25_pad_type_0, strides = input_25_strides_0, weight = stage2_1_reparam_conv_weight_cast_fp16, x = dequantize_6)[name = string("input_25_cast_fp16")]; tensor input_27_cast_fp16 = relu(x = input_25_cast_fp16)[name = string("input_27_cast_fp16")]; string input_29_pad_type_0 = const()[name = string("input_29_pad_type_0"), val = string("custom")]; tensor input_29_pad_0 = const()[name = string("input_29_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_29_groups_0 = const()[name = string("input_29_groups_0"), val = int32(128)]; tensor input_29_strides_0 = const()[name = string("input_29_strides_0"), val = tensor([1, 1])]; tensor input_29_dilations_0 = const()[name = string("input_29_dilations_0"), val = tensor([1, 1])]; tensor stage2_2_reparam_conv_bias_to_fp16 = const()[name = string("stage2_2_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1070400)))]; fp16 quantize_7_scale_1 = const()[name = string("quantize_7_scale_1"), val = fp16(0x1.57cp-9)]; string quantize_7_output_dtype_1 = const()[name = string("quantize_7_output_dtype_1"), val = string("int8")]; tensor quantize_7 = quantize(input = input_27_cast_fp16, output_dtype = quantize_7_output_dtype_1, scale = quantize_7_scale_1)[name = string("quantize_7")]; fp16 dequantize_7_scale_1 = const()[name = string("dequantize_7_scale_1"), val = fp16(0x1.57cp-9)]; tensor dequantize_7 = dequantize(input = quantize_7, scale = dequantize_7_scale_1)[name = string("dequantize_7")]; tensor input_29_cast_fp16 = conv(bias = stage2_2_reparam_conv_bias_to_fp16, dilations = input_29_dilations_0, groups = input_29_groups_0, pad = input_29_pad_0, pad_type = input_29_pad_type_0, strides = input_29_strides_0, weight = stage2_2_reparam_conv_weight_cast_fp16, x = dequantize_7)[name = string("input_29_cast_fp16")]; tensor input_31_cast_fp16 = relu(x = input_29_cast_fp16)[name = string("input_31_cast_fp16")]; string input_33_pad_type_0 = const()[name = string("input_33_pad_type_0"), val = string("valid")]; tensor input_33_strides_0 = const()[name = string("input_33_strides_0"), val = tensor([1, 1])]; tensor input_33_pad_0 = const()[name = string("input_33_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_33_dilations_0 = const()[name = string("input_33_dilations_0"), val = tensor([1, 1])]; int32 input_33_groups_0 = const()[name = string("input_33_groups_0"), val = int32(1)]; tensor stage2_3_reparam_conv_bias_to_fp16 = const()[name = string("stage2_3_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1070720)))]; fp16 quantize_8_scale_1 = const()[name = string("quantize_8_scale_1"), val = fp16(0x1.37cp-9)]; string quantize_8_output_dtype_1 = const()[name = string("quantize_8_output_dtype_1"), val = string("int8")]; tensor quantize_8 = quantize(input = input_31_cast_fp16, output_dtype = quantize_8_output_dtype_1, scale = quantize_8_scale_1)[name = string("quantize_8")]; fp16 dequantize_8_scale_1 = const()[name = string("dequantize_8_scale_1"), val = fp16(0x1.37cp-9)]; tensor dequantize_8 = dequantize(input = quantize_8, scale = dequantize_8_scale_1)[name = string("dequantize_8")]; tensor input_33_cast_fp16 = conv(bias = stage2_3_reparam_conv_bias_to_fp16, dilations = input_33_dilations_0, groups = input_33_groups_0, pad = input_33_pad_0, pad_type = input_33_pad_type_0, strides = input_33_strides_0, weight = stage2_3_reparam_conv_weight_cast_fp16, x = dequantize_8)[name = string("input_33_cast_fp16")]; tensor input_35_cast_fp16 = relu(x = input_33_cast_fp16)[name = string("input_35_cast_fp16")]; string input_37_pad_type_0 = const()[name = string("input_37_pad_type_0"), val = string("custom")]; tensor input_37_pad_0 = const()[name = string("input_37_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_37_groups_0 = const()[name = string("input_37_groups_0"), val = int32(128)]; tensor input_37_strides_0 = const()[name = string("input_37_strides_0"), val = tensor([1, 1])]; tensor input_37_dilations_0 = const()[name = string("input_37_dilations_0"), val = tensor([1, 1])]; tensor stage2_4_reparam_conv_bias_to_fp16 = const()[name = string("stage2_4_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1071040)))]; fp16 quantize_9_scale_1 = const()[name = string("quantize_9_scale_1"), val = fp16(0x1.494p-9)]; string quantize_9_output_dtype_1 = const()[name = string("quantize_9_output_dtype_1"), val = string("int8")]; tensor quantize_9 = quantize(input = input_35_cast_fp16, output_dtype = quantize_9_output_dtype_1, scale = quantize_9_scale_1)[name = string("quantize_9")]; fp16 dequantize_9_scale_1 = const()[name = string("dequantize_9_scale_1"), val = fp16(0x1.494p-9)]; tensor dequantize_9 = dequantize(input = quantize_9, scale = dequantize_9_scale_1)[name = string("dequantize_9")]; tensor input_37_cast_fp16 = conv(bias = stage2_4_reparam_conv_bias_to_fp16, dilations = input_37_dilations_0, groups = input_37_groups_0, pad = input_37_pad_0, pad_type = input_37_pad_type_0, strides = input_37_strides_0, weight = stage2_4_reparam_conv_weight_cast_fp16, x = dequantize_9)[name = string("input_37_cast_fp16")]; tensor input_39_cast_fp16 = relu(x = input_37_cast_fp16)[name = string("input_39_cast_fp16")]; string input_41_pad_type_0 = const()[name = string("input_41_pad_type_0"), val = string("valid")]; tensor input_41_strides_0 = const()[name = string("input_41_strides_0"), val = tensor([1, 1])]; tensor input_41_pad_0 = const()[name = string("input_41_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_41_dilations_0 = const()[name = string("input_41_dilations_0"), val = tensor([1, 1])]; int32 input_41_groups_0 = const()[name = string("input_41_groups_0"), val = int32(1)]; tensor stage2_5_reparam_conv_bias_to_fp16 = const()[name = string("stage2_5_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1071360)))]; fp16 quantize_10_scale_1 = const()[name = string("quantize_10_scale_1"), val = fp16(0x1.51cp-9)]; string quantize_10_output_dtype_1 = const()[name = string("quantize_10_output_dtype_1"), val = string("int8")]; tensor quantize_10 = quantize(input = input_39_cast_fp16, output_dtype = quantize_10_output_dtype_1, scale = quantize_10_scale_1)[name = string("quantize_10")]; fp16 dequantize_10_scale_1 = const()[name = string("dequantize_10_scale_1"), val = fp16(0x1.51cp-9)]; tensor dequantize_10 = dequantize(input = quantize_10, scale = dequantize_10_scale_1)[name = string("dequantize_10")]; tensor input_41_cast_fp16 = conv(bias = stage2_5_reparam_conv_bias_to_fp16, dilations = input_41_dilations_0, groups = input_41_groups_0, pad = input_41_pad_0, pad_type = input_41_pad_type_0, strides = input_41_strides_0, weight = stage2_5_reparam_conv_weight_cast_fp16, x = dequantize_10)[name = string("input_41_cast_fp16")]; tensor input_43_cast_fp16 = relu(x = input_41_cast_fp16)[name = string("input_43_cast_fp16")]; string input_45_pad_type_0 = const()[name = string("input_45_pad_type_0"), val = string("custom")]; tensor input_45_pad_0 = const()[name = string("input_45_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_45_groups_0 = const()[name = string("input_45_groups_0"), val = int32(128)]; tensor input_45_strides_0 = const()[name = string("input_45_strides_0"), val = tensor([1, 1])]; tensor input_45_dilations_0 = const()[name = string("input_45_dilations_0"), val = tensor([1, 1])]; tensor stage2_6_reparam_conv_bias_to_fp16 = const()[name = string("stage2_6_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1071680)))]; fp16 quantize_11_scale_1 = const()[name = string("quantize_11_scale_1"), val = fp16(0x1.884p-9)]; string quantize_11_output_dtype_1 = const()[name = string("quantize_11_output_dtype_1"), val = string("int8")]; tensor quantize_11 = quantize(input = input_43_cast_fp16, output_dtype = quantize_11_output_dtype_1, scale = quantize_11_scale_1)[name = string("quantize_11")]; fp16 dequantize_11_scale_1 = const()[name = string("dequantize_11_scale_1"), val = fp16(0x1.884p-9)]; tensor dequantize_11 = dequantize(input = quantize_11, scale = dequantize_11_scale_1)[name = string("dequantize_11")]; tensor input_45_cast_fp16 = conv(bias = stage2_6_reparam_conv_bias_to_fp16, dilations = input_45_dilations_0, groups = input_45_groups_0, pad = input_45_pad_0, pad_type = input_45_pad_type_0, strides = input_45_strides_0, weight = stage2_6_reparam_conv_weight_cast_fp16, x = dequantize_11)[name = string("input_45_cast_fp16")]; tensor input_47_cast_fp16 = relu(x = input_45_cast_fp16)[name = string("input_47_cast_fp16")]; string input_49_pad_type_0 = const()[name = string("input_49_pad_type_0"), val = string("valid")]; tensor input_49_strides_0 = const()[name = string("input_49_strides_0"), val = tensor([1, 1])]; tensor input_49_pad_0 = const()[name = string("input_49_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_49_dilations_0 = const()[name = string("input_49_dilations_0"), val = tensor([1, 1])]; int32 input_49_groups_0 = const()[name = string("input_49_groups_0"), val = int32(1)]; tensor stage2_7_reparam_conv_bias_to_fp16 = const()[name = string("stage2_7_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1072000)))]; fp16 quantize_12_scale_1 = const()[name = string("quantize_12_scale_1"), val = fp16(0x1.96cp-9)]; string quantize_12_output_dtype_1 = const()[name = string("quantize_12_output_dtype_1"), val = string("int8")]; tensor quantize_12 = quantize(input = input_47_cast_fp16, output_dtype = quantize_12_output_dtype_1, scale = quantize_12_scale_1)[name = string("quantize_12")]; fp16 dequantize_12_scale_1 = const()[name = string("dequantize_12_scale_1"), val = fp16(0x1.96cp-9)]; tensor dequantize_12 = dequantize(input = quantize_12, scale = dequantize_12_scale_1)[name = string("dequantize_12")]; tensor input_49_cast_fp16 = conv(bias = stage2_7_reparam_conv_bias_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = stage2_7_reparam_conv_weight_cast_fp16, x = dequantize_12)[name = string("input_49_cast_fp16")]; tensor input_51_cast_fp16 = relu(x = input_49_cast_fp16)[name = string("input_51_cast_fp16")]; string input_53_pad_type_0 = const()[name = string("input_53_pad_type_0"), val = string("custom")]; tensor input_53_pad_0 = const()[name = string("input_53_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_53_groups_0 = const()[name = string("input_53_groups_0"), val = int32(128)]; tensor input_53_strides_0 = const()[name = string("input_53_strides_0"), val = tensor([1, 1])]; tensor input_53_dilations_0 = const()[name = string("input_53_dilations_0"), val = tensor([1, 1])]; tensor stage2_8_reparam_conv_bias_to_fp16 = const()[name = string("stage2_8_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1072320)))]; fp16 quantize_13_scale_1 = const()[name = string("quantize_13_scale_1"), val = fp16(0x1.87cp-9)]; string quantize_13_output_dtype_1 = const()[name = string("quantize_13_output_dtype_1"), val = string("int8")]; tensor quantize_13 = quantize(input = input_51_cast_fp16, output_dtype = quantize_13_output_dtype_1, scale = quantize_13_scale_1)[name = string("quantize_13")]; fp16 dequantize_13_scale_1 = const()[name = string("dequantize_13_scale_1"), val = fp16(0x1.87cp-9)]; tensor dequantize_13 = dequantize(input = quantize_13, scale = dequantize_13_scale_1)[name = string("dequantize_13")]; tensor input_53_cast_fp16 = conv(bias = stage2_8_reparam_conv_bias_to_fp16, dilations = input_53_dilations_0, groups = input_53_groups_0, pad = input_53_pad_0, pad_type = input_53_pad_type_0, strides = input_53_strides_0, weight = stage2_8_reparam_conv_weight_cast_fp16, x = dequantize_13)[name = string("input_53_cast_fp16")]; tensor input_55_cast_fp16 = relu(x = input_53_cast_fp16)[name = string("input_55_cast_fp16")]; string input_57_pad_type_0 = const()[name = string("input_57_pad_type_0"), val = string("valid")]; tensor input_57_strides_0 = const()[name = string("input_57_strides_0"), val = tensor([1, 1])]; tensor input_57_pad_0 = const()[name = string("input_57_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_57_dilations_0 = const()[name = string("input_57_dilations_0"), val = tensor([1, 1])]; int32 input_57_groups_0 = const()[name = string("input_57_groups_0"), val = int32(1)]; tensor stage2_9_reparam_conv_bias_to_fp16 = const()[name = string("stage2_9_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1072640)))]; fp16 quantize_14_scale_1 = const()[name = string("quantize_14_scale_1"), val = fp16(0x1.318p-9)]; string quantize_14_output_dtype_1 = const()[name = string("quantize_14_output_dtype_1"), val = string("int8")]; tensor quantize_14 = quantize(input = input_55_cast_fp16, output_dtype = quantize_14_output_dtype_1, scale = quantize_14_scale_1)[name = string("quantize_14")]; fp16 dequantize_14_scale_1 = const()[name = string("dequantize_14_scale_1"), val = fp16(0x1.318p-9)]; tensor dequantize_14 = dequantize(input = quantize_14, scale = dequantize_14_scale_1)[name = string("dequantize_14")]; tensor input_57_cast_fp16 = conv(bias = stage2_9_reparam_conv_bias_to_fp16, dilations = input_57_dilations_0, groups = input_57_groups_0, pad = input_57_pad_0, pad_type = input_57_pad_type_0, strides = input_57_strides_0, weight = stage2_9_reparam_conv_weight_cast_fp16, x = dequantize_14)[name = string("input_57_cast_fp16")]; tensor input_59_cast_fp16 = relu(x = input_57_cast_fp16)[name = string("input_59_cast_fp16")]; string input_61_pad_type_0 = const()[name = string("input_61_pad_type_0"), val = string("custom")]; tensor input_61_pad_0 = const()[name = string("input_61_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_61_groups_0 = const()[name = string("input_61_groups_0"), val = int32(128)]; tensor input_61_strides_0 = const()[name = string("input_61_strides_0"), val = tensor([1, 1])]; tensor input_61_dilations_0 = const()[name = string("input_61_dilations_0"), val = tensor([1, 1])]; tensor stage2_10_reparam_conv_bias_to_fp16 = const()[name = string("stage2_10_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1072960)))]; fp16 quantize_15_scale_1 = const()[name = string("quantize_15_scale_1"), val = fp16(0x1.278p-9)]; string quantize_15_output_dtype_1 = const()[name = string("quantize_15_output_dtype_1"), val = string("int8")]; tensor quantize_15 = quantize(input = input_59_cast_fp16, output_dtype = quantize_15_output_dtype_1, scale = quantize_15_scale_1)[name = string("quantize_15")]; fp16 dequantize_15_scale_1 = const()[name = string("dequantize_15_scale_1"), val = fp16(0x1.278p-9)]; tensor dequantize_15 = dequantize(input = quantize_15, scale = dequantize_15_scale_1)[name = string("dequantize_15")]; tensor input_61_cast_fp16 = conv(bias = stage2_10_reparam_conv_bias_to_fp16, dilations = input_61_dilations_0, groups = input_61_groups_0, pad = input_61_pad_0, pad_type = input_61_pad_type_0, strides = input_61_strides_0, weight = stage2_10_reparam_conv_weight_cast_fp16, x = dequantize_15)[name = string("input_61_cast_fp16")]; tensor input_63_cast_fp16 = relu(x = input_61_cast_fp16)[name = string("input_63_cast_fp16")]; string input_65_pad_type_0 = const()[name = string("input_65_pad_type_0"), val = string("valid")]; tensor input_65_strides_0 = const()[name = string("input_65_strides_0"), val = tensor([1, 1])]; tensor input_65_pad_0 = const()[name = string("input_65_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_65_dilations_0 = const()[name = string("input_65_dilations_0"), val = tensor([1, 1])]; int32 input_65_groups_0 = const()[name = string("input_65_groups_0"), val = int32(1)]; tensor stage2_11_reparam_conv_bias_to_fp16 = const()[name = string("stage2_11_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1073280)))]; fp16 quantize_16_scale_1 = const()[name = string("quantize_16_scale_1"), val = fp16(0x1.7e8p-9)]; string quantize_16_output_dtype_1 = const()[name = string("quantize_16_output_dtype_1"), val = string("int8")]; tensor quantize_16 = quantize(input = input_63_cast_fp16, output_dtype = quantize_16_output_dtype_1, scale = quantize_16_scale_1)[name = string("quantize_16")]; fp16 dequantize_16_scale_1 = const()[name = string("dequantize_16_scale_1"), val = fp16(0x1.7e8p-9)]; tensor dequantize_16 = dequantize(input = quantize_16, scale = dequantize_16_scale_1)[name = string("dequantize_16")]; tensor input_65_cast_fp16 = conv(bias = stage2_11_reparam_conv_bias_to_fp16, dilations = input_65_dilations_0, groups = input_65_groups_0, pad = input_65_pad_0, pad_type = input_65_pad_type_0, strides = input_65_strides_0, weight = stage2_11_reparam_conv_weight_cast_fp16, x = dequantize_16)[name = string("input_65_cast_fp16")]; tensor input_67_cast_fp16 = relu(x = input_65_cast_fp16)[name = string("input_67_cast_fp16")]; string input_69_pad_type_0 = const()[name = string("input_69_pad_type_0"), val = string("custom")]; tensor input_69_pad_0 = const()[name = string("input_69_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_69_groups_0 = const()[name = string("input_69_groups_0"), val = int32(128)]; tensor input_69_strides_0 = const()[name = string("input_69_strides_0"), val = tensor([1, 1])]; tensor input_69_dilations_0 = const()[name = string("input_69_dilations_0"), val = tensor([1, 1])]; tensor stage2_12_reparam_conv_bias_to_fp16 = const()[name = string("stage2_12_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1073600)))]; fp16 quantize_17_scale_1 = const()[name = string("quantize_17_scale_1"), val = fp16(0x1.0f8p-8)]; string quantize_17_output_dtype_1 = const()[name = string("quantize_17_output_dtype_1"), val = string("int8")]; tensor quantize_17 = quantize(input = input_67_cast_fp16, output_dtype = quantize_17_output_dtype_1, scale = quantize_17_scale_1)[name = string("quantize_17")]; fp16 dequantize_17_scale_1 = const()[name = string("dequantize_17_scale_1"), val = fp16(0x1.0f8p-8)]; tensor dequantize_17 = dequantize(input = quantize_17, scale = dequantize_17_scale_1)[name = string("dequantize_17")]; tensor input_69_cast_fp16 = conv(bias = stage2_12_reparam_conv_bias_to_fp16, dilations = input_69_dilations_0, groups = input_69_groups_0, pad = input_69_pad_0, pad_type = input_69_pad_type_0, strides = input_69_strides_0, weight = stage2_12_reparam_conv_weight_cast_fp16, x = dequantize_17)[name = string("input_69_cast_fp16")]; tensor input_71_cast_fp16 = relu(x = input_69_cast_fp16)[name = string("input_71_cast_fp16")]; string input_73_pad_type_0 = const()[name = string("input_73_pad_type_0"), val = string("valid")]; tensor input_73_strides_0 = const()[name = string("input_73_strides_0"), val = tensor([1, 1])]; tensor input_73_pad_0 = const()[name = string("input_73_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_73_dilations_0 = const()[name = string("input_73_dilations_0"), val = tensor([1, 1])]; int32 input_73_groups_0 = const()[name = string("input_73_groups_0"), val = int32(1)]; tensor stage2_13_reparam_conv_bias_to_fp16 = const()[name = string("stage2_13_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1073920)))]; fp16 quantize_18_scale_1 = const()[name = string("quantize_18_scale_1"), val = fp16(0x1.4bcp-8)]; string quantize_18_output_dtype_1 = const()[name = string("quantize_18_output_dtype_1"), val = string("int8")]; tensor quantize_18 = quantize(input = input_71_cast_fp16, output_dtype = quantize_18_output_dtype_1, scale = quantize_18_scale_1)[name = string("quantize_18")]; fp16 dequantize_18_scale_1 = const()[name = string("dequantize_18_scale_1"), val = fp16(0x1.4bcp-8)]; tensor dequantize_18 = dequantize(input = quantize_18, scale = dequantize_18_scale_1)[name = string("dequantize_18")]; tensor input_73_cast_fp16 = conv(bias = stage2_13_reparam_conv_bias_to_fp16, dilations = input_73_dilations_0, groups = input_73_groups_0, pad = input_73_pad_0, pad_type = input_73_pad_type_0, strides = input_73_strides_0, weight = stage2_13_reparam_conv_weight_cast_fp16, x = dequantize_18)[name = string("input_73_cast_fp16")]; tensor input_75_cast_fp16 = relu(x = input_73_cast_fp16)[name = string("input_75_cast_fp16")]; string input_77_pad_type_0 = const()[name = string("input_77_pad_type_0"), val = string("custom")]; tensor input_77_pad_0 = const()[name = string("input_77_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_77_groups_0 = const()[name = string("input_77_groups_0"), val = int32(128)]; tensor input_77_strides_0 = const()[name = string("input_77_strides_0"), val = tensor([1, 1])]; tensor input_77_dilations_0 = const()[name = string("input_77_dilations_0"), val = tensor([1, 1])]; tensor stage2_14_reparam_conv_bias_to_fp16 = const()[name = string("stage2_14_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1074240)))]; fp16 quantize_19_scale_1 = const()[name = string("quantize_19_scale_1"), val = fp16(0x1.1cp-8)]; string quantize_19_output_dtype_1 = const()[name = string("quantize_19_output_dtype_1"), val = string("int8")]; tensor quantize_19 = quantize(input = input_75_cast_fp16, output_dtype = quantize_19_output_dtype_1, scale = quantize_19_scale_1)[name = string("quantize_19")]; fp16 dequantize_19_scale_1 = const()[name = string("dequantize_19_scale_1"), val = fp16(0x1.1cp-8)]; tensor dequantize_19 = dequantize(input = quantize_19, scale = dequantize_19_scale_1)[name = string("dequantize_19")]; tensor input_77_cast_fp16 = conv(bias = stage2_14_reparam_conv_bias_to_fp16, dilations = input_77_dilations_0, groups = input_77_groups_0, pad = input_77_pad_0, pad_type = input_77_pad_type_0, strides = input_77_strides_0, weight = stage2_14_reparam_conv_weight_cast_fp16, x = dequantize_19)[name = string("input_77_cast_fp16")]; tensor input_79_cast_fp16 = relu(x = input_77_cast_fp16)[name = string("input_79_cast_fp16")]; string input_81_pad_type_0 = const()[name = string("input_81_pad_type_0"), val = string("valid")]; tensor input_81_strides_0 = const()[name = string("input_81_strides_0"), val = tensor([1, 1])]; tensor input_81_pad_0 = const()[name = string("input_81_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_81_dilations_0 = const()[name = string("input_81_dilations_0"), val = tensor([1, 1])]; int32 input_81_groups_0 = const()[name = string("input_81_groups_0"), val = int32(1)]; tensor stage2_15_reparam_conv_bias_to_fp16 = const()[name = string("stage2_15_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1074560)))]; fp16 quantize_20_scale_1 = const()[name = string("quantize_20_scale_1"), val = fp16(0x1.0bcp-7)]; string quantize_20_output_dtype_1 = const()[name = string("quantize_20_output_dtype_1"), val = string("int8")]; tensor quantize_20 = quantize(input = input_79_cast_fp16, output_dtype = quantize_20_output_dtype_1, scale = quantize_20_scale_1)[name = string("quantize_20")]; fp16 dequantize_20_scale_1 = const()[name = string("dequantize_20_scale_1"), val = fp16(0x1.0bcp-7)]; tensor dequantize_20 = dequantize(input = quantize_20, scale = dequantize_20_scale_1)[name = string("dequantize_20")]; tensor input_81_cast_fp16 = conv(bias = stage2_15_reparam_conv_bias_to_fp16, dilations = input_81_dilations_0, groups = input_81_groups_0, pad = input_81_pad_0, pad_type = input_81_pad_type_0, strides = input_81_strides_0, weight = stage2_15_reparam_conv_weight_cast_fp16, x = dequantize_20)[name = string("input_81_cast_fp16")]; tensor input_83_cast_fp16 = relu(x = input_81_cast_fp16)[name = string("input_83_cast_fp16")]; string input_85_pad_type_0 = const()[name = string("input_85_pad_type_0"), val = string("custom")]; tensor input_85_pad_0 = const()[name = string("input_85_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_85_strides_0 = const()[name = string("input_85_strides_0"), val = tensor([2, 2])]; int32 input_85_groups_0 = const()[name = string("input_85_groups_0"), val = int32(128)]; tensor input_85_dilations_0 = const()[name = string("input_85_dilations_0"), val = tensor([1, 1])]; tensor stage3_0_reparam_conv_bias_to_fp16 = const()[name = string("stage3_0_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1074880)))]; fp16 quantize_21_scale_1 = const()[name = string("quantize_21_scale_1"), val = fp16(0x1.518p-8)]; string quantize_21_output_dtype_1 = const()[name = string("quantize_21_output_dtype_1"), val = string("int8")]; tensor quantize_21 = quantize(input = input_83_cast_fp16, output_dtype = quantize_21_output_dtype_1, scale = quantize_21_scale_1)[name = string("quantize_21")]; fp16 dequantize_21_scale_1 = const()[name = string("dequantize_21_scale_1"), val = fp16(0x1.518p-8)]; tensor dequantize_21 = dequantize(input = quantize_21, scale = dequantize_21_scale_1)[name = string("dequantize_21")]; tensor input_85_cast_fp16 = conv(bias = stage3_0_reparam_conv_bias_to_fp16, dilations = input_85_dilations_0, groups = input_85_groups_0, pad = input_85_pad_0, pad_type = input_85_pad_type_0, strides = input_85_strides_0, weight = stage3_0_reparam_conv_weight_cast_fp16, x = dequantize_21)[name = string("input_85_cast_fp16")]; tensor input_87_cast_fp16 = relu(x = input_85_cast_fp16)[name = string("input_87_cast_fp16")]; string input_89_pad_type_0 = const()[name = string("input_89_pad_type_0"), val = string("valid")]; tensor input_89_strides_0 = const()[name = string("input_89_strides_0"), val = tensor([1, 1])]; tensor input_89_pad_0 = const()[name = string("input_89_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_89_dilations_0 = const()[name = string("input_89_dilations_0"), val = tensor([1, 1])]; int32 input_89_groups_0 = const()[name = string("input_89_groups_0"), val = int32(1)]; tensor stage3_1_reparam_conv_bias_to_fp16 = const()[name = string("stage3_1_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1075200)))]; fp16 quantize_22_scale_1 = const()[name = string("quantize_22_scale_1"), val = fp16(0x1.bap-9)]; string quantize_22_output_dtype_1 = const()[name = string("quantize_22_output_dtype_1"), val = string("int8")]; tensor quantize_22 = quantize(input = input_87_cast_fp16, output_dtype = quantize_22_output_dtype_1, scale = quantize_22_scale_1)[name = string("quantize_22")]; fp16 dequantize_22_scale_1 = const()[name = string("dequantize_22_scale_1"), val = fp16(0x1.bap-9)]; tensor dequantize_22 = dequantize(input = quantize_22, scale = dequantize_22_scale_1)[name = string("dequantize_22")]; tensor input_89_cast_fp16 = conv(bias = stage3_1_reparam_conv_bias_to_fp16, dilations = input_89_dilations_0, groups = input_89_groups_0, pad = input_89_pad_0, pad_type = input_89_pad_type_0, strides = input_89_strides_0, weight = stage3_1_reparam_conv_weight_cast_fp16, x = dequantize_22)[name = string("input_89_cast_fp16")]; tensor input_91_cast_fp16 = relu(x = input_89_cast_fp16)[name = string("input_91_cast_fp16")]; string input_93_pad_type_0 = const()[name = string("input_93_pad_type_0"), val = string("custom")]; tensor input_93_pad_0 = const()[name = string("input_93_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_93_groups_0 = const()[name = string("input_93_groups_0"), val = int32(256)]; tensor input_93_strides_0 = const()[name = string("input_93_strides_0"), val = tensor([1, 1])]; tensor input_93_dilations_0 = const()[name = string("input_93_dilations_0"), val = tensor([1, 1])]; tensor stage3_2_reparam_conv_bias_to_fp16 = const()[name = string("stage3_2_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1075776)))]; fp16 quantize_23_scale_1 = const()[name = string("quantize_23_scale_1"), val = fp16(0x1.634p-9)]; string quantize_23_output_dtype_1 = const()[name = string("quantize_23_output_dtype_1"), val = string("int8")]; tensor quantize_23 = quantize(input = input_91_cast_fp16, output_dtype = quantize_23_output_dtype_1, scale = quantize_23_scale_1)[name = string("quantize_23")]; fp16 dequantize_23_scale_1 = const()[name = string("dequantize_23_scale_1"), val = fp16(0x1.634p-9)]; tensor dequantize_23 = dequantize(input = quantize_23, scale = dequantize_23_scale_1)[name = string("dequantize_23")]; tensor input_93_cast_fp16 = conv(bias = stage3_2_reparam_conv_bias_to_fp16, dilations = input_93_dilations_0, groups = input_93_groups_0, pad = input_93_pad_0, pad_type = input_93_pad_type_0, strides = input_93_strides_0, weight = stage3_2_reparam_conv_weight_cast_fp16, x = dequantize_23)[name = string("input_93_cast_fp16")]; tensor input_95_cast_fp16 = relu(x = input_93_cast_fp16)[name = string("input_95_cast_fp16")]; string input_97_pad_type_0 = const()[name = string("input_97_pad_type_0"), val = string("valid")]; tensor input_97_strides_0 = const()[name = string("input_97_strides_0"), val = tensor([1, 1])]; tensor input_97_pad_0 = const()[name = string("input_97_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_97_dilations_0 = const()[name = string("input_97_dilations_0"), val = tensor([1, 1])]; int32 input_97_groups_0 = const()[name = string("input_97_groups_0"), val = int32(1)]; tensor stage3_3_reparam_conv_bias_to_fp16 = const()[name = string("stage3_3_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1076352)))]; fp16 quantize_24_scale_1 = const()[name = string("quantize_24_scale_1"), val = fp16(0x1.4ep-9)]; string quantize_24_output_dtype_1 = const()[name = string("quantize_24_output_dtype_1"), val = string("int8")]; tensor quantize_24 = quantize(input = input_95_cast_fp16, output_dtype = quantize_24_output_dtype_1, scale = quantize_24_scale_1)[name = string("quantize_24")]; fp16 dequantize_24_scale_1 = const()[name = string("dequantize_24_scale_1"), val = fp16(0x1.4ep-9)]; tensor dequantize_24 = dequantize(input = quantize_24, scale = dequantize_24_scale_1)[name = string("dequantize_24")]; tensor input_97_cast_fp16 = conv(bias = stage3_3_reparam_conv_bias_to_fp16, dilations = input_97_dilations_0, groups = input_97_groups_0, pad = input_97_pad_0, pad_type = input_97_pad_type_0, strides = input_97_strides_0, weight = stage3_3_reparam_conv_weight_cast_fp16, x = dequantize_24)[name = string("input_97_cast_fp16")]; tensor input_99_cast_fp16 = relu(x = input_97_cast_fp16)[name = string("input_99_cast_fp16")]; string input_101_pad_type_0 = const()[name = string("input_101_pad_type_0"), val = string("custom")]; tensor input_101_pad_0 = const()[name = string("input_101_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_101_groups_0 = const()[name = string("input_101_groups_0"), val = int32(256)]; tensor input_101_strides_0 = const()[name = string("input_101_strides_0"), val = tensor([1, 1])]; tensor input_101_dilations_0 = const()[name = string("input_101_dilations_0"), val = tensor([1, 1])]; tensor stage3_4_reparam_conv_bias_to_fp16 = const()[name = string("stage3_4_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1076928)))]; fp16 quantize_25_scale_1 = const()[name = string("quantize_25_scale_1"), val = fp16(0x1.cbcp-9)]; string quantize_25_output_dtype_1 = const()[name = string("quantize_25_output_dtype_1"), val = string("int8")]; tensor quantize_25 = quantize(input = input_99_cast_fp16, output_dtype = quantize_25_output_dtype_1, scale = quantize_25_scale_1)[name = string("quantize_25")]; fp16 dequantize_25_scale_1 = const()[name = string("dequantize_25_scale_1"), val = fp16(0x1.cbcp-9)]; tensor dequantize_25 = dequantize(input = quantize_25, scale = dequantize_25_scale_1)[name = string("dequantize_25")]; tensor input_101_cast_fp16 = conv(bias = stage3_4_reparam_conv_bias_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = stage3_4_reparam_conv_weight_cast_fp16, x = dequantize_25)[name = string("input_101_cast_fp16")]; tensor input_103_cast_fp16 = relu(x = input_101_cast_fp16)[name = string("input_103_cast_fp16")]; string input_105_pad_type_0 = const()[name = string("input_105_pad_type_0"), val = string("valid")]; tensor input_105_strides_0 = const()[name = string("input_105_strides_0"), val = tensor([1, 1])]; tensor input_105_pad_0 = const()[name = string("input_105_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_105_dilations_0 = const()[name = string("input_105_dilations_0"), val = tensor([1, 1])]; int32 input_105_groups_0 = const()[name = string("input_105_groups_0"), val = int32(1)]; tensor stage3_5_reparam_conv_bias_to_fp16 = const()[name = string("stage3_5_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1077504)))]; fp16 quantize_26_scale_1 = const()[name = string("quantize_26_scale_1"), val = fp16(0x1.dep-9)]; string quantize_26_output_dtype_1 = const()[name = string("quantize_26_output_dtype_1"), val = string("int8")]; tensor quantize_26 = quantize(input = input_103_cast_fp16, output_dtype = quantize_26_output_dtype_1, scale = quantize_26_scale_1)[name = string("quantize_26")]; fp16 dequantize_26_scale_1 = const()[name = string("dequantize_26_scale_1"), val = fp16(0x1.dep-9)]; tensor dequantize_26 = dequantize(input = quantize_26, scale = dequantize_26_scale_1)[name = string("dequantize_26")]; tensor input_105_cast_fp16 = conv(bias = stage3_5_reparam_conv_bias_to_fp16, dilations = input_105_dilations_0, groups = input_105_groups_0, pad = input_105_pad_0, pad_type = input_105_pad_type_0, strides = input_105_strides_0, weight = stage3_5_reparam_conv_weight_cast_fp16, x = dequantize_26)[name = string("input_105_cast_fp16")]; tensor input_107_cast_fp16 = relu(x = input_105_cast_fp16)[name = string("input_107_cast_fp16")]; string input_109_pad_type_0 = const()[name = string("input_109_pad_type_0"), val = string("custom")]; tensor input_109_pad_0 = const()[name = string("input_109_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_109_groups_0 = const()[name = string("input_109_groups_0"), val = int32(256)]; tensor input_109_strides_0 = const()[name = string("input_109_strides_0"), val = tensor([1, 1])]; tensor input_109_dilations_0 = const()[name = string("input_109_dilations_0"), val = tensor([1, 1])]; tensor stage3_6_reparam_conv_bias_to_fp16 = const()[name = string("stage3_6_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1078080)))]; fp16 quantize_27_scale_1 = const()[name = string("quantize_27_scale_1"), val = fp16(0x1.59p-9)]; string quantize_27_output_dtype_1 = const()[name = string("quantize_27_output_dtype_1"), val = string("int8")]; tensor quantize_27 = quantize(input = input_107_cast_fp16, output_dtype = quantize_27_output_dtype_1, scale = quantize_27_scale_1)[name = string("quantize_27")]; fp16 dequantize_27_scale_1 = const()[name = string("dequantize_27_scale_1"), val = fp16(0x1.59p-9)]; tensor dequantize_27 = dequantize(input = quantize_27, scale = dequantize_27_scale_1)[name = string("dequantize_27")]; tensor input_109_cast_fp16 = conv(bias = stage3_6_reparam_conv_bias_to_fp16, dilations = input_109_dilations_0, groups = input_109_groups_0, pad = input_109_pad_0, pad_type = input_109_pad_type_0, strides = input_109_strides_0, weight = stage3_6_reparam_conv_weight_cast_fp16, x = dequantize_27)[name = string("input_109_cast_fp16")]; tensor input_111_cast_fp16 = relu(x = input_109_cast_fp16)[name = string("input_111_cast_fp16")]; string input_113_pad_type_0 = const()[name = string("input_113_pad_type_0"), val = string("valid")]; tensor input_113_strides_0 = const()[name = string("input_113_strides_0"), val = tensor([1, 1])]; tensor input_113_pad_0 = const()[name = string("input_113_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_113_dilations_0 = const()[name = string("input_113_dilations_0"), val = tensor([1, 1])]; int32 input_113_groups_0 = const()[name = string("input_113_groups_0"), val = int32(1)]; tensor stage3_7_reparam_conv_bias_to_fp16 = const()[name = string("stage3_7_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1078656)))]; fp16 quantize_28_scale_1 = const()[name = string("quantize_28_scale_1"), val = fp16(0x1.03p-8)]; string quantize_28_output_dtype_1 = const()[name = string("quantize_28_output_dtype_1"), val = string("int8")]; tensor quantize_28 = quantize(input = input_111_cast_fp16, output_dtype = quantize_28_output_dtype_1, scale = quantize_28_scale_1)[name = string("quantize_28")]; fp16 dequantize_28_scale_1 = const()[name = string("dequantize_28_scale_1"), val = fp16(0x1.03p-8)]; tensor dequantize_28 = dequantize(input = quantize_28, scale = dequantize_28_scale_1)[name = string("dequantize_28")]; tensor input_113_cast_fp16 = conv(bias = stage3_7_reparam_conv_bias_to_fp16, dilations = input_113_dilations_0, groups = input_113_groups_0, pad = input_113_pad_0, pad_type = input_113_pad_type_0, strides = input_113_strides_0, weight = stage3_7_reparam_conv_weight_cast_fp16, x = dequantize_28)[name = string("input_113_cast_fp16")]; tensor input_115_cast_fp16 = relu(x = input_113_cast_fp16)[name = string("input_115_cast_fp16")]; string input_117_pad_type_0 = const()[name = string("input_117_pad_type_0"), val = string("custom")]; tensor input_117_pad_0 = const()[name = string("input_117_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_117_groups_0 = const()[name = string("input_117_groups_0"), val = int32(256)]; tensor input_117_strides_0 = const()[name = string("input_117_strides_0"), val = tensor([1, 1])]; tensor input_117_dilations_0 = const()[name = string("input_117_dilations_0"), val = tensor([1, 1])]; tensor stage3_8_reparam_conv_bias_to_fp16 = const()[name = string("stage3_8_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1079232)))]; fp16 quantize_29_scale_1 = const()[name = string("quantize_29_scale_1"), val = fp16(0x1.32p-9)]; string quantize_29_output_dtype_1 = const()[name = string("quantize_29_output_dtype_1"), val = string("int8")]; tensor quantize_29 = quantize(input = input_115_cast_fp16, output_dtype = quantize_29_output_dtype_1, scale = quantize_29_scale_1)[name = string("quantize_29")]; fp16 dequantize_29_scale_1 = const()[name = string("dequantize_29_scale_1"), val = fp16(0x1.32p-9)]; tensor dequantize_29 = dequantize(input = quantize_29, scale = dequantize_29_scale_1)[name = string("dequantize_29")]; tensor input_117_cast_fp16 = conv(bias = stage3_8_reparam_conv_bias_to_fp16, dilations = input_117_dilations_0, groups = input_117_groups_0, pad = input_117_pad_0, pad_type = input_117_pad_type_0, strides = input_117_strides_0, weight = stage3_8_reparam_conv_weight_cast_fp16, x = dequantize_29)[name = string("input_117_cast_fp16")]; tensor input_119_cast_fp16 = relu(x = input_117_cast_fp16)[name = string("input_119_cast_fp16")]; string input_121_pad_type_0 = const()[name = string("input_121_pad_type_0"), val = string("valid")]; tensor input_121_strides_0 = const()[name = string("input_121_strides_0"), val = tensor([1, 1])]; tensor input_121_pad_0 = const()[name = string("input_121_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_121_dilations_0 = const()[name = string("input_121_dilations_0"), val = tensor([1, 1])]; int32 input_121_groups_0 = const()[name = string("input_121_groups_0"), val = int32(1)]; tensor stage3_9_reparam_conv_bias_to_fp16 = const()[name = string("stage3_9_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1079808)))]; fp16 quantize_30_scale_1 = const()[name = string("quantize_30_scale_1"), val = fp16(0x1.054p-7)]; string quantize_30_output_dtype_1 = const()[name = string("quantize_30_output_dtype_1"), val = string("int8")]; tensor quantize_30 = quantize(input = input_119_cast_fp16, output_dtype = quantize_30_output_dtype_1, scale = quantize_30_scale_1)[name = string("quantize_30")]; fp16 dequantize_30_scale_1 = const()[name = string("dequantize_30_scale_1"), val = fp16(0x1.054p-7)]; tensor dequantize_30 = dequantize(input = quantize_30, scale = dequantize_30_scale_1)[name = string("dequantize_30")]; tensor input_121_cast_fp16 = conv(bias = stage3_9_reparam_conv_bias_to_fp16, dilations = input_121_dilations_0, groups = input_121_groups_0, pad = input_121_pad_0, pad_type = input_121_pad_type_0, strides = input_121_strides_0, weight = stage3_9_reparam_conv_weight_cast_fp16, x = dequantize_30)[name = string("input_121_cast_fp16")]; tensor input_123_cast_fp16 = relu(x = input_121_cast_fp16)[name = string("input_123_cast_fp16")]; string input_125_pad_type_0 = const()[name = string("input_125_pad_type_0"), val = string("custom")]; tensor input_125_pad_0 = const()[name = string("input_125_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_125_groups_0 = const()[name = string("input_125_groups_0"), val = int32(256)]; tensor input_125_strides_0 = const()[name = string("input_125_strides_0"), val = tensor([1, 1])]; tensor input_125_dilations_0 = const()[name = string("input_125_dilations_0"), val = tensor([1, 1])]; tensor stage3_10_reparam_conv_bias_to_fp16 = const()[name = string("stage3_10_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1080384)))]; fp16 quantize_31_scale_1 = const()[name = string("quantize_31_scale_1"), val = fp16(0x1.23p-8)]; string quantize_31_output_dtype_1 = const()[name = string("quantize_31_output_dtype_1"), val = string("int8")]; tensor quantize_31 = quantize(input = input_123_cast_fp16, output_dtype = quantize_31_output_dtype_1, scale = quantize_31_scale_1)[name = string("quantize_31")]; fp16 dequantize_31_scale_1 = const()[name = string("dequantize_31_scale_1"), val = fp16(0x1.23p-8)]; tensor dequantize_31 = dequantize(input = quantize_31, scale = dequantize_31_scale_1)[name = string("dequantize_31")]; tensor input_125_cast_fp16 = conv(bias = stage3_10_reparam_conv_bias_to_fp16, dilations = input_125_dilations_0, groups = input_125_groups_0, pad = input_125_pad_0, pad_type = input_125_pad_type_0, strides = input_125_strides_0, weight = stage3_10_reparam_conv_weight_cast_fp16, x = dequantize_31)[name = string("input_125_cast_fp16")]; tensor input_127_cast_fp16 = relu(x = input_125_cast_fp16)[name = string("input_127_cast_fp16")]; string input_129_pad_type_0 = const()[name = string("input_129_pad_type_0"), val = string("valid")]; tensor input_129_strides_0 = const()[name = string("input_129_strides_0"), val = tensor([1, 1])]; tensor input_129_pad_0 = const()[name = string("input_129_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_129_dilations_0 = const()[name = string("input_129_dilations_0"), val = tensor([1, 1])]; int32 input_129_groups_0 = const()[name = string("input_129_groups_0"), val = int32(1)]; tensor stage3_11_reparam_conv_bias_to_fp16 = const()[name = string("stage3_11_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1080960)))]; fp16 quantize_32_scale_1 = const()[name = string("quantize_32_scale_1"), val = fp16(0x1.c9cp-9)]; string quantize_32_output_dtype_1 = const()[name = string("quantize_32_output_dtype_1"), val = string("int8")]; tensor quantize_32 = quantize(input = input_127_cast_fp16, output_dtype = quantize_32_output_dtype_1, scale = quantize_32_scale_1)[name = string("quantize_32")]; fp16 dequantize_32_scale_1 = const()[name = string("dequantize_32_scale_1"), val = fp16(0x1.c9cp-9)]; tensor dequantize_32 = dequantize(input = quantize_32, scale = dequantize_32_scale_1)[name = string("dequantize_32")]; tensor input_129_cast_fp16 = conv(bias = stage3_11_reparam_conv_bias_to_fp16, dilations = input_129_dilations_0, groups = input_129_groups_0, pad = input_129_pad_0, pad_type = input_129_pad_type_0, strides = input_129_strides_0, weight = stage3_11_reparam_conv_weight_cast_fp16, x = dequantize_32)[name = string("input_129_cast_fp16")]; tensor input_131_cast_fp16 = relu(x = input_129_cast_fp16)[name = string("input_131_cast_fp16")]; string input_133_pad_type_0 = const()[name = string("input_133_pad_type_0"), val = string("custom")]; tensor input_133_pad_0 = const()[name = string("input_133_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_133_groups_0 = const()[name = string("input_133_groups_0"), val = int32(256)]; tensor input_133_strides_0 = const()[name = string("input_133_strides_0"), val = tensor([1, 1])]; tensor input_133_dilations_0 = const()[name = string("input_133_dilations_0"), val = tensor([1, 1])]; tensor stage3_12_reparam_conv_bias_to_fp16 = const()[name = string("stage3_12_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1081536)))]; fp16 quantize_33_scale_1 = const()[name = string("quantize_33_scale_1"), val = fp16(0x1.a94p-9)]; string quantize_33_output_dtype_1 = const()[name = string("quantize_33_output_dtype_1"), val = string("int8")]; tensor quantize_33 = quantize(input = input_131_cast_fp16, output_dtype = quantize_33_output_dtype_1, scale = quantize_33_scale_1)[name = string("quantize_33")]; fp16 dequantize_33_scale_1 = const()[name = string("dequantize_33_scale_1"), val = fp16(0x1.a94p-9)]; tensor dequantize_33 = dequantize(input = quantize_33, scale = dequantize_33_scale_1)[name = string("dequantize_33")]; tensor input_133_cast_fp16 = conv(bias = stage3_12_reparam_conv_bias_to_fp16, dilations = input_133_dilations_0, groups = input_133_groups_0, pad = input_133_pad_0, pad_type = input_133_pad_type_0, strides = input_133_strides_0, weight = stage3_12_reparam_conv_weight_cast_fp16, x = dequantize_33)[name = string("input_133_cast_fp16")]; tensor input_135_cast_fp16 = relu(x = input_133_cast_fp16)[name = string("input_135_cast_fp16")]; string input_137_pad_type_0 = const()[name = string("input_137_pad_type_0"), val = string("valid")]; tensor input_137_strides_0 = const()[name = string("input_137_strides_0"), val = tensor([1, 1])]; tensor input_137_pad_0 = const()[name = string("input_137_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_137_dilations_0 = const()[name = string("input_137_dilations_0"), val = tensor([1, 1])]; int32 input_137_groups_0 = const()[name = string("input_137_groups_0"), val = int32(1)]; tensor stage3_13_reparam_conv_bias_to_fp16 = const()[name = string("stage3_13_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1082112)))]; fp16 quantize_34_scale_1 = const()[name = string("quantize_34_scale_1"), val = fp16(0x1.a04p-9)]; string quantize_34_output_dtype_1 = const()[name = string("quantize_34_output_dtype_1"), val = string("int8")]; tensor quantize_34 = quantize(input = input_135_cast_fp16, output_dtype = quantize_34_output_dtype_1, scale = quantize_34_scale_1)[name = string("quantize_34")]; fp16 dequantize_34_scale_1 = const()[name = string("dequantize_34_scale_1"), val = fp16(0x1.a04p-9)]; tensor dequantize_34 = dequantize(input = quantize_34, scale = dequantize_34_scale_1)[name = string("dequantize_34")]; tensor input_137_cast_fp16 = conv(bias = stage3_13_reparam_conv_bias_to_fp16, dilations = input_137_dilations_0, groups = input_137_groups_0, pad = input_137_pad_0, pad_type = input_137_pad_type_0, strides = input_137_strides_0, weight = stage3_13_reparam_conv_weight_cast_fp16, x = dequantize_34)[name = string("input_137_cast_fp16")]; tensor input_139_cast_fp16 = relu(x = input_137_cast_fp16)[name = string("input_139_cast_fp16")]; string input_141_pad_type_0 = const()[name = string("input_141_pad_type_0"), val = string("custom")]; tensor input_141_pad_0 = const()[name = string("input_141_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_141_groups_0 = const()[name = string("input_141_groups_0"), val = int32(256)]; tensor input_141_strides_0 = const()[name = string("input_141_strides_0"), val = tensor([1, 1])]; tensor input_141_dilations_0 = const()[name = string("input_141_dilations_0"), val = tensor([1, 1])]; tensor stage3_14_reparam_conv_bias_to_fp16 = const()[name = string("stage3_14_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1082688)))]; fp16 quantize_35_scale_1 = const()[name = string("quantize_35_scale_1"), val = fp16(0x1.698p-9)]; string quantize_35_output_dtype_1 = const()[name = string("quantize_35_output_dtype_1"), val = string("int8")]; tensor quantize_35 = quantize(input = input_139_cast_fp16, output_dtype = quantize_35_output_dtype_1, scale = quantize_35_scale_1)[name = string("quantize_35")]; fp16 dequantize_35_scale_1 = const()[name = string("dequantize_35_scale_1"), val = fp16(0x1.698p-9)]; tensor dequantize_35 = dequantize(input = quantize_35, scale = dequantize_35_scale_1)[name = string("dequantize_35")]; tensor input_141_cast_fp16 = conv(bias = stage3_14_reparam_conv_bias_to_fp16, dilations = input_141_dilations_0, groups = input_141_groups_0, pad = input_141_pad_0, pad_type = input_141_pad_type_0, strides = input_141_strides_0, weight = stage3_14_reparam_conv_weight_cast_fp16, x = dequantize_35)[name = string("input_141_cast_fp16")]; tensor input_143_cast_fp16 = relu(x = input_141_cast_fp16)[name = string("input_143_cast_fp16")]; string input_145_pad_type_0 = const()[name = string("input_145_pad_type_0"), val = string("valid")]; tensor input_145_strides_0 = const()[name = string("input_145_strides_0"), val = tensor([1, 1])]; tensor input_145_pad_0 = const()[name = string("input_145_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_145_dilations_0 = const()[name = string("input_145_dilations_0"), val = tensor([1, 1])]; int32 input_145_groups_0 = const()[name = string("input_145_groups_0"), val = int32(1)]; tensor stage3_15_reparam_conv_bias_to_fp16 = const()[name = string("stage3_15_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1083264)))]; fp16 quantize_36_scale_1 = const()[name = string("quantize_36_scale_1"), val = fp16(0x1.6ecp-9)]; string quantize_36_output_dtype_1 = const()[name = string("quantize_36_output_dtype_1"), val = string("int8")]; tensor quantize_36 = quantize(input = input_143_cast_fp16, output_dtype = quantize_36_output_dtype_1, scale = quantize_36_scale_1)[name = string("quantize_36")]; fp16 dequantize_36_scale_1 = const()[name = string("dequantize_36_scale_1"), val = fp16(0x1.6ecp-9)]; tensor dequantize_36 = dequantize(input = quantize_36, scale = dequantize_36_scale_1)[name = string("dequantize_36")]; tensor input_145_cast_fp16 = conv(bias = stage3_15_reparam_conv_bias_to_fp16, dilations = input_145_dilations_0, groups = input_145_groups_0, pad = input_145_pad_0, pad_type = input_145_pad_type_0, strides = input_145_strides_0, weight = stage3_15_reparam_conv_weight_cast_fp16, x = dequantize_36)[name = string("input_145_cast_fp16")]; tensor input_147_cast_fp16 = relu(x = input_145_cast_fp16)[name = string("input_147_cast_fp16")]; string input_149_pad_type_0 = const()[name = string("input_149_pad_type_0"), val = string("custom")]; tensor input_149_pad_0 = const()[name = string("input_149_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_149_groups_0 = const()[name = string("input_149_groups_0"), val = int32(256)]; tensor input_149_strides_0 = const()[name = string("input_149_strides_0"), val = tensor([1, 1])]; tensor input_149_dilations_0 = const()[name = string("input_149_dilations_0"), val = tensor([1, 1])]; tensor stage3_16_reparam_conv_bias_to_fp16 = const()[name = string("stage3_16_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1083840)))]; fp16 quantize_37_scale_1 = const()[name = string("quantize_37_scale_1"), val = fp16(0x1.65p-9)]; string quantize_37_output_dtype_1 = const()[name = string("quantize_37_output_dtype_1"), val = string("int8")]; tensor quantize_37 = quantize(input = input_147_cast_fp16, output_dtype = quantize_37_output_dtype_1, scale = quantize_37_scale_1)[name = string("quantize_37")]; fp16 dequantize_37_scale_1 = const()[name = string("dequantize_37_scale_1"), val = fp16(0x1.65p-9)]; tensor dequantize_37 = dequantize(input = quantize_37, scale = dequantize_37_scale_1)[name = string("dequantize_37")]; tensor input_149_cast_fp16 = conv(bias = stage3_16_reparam_conv_bias_to_fp16, dilations = input_149_dilations_0, groups = input_149_groups_0, pad = input_149_pad_0, pad_type = input_149_pad_type_0, strides = input_149_strides_0, weight = stage3_16_reparam_conv_weight_cast_fp16, x = dequantize_37)[name = string("input_149_cast_fp16")]; tensor input_151_cast_fp16 = relu(x = input_149_cast_fp16)[name = string("input_151_cast_fp16")]; string input_153_pad_type_0 = const()[name = string("input_153_pad_type_0"), val = string("valid")]; tensor input_153_strides_0 = const()[name = string("input_153_strides_0"), val = tensor([1, 1])]; tensor input_153_pad_0 = const()[name = string("input_153_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_153_dilations_0 = const()[name = string("input_153_dilations_0"), val = tensor([1, 1])]; int32 input_153_groups_0 = const()[name = string("input_153_groups_0"), val = int32(1)]; tensor stage3_17_reparam_conv_bias_to_fp16 = const()[name = string("stage3_17_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1084416)))]; fp16 quantize_38_scale_1 = const()[name = string("quantize_38_scale_1"), val = fp16(0x1.a9cp-9)]; string quantize_38_output_dtype_1 = const()[name = string("quantize_38_output_dtype_1"), val = string("int8")]; tensor quantize_38 = quantize(input = input_151_cast_fp16, output_dtype = quantize_38_output_dtype_1, scale = quantize_38_scale_1)[name = string("quantize_38")]; fp16 dequantize_38_scale_1 = const()[name = string("dequantize_38_scale_1"), val = fp16(0x1.a9cp-9)]; tensor dequantize_38 = dequantize(input = quantize_38, scale = dequantize_38_scale_1)[name = string("dequantize_38")]; tensor input_153_cast_fp16 = conv(bias = stage3_17_reparam_conv_bias_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = stage3_17_reparam_conv_weight_cast_fp16, x = dequantize_38)[name = string("input_153_cast_fp16")]; tensor input_155_cast_fp16 = relu(x = input_153_cast_fp16)[name = string("input_155_cast_fp16")]; string input_157_pad_type_0 = const()[name = string("input_157_pad_type_0"), val = string("custom")]; tensor input_157_pad_0 = const()[name = string("input_157_pad_0"), val = tensor([1, 1, 1, 1])]; int32 input_157_groups_0 = const()[name = string("input_157_groups_0"), val = int32(256)]; tensor input_157_strides_0 = const()[name = string("input_157_strides_0"), val = tensor([1, 1])]; tensor input_157_dilations_0 = const()[name = string("input_157_dilations_0"), val = tensor([1, 1])]; tensor stage3_18_reparam_conv_bias_to_fp16 = const()[name = string("stage3_18_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1084992)))]; fp16 quantize_39_scale_1 = const()[name = string("quantize_39_scale_1"), val = fp16(0x1.b34p-8)]; string quantize_39_output_dtype_1 = const()[name = string("quantize_39_output_dtype_1"), val = string("int8")]; tensor quantize_39 = quantize(input = input_155_cast_fp16, output_dtype = quantize_39_output_dtype_1, scale = quantize_39_scale_1)[name = string("quantize_39")]; fp16 dequantize_39_scale_1 = const()[name = string("dequantize_39_scale_1"), val = fp16(0x1.b34p-8)]; tensor dequantize_39 = dequantize(input = quantize_39, scale = dequantize_39_scale_1)[name = string("dequantize_39")]; tensor input_157_cast_fp16 = conv(bias = stage3_18_reparam_conv_bias_to_fp16, dilations = input_157_dilations_0, groups = input_157_groups_0, pad = input_157_pad_0, pad_type = input_157_pad_type_0, strides = input_157_strides_0, weight = stage3_18_reparam_conv_weight_cast_fp16, x = dequantize_39)[name = string("input_157_cast_fp16")]; tensor input_159_cast_fp16 = relu(x = input_157_cast_fp16)[name = string("input_159_cast_fp16")]; string input_161_pad_type_0 = const()[name = string("input_161_pad_type_0"), val = string("valid")]; tensor input_161_strides_0 = const()[name = string("input_161_strides_0"), val = tensor([1, 1])]; tensor input_161_pad_0 = const()[name = string("input_161_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_161_dilations_0 = const()[name = string("input_161_dilations_0"), val = tensor([1, 1])]; int32 input_161_groups_0 = const()[name = string("input_161_groups_0"), val = int32(1)]; tensor stage3_19_reparam_conv_bias_to_fp16 = const()[name = string("stage3_19_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1085568)))]; fp16 quantize_40_scale_1 = const()[name = string("quantize_40_scale_1"), val = fp16(0x1.91cp-9)]; string quantize_40_output_dtype_1 = const()[name = string("quantize_40_output_dtype_1"), val = string("int8")]; tensor quantize_40 = quantize(input = input_159_cast_fp16, output_dtype = quantize_40_output_dtype_1, scale = quantize_40_scale_1)[name = string("quantize_40")]; fp16 dequantize_40_scale_1 = const()[name = string("dequantize_40_scale_1"), val = fp16(0x1.91cp-9)]; tensor dequantize_40 = dequantize(input = quantize_40, scale = dequantize_40_scale_1)[name = string("dequantize_40")]; tensor input_161_cast_fp16 = conv(bias = stage3_19_reparam_conv_bias_to_fp16, dilations = input_161_dilations_0, groups = input_161_groups_0, pad = input_161_pad_0, pad_type = input_161_pad_type_0, strides = input_161_strides_0, weight = stage3_19_reparam_conv_weight_cast_fp16, x = dequantize_40)[name = string("input_161_cast_fp16")]; tensor input_163_cast_fp16 = relu(x = input_161_cast_fp16)[name = string("input_163_cast_fp16")]; string input_165_pad_type_0 = const()[name = string("input_165_pad_type_0"), val = string("custom")]; tensor input_165_pad_0 = const()[name = string("input_165_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_165_strides_0 = const()[name = string("input_165_strides_0"), val = tensor([2, 2])]; int32 input_165_groups_0 = const()[name = string("input_165_groups_0"), val = int32(256)]; tensor input_165_dilations_0 = const()[name = string("input_165_dilations_0"), val = tensor([1, 1])]; tensor stage4_0_reparam_conv_bias_to_fp16 = const()[name = string("stage4_0_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1086144)))]; fp16 quantize_41_scale_1 = const()[name = string("quantize_41_scale_1"), val = fp16(0x1.06cp-8)]; string quantize_41_output_dtype_1 = const()[name = string("quantize_41_output_dtype_1"), val = string("int8")]; tensor quantize_41 = quantize(input = input_163_cast_fp16, output_dtype = quantize_41_output_dtype_1, scale = quantize_41_scale_1)[name = string("quantize_41")]; fp16 dequantize_41_scale_1 = const()[name = string("dequantize_41_scale_1"), val = fp16(0x1.06cp-8)]; tensor dequantize_41 = dequantize(input = quantize_41, scale = dequantize_41_scale_1)[name = string("dequantize_41")]; tensor input_165_cast_fp16 = conv(bias = stage4_0_reparam_conv_bias_to_fp16, dilations = input_165_dilations_0, groups = input_165_groups_0, pad = input_165_pad_0, pad_type = input_165_pad_type_0, strides = input_165_strides_0, weight = stage4_0_reparam_conv_weight_cast_fp16, x = dequantize_41)[name = string("input_165_cast_fp16")]; tensor input_167_cast_fp16 = relu(x = input_165_cast_fp16)[name = string("input_167_cast_fp16")]; string input_169_pad_type_0 = const()[name = string("input_169_pad_type_0"), val = string("valid")]; tensor input_169_strides_0 = const()[name = string("input_169_strides_0"), val = tensor([1, 1])]; tensor input_169_pad_0 = const()[name = string("input_169_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_169_dilations_0 = const()[name = string("input_169_dilations_0"), val = tensor([1, 1])]; int32 input_169_groups_0 = const()[name = string("input_169_groups_0"), val = int32(1)]; tensor stage4_1_reparam_conv_bias_to_fp16 = const()[name = string("stage4_1_reparam_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = string("@model_path/weights/weight.weights_nonane"), offset = uint64(1086720)))]; fp16 quantize_42_scale_1 = const()[name = string("quantize_42_scale_1"), val = fp16(0x1.014p-7)]; string quantize_42_output_dtype_1 = const()[name = string("quantize_42_output_dtype_1"), val = string("int8")]; tensor quantize_42 = quantize(input = input_167_cast_fp16, output_dtype = quantize_42_output_dtype_1, scale = quantize_42_scale_1)[name = string("quantize_42")]; fp16 dequantize_42_scale_1 = const()[name = string("dequantize_42_scale_1"), val = fp16(0x1.014p-7)]; tensor dequantize_42 = dequantize(input = quantize_42, scale = dequantize_42_scale_1)[name = string("dequantize_42")]; tensor input_169_cast_fp16 = conv(bias = stage4_1_reparam_conv_bias_to_fp16, dilations = input_169_dilations_0, groups = input_169_groups_0, pad = input_169_pad_0, pad_type = input_169_pad_type_0, strides = input_169_strides_0, weight = stage4_1_reparam_conv_weight_cast_fp16, x = dequantize_42)[name = string("input_169_cast_fp16")]; tensor input_171_cast_fp16 = relu(x = input_169_cast_fp16)[name = string("input_171_cast_fp16")]; fp16 quantize_43_scale_1 = const()[name = string("quantize_43_scale_1"), val = fp16(0x1.ab4p-8)]; string quantize_43_output_dtype_1 = const()[name = string("quantize_43_output_dtype_1"), val = string("int8")]; tensor quantize_43 = quantize(input = input_171_cast_fp16, output_dtype = quantize_43_output_dtype_1, scale = quantize_43_scale_1)[name = string("quantize_43")]; fp16 dequantize_43_scale_1 = const()[name = string("dequantize_43_scale_1"), val = fp16(0x1.ab4p-8)]; tensor dequantize_43 = dequantize(input = quantize_43, scale = dequantize_43_scale_1)[name = string("dequantize_43")]; tensor x_axes_0 = const()[name = string("x_axes_0"), val = tensor([-2, -1])]; bool x_keep_dims_0 = const()[name = string("x_keep_dims_0"), val = bool(true)]; tensor x_cast_fp16 = reduce_mean(axes = x_axes_0, keep_dims = x_keep_dims_0, x = dequantize_43)[name = string("x_cast_fp16")]; tensor var_476 = const()[name = string("op_476"), val = tensor([1, -1])]; tensor input_173_cast_fp16 = reshape(shape = var_476, x = x_cast_fp16)[name = string("input_173_cast_fp16")]; tensor linear_bias_to_fp16 = const()[name = string("linear_bias_to_fp16"), val = tensor([-0x1.28p-5])]; tensor linear_0_cast_fp16 = linear(bias = linear_bias_to_fp16, weight = linear_weight_cast_fp16, x = input_173_cast_fp16)[name = string("linear_0_cast_fp16")]; tensor smudge_probabilities = sigmoid(x = linear_0_cast_fp16)[name = string("op_481_cast_fp16")]; } -> (smudge_probabilities); }