program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.1"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.2"}, {"mldb_token", "mldb-rva93zja72"}})] { func main(tensor input) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"input", [1, 512]}}), ("RangeDims", {{"input", [[1, 64], [512, 512]]}})))] { tensor input_to_fp16_dtype_0 = const()[name = tensor("input_to_fp16_dtype_0"), val = tensor("fp16")]; tensor model_encoder_0_weight_to_fp16 = const()[name = tensor("model_encoder_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor model_encoder_0_bias_to_fp16 = const()[name = tensor("model_encoder_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131200)))]; tensor input_to_fp16 = cast(dtype = input_to_fp16_dtype_0, x = input)[name = tensor("cast_1")]; tensor linear_0_cast_fp16 = linear(bias = model_encoder_0_bias_to_fp16, weight = model_encoder_0_weight_to_fp16, x = input_to_fp16)[name = tensor("linear_0_cast_fp16")]; tensor input_9_cast_fp16 = relu(x = linear_0_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor model_encoder_3_weight_to_fp16 = const()[name = tensor("model_encoder_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131520)))]; tensor model_encoder_3_bias_to_fp16 = const()[name = tensor("model_encoder_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164352)))]; tensor linear_1_cast_fp16 = linear(bias = model_encoder_3_bias_to_fp16, weight = model_encoder_3_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_21_cast_fp16 = relu(x = linear_1_cast_fp16)[name = tensor("input_21_cast_fp16")]; tensor model_encoder_6_weight_to_fp16 = const()[name = tensor("model_encoder_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(164672)))]; tensor model_encoder_6_bias_to_fp16 = const()[name = tensor("model_encoder_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181120)))]; tensor linear_2_cast_fp16 = linear(bias = model_encoder_6_bias_to_fp16, weight = model_encoder_6_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor classification_head_weight_to_fp16 = const()[name = tensor("classification_head_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181312)))]; tensor classification_head_bias_to_fp16 = const()[name = tensor("classification_head_bias_to_fp16"), val = tensor([0x1.e08p-4, -0x1.d1cp-4, 0x1.c1p-4, -0x1.474p-4, -0x1.26cp-5])]; tensor linear_3_cast_fp16 = linear(bias = classification_head_bias_to_fp16, weight = classification_head_weight_to_fp16, x = linear_2_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor linear_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("linear_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor logits_1 = cast(dtype = linear_3_cast_fp16_to_fp32_dtype_0, x = linear_3_cast_fp16)[name = tensor("cast_0")]; } -> (logits_1); }