program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3520.2.1"}, {"coremlc-version", "3520.2.1"}, {"mldb_token", "mldb-jvakw8rano"}})] { func main(tensor token_ids) [FlexibleShapeInformation = tuple, dict, tensor>>, tuple, dict, list, ?>>>>((("DefaultShapes", {{"token_ids", [1, 1]}}), ("RangeDims", {{"token_ids", [[1, 1], [1, 64]]}})))] { tensor embedded_input_axis_0 = const()[name = tensor("embedded_input_axis_0"), val = tensor(0)]; tensor word_embs_weight_to_fp16 = const()[name = tensor("word_embs_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor embedded_input_cast = gather(axis = embedded_input_axis_0, indices = token_ids, x = word_embs_weight_to_fp16); tensor embedded_input_batch_first_transpose_perm_0 = const()[name = tensor("embedded_input_batch_first_transpose_perm_0"), val = tensor([1, 0, 2])]; tensor embedded_input_batch_first_transpose_cast_to_fp32_dtype_0 = const()[name = tensor("embedded_input_batch_first_transpose_cast_to_fp32_dtype_0"), val = tensor("fp32")]; tensor add_0 = const()[name = tensor("add_0"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1600320)))]; tensor add_1 = const()[name = tensor("add_1"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1601408)))]; tensor concat_4 = const()[name = tensor("concat_4"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1602496)))]; tensor concat_5 = const()[name = tensor("concat_5"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1684480)))]; tensor concat_6 = const()[name = tensor("concat_6"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1750080)))]; tensor concat_7 = const()[name = tensor("concat_7"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1832064)))]; tensor lstm_output_lstm_layer_0_lstm_h0_reshaped = const()[name = tensor("lstm_output_lstm_layer_0_lstm_h0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1897664)))]; tensor lstm_output_lstm_layer_0_lstm_c0_reshaped = const()[name = tensor("lstm_output_lstm_layer_0_lstm_c0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1898240)))]; tensor lstm_output_lstm_layer_0_direction_0 = const()[name = tensor("lstm_output_lstm_layer_0_direction_0"), val = tensor("bidirectional")]; tensor lstm_output_lstm_layer_0_output_sequence_0 = const()[name = tensor("lstm_output_lstm_layer_0_output_sequence_0"), val = tensor(true)]; tensor lstm_output_lstm_layer_0_recurrent_activation_0 = const()[name = tensor("lstm_output_lstm_layer_0_recurrent_activation_0"), val = tensor("sigmoid")]; tensor lstm_output_lstm_layer_0_cell_activation_0 = const()[name = tensor("lstm_output_lstm_layer_0_cell_activation_0"), val = tensor("tanh")]; tensor lstm_output_lstm_layer_0_activation_0 = const()[name = tensor("lstm_output_lstm_layer_0_activation_0"), val = tensor("tanh")]; tensor transpose_1 = transpose(perm = embedded_input_batch_first_transpose_perm_0, x = embedded_input_cast); tensor cast_27 = cast(dtype = embedded_input_batch_first_transpose_cast_to_fp32_dtype_0, x = transpose_1); tensor lstm_output_lstm_layer_0_0, tensor lstm_output_lstm_layer_0_1, tensor lstm_output_lstm_layer_0_2 = lstm(activation = lstm_output_lstm_layer_0_activation_0, bias = add_0, bias_back = add_1, cell_activation = lstm_output_lstm_layer_0_cell_activation_0, direction = lstm_output_lstm_layer_0_direction_0, initial_c = lstm_output_lstm_layer_0_lstm_c0_reshaped, initial_h = lstm_output_lstm_layer_0_lstm_h0_reshaped, output_sequence = lstm_output_lstm_layer_0_output_sequence_0, recurrent_activation = lstm_output_lstm_layer_0_recurrent_activation_0, weight_hh = concat_5, weight_hh_back = concat_7, weight_ih = concat_4, weight_ih_back = concat_6, x = cast_27); tensor add_2 = const()[name = tensor("add_2"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1898816)))]; tensor add_3 = const()[name = tensor("add_3"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1899904)))]; tensor concat_14 = const()[name = tensor("concat_14"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1900992)))]; tensor concat_15 = const()[name = tensor("concat_15"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2032128)))]; tensor concat_16 = const()[name = tensor("concat_16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2097728)))]; tensor concat_17 = const()[name = tensor("concat_17"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2228864)))]; tensor lstm_output_lstm_layer_1_lstm_h0_reshaped = const()[name = tensor("lstm_output_lstm_layer_1_lstm_h0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2294464)))]; tensor lstm_output_lstm_layer_1_lstm_c0_reshaped = const()[name = tensor("lstm_output_lstm_layer_1_lstm_c0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2295040)))]; tensor lstm_output_lstm_layer_1_direction_0 = const()[name = tensor("lstm_output_lstm_layer_1_direction_0"), val = tensor("bidirectional")]; tensor lstm_output_lstm_layer_1_output_sequence_0 = const()[name = tensor("lstm_output_lstm_layer_1_output_sequence_0"), val = tensor(true)]; tensor lstm_output_lstm_layer_1_recurrent_activation_0 = const()[name = tensor("lstm_output_lstm_layer_1_recurrent_activation_0"), val = tensor("sigmoid")]; tensor lstm_output_lstm_layer_1_cell_activation_0 = const()[name = tensor("lstm_output_lstm_layer_1_cell_activation_0"), val = tensor("tanh")]; tensor lstm_output_lstm_layer_1_activation_0 = const()[name = tensor("lstm_output_lstm_layer_1_activation_0"), val = tensor("tanh")]; tensor lstm_output_lstm_layer_1_0, tensor lstm_output_lstm_layer_1_1, tensor lstm_output_lstm_layer_1_2 = lstm(activation = lstm_output_lstm_layer_1_activation_0, bias = add_2, bias_back = add_3, cell_activation = lstm_output_lstm_layer_1_cell_activation_0, direction = lstm_output_lstm_layer_1_direction_0, initial_c = lstm_output_lstm_layer_1_lstm_c0_reshaped, initial_h = lstm_output_lstm_layer_1_lstm_h0_reshaped, output_sequence = lstm_output_lstm_layer_1_output_sequence_0, recurrent_activation = lstm_output_lstm_layer_1_recurrent_activation_0, weight_hh = concat_15, weight_hh_back = concat_17, weight_ih = concat_14, weight_ih_back = concat_16, x = lstm_output_lstm_layer_0_0); tensor add_4 = const()[name = tensor("add_4"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2295616)))]; tensor add_5 = const()[name = tensor("add_5"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2296704)))]; tensor concat_24 = const()[name = tensor("concat_24"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2297792)))]; tensor concat_25 = const()[name = tensor("concat_25"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2428928)))]; tensor concat_26 = const()[name = tensor("concat_26"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2494528)))]; tensor concat_27 = const()[name = tensor("concat_27"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2625664)))]; tensor lstm_output_batch_first_lstm_h0_reshaped = const()[name = tensor("lstm_output_batch_first_lstm_h0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2691264)))]; tensor lstm_output_batch_first_lstm_c0_reshaped = const()[name = tensor("lstm_output_batch_first_lstm_c0_reshaped"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2691840)))]; tensor lstm_output_batch_first_direction_0 = const()[name = tensor("lstm_output_batch_first_direction_0"), val = tensor("bidirectional")]; tensor lstm_output_batch_first_output_sequence_0 = const()[name = tensor("lstm_output_batch_first_output_sequence_0"), val = tensor(true)]; tensor lstm_output_batch_first_recurrent_activation_0 = const()[name = tensor("lstm_output_batch_first_recurrent_activation_0"), val = tensor("sigmoid")]; tensor lstm_output_batch_first_cell_activation_0 = const()[name = tensor("lstm_output_batch_first_cell_activation_0"), val = tensor("tanh")]; tensor lstm_output_batch_first_activation_0 = const()[name = tensor("lstm_output_batch_first_activation_0"), val = tensor("tanh")]; tensor lstm_output_batch_first_0, tensor lstm_output_batch_first_1, tensor lstm_output_batch_first_2 = lstm(activation = lstm_output_batch_first_activation_0, bias = add_4, bias_back = add_5, cell_activation = lstm_output_batch_first_cell_activation_0, direction = lstm_output_batch_first_direction_0, initial_c = lstm_output_batch_first_lstm_c0_reshaped, initial_h = lstm_output_batch_first_lstm_h0_reshaped, output_sequence = lstm_output_batch_first_output_sequence_0, recurrent_activation = lstm_output_batch_first_recurrent_activation_0, weight_hh = concat_25, weight_hh_back = concat_27, weight_ih = concat_24, weight_ih_back = concat_26, x = lstm_output_lstm_layer_1_0); tensor lstm_output_perm_0 = const()[name = tensor("lstm_output_perm_0"), val = tensor([1, 0, 2])]; tensor lstm_output_batch_first_0_to_fp16_dtype_0 = const()[name = tensor("lstm_output_batch_first_0_to_fp16_dtype_0"), val = tensor("fp16")]; tensor lstm_output_cast_to_fp32_dtype_0 = const()[name = tensor("lstm_output_cast_to_fp32_dtype_0"), val = tensor("fp32")]; tensor split_32_num_splits_0 = const()[name = tensor("split_32_num_splits_0"), val = tensor(2)]; tensor split_32_axis_0 = const()[name = tensor("split_32_axis_0"), val = tensor(1)]; tensor lstm_output_lstm_layer_0_1_to_fp16_dtype_0 = const()[name = tensor("lstm_output_lstm_layer_0_1_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cast_24 = cast(dtype = lstm_output_lstm_layer_0_1_to_fp16_dtype_0, x = lstm_output_lstm_layer_0_1); tensor split_32_cast_0, tensor split_32_cast_1 = split(axis = split_32_axis_0, num_splits = split_32_num_splits_0, x = cast_24); tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(0)]; tensor stack_0_cast = stack(axis = stack_0_axis_0, values = (split_32_cast_0, split_32_cast_1)); tensor split_33_num_splits_0 = const()[name = tensor("split_33_num_splits_0"), val = tensor(2)]; tensor split_33_axis_0 = const()[name = tensor("split_33_axis_0"), val = tensor(1)]; tensor lstm_output_lstm_layer_1_1_to_fp16_dtype_0 = const()[name = tensor("lstm_output_lstm_layer_1_1_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cast_20 = cast(dtype = lstm_output_lstm_layer_1_1_to_fp16_dtype_0, x = lstm_output_lstm_layer_1_1); tensor split_33_cast_0, tensor split_33_cast_1 = split(axis = split_33_axis_0, num_splits = split_33_num_splits_0, x = cast_20); tensor stack_1_axis_0 = const()[name = tensor("stack_1_axis_0"), val = tensor(0)]; tensor stack_1_cast = stack(axis = stack_1_axis_0, values = (split_33_cast_0, split_33_cast_1)); tensor split_34_num_splits_0 = const()[name = tensor("split_34_num_splits_0"), val = tensor(2)]; tensor split_34_axis_0 = const()[name = tensor("split_34_axis_0"), val = tensor(1)]; tensor lstm_output_batch_first_1_to_fp16_dtype_0 = const()[name = tensor("lstm_output_batch_first_1_to_fp16_dtype_0"), val = tensor("fp16")]; tensor cast_16 = cast(dtype = lstm_output_batch_first_1_to_fp16_dtype_0, x = lstm_output_batch_first_1); tensor split_34_cast_0, tensor split_34_cast_1 = split(axis = split_34_axis_0, num_splits = split_34_num_splits_0, x = cast_16); tensor stack_2_axis_0 = const()[name = tensor("stack_2_axis_0"), val = tensor(0)]; tensor stack_2_cast = stack(axis = stack_2_axis_0, values = (split_34_cast_0, split_34_cast_1)); tensor hn_1_axis_0 = const()[name = tensor("hn_1_axis_0"), val = tensor(0)]; tensor hn_1_interleave_0 = const()[name = tensor("hn_1_interleave_0"), val = tensor(false)]; tensor hn_1_cast = concat(axis = hn_1_axis_0, interleave = hn_1_interleave_0, values = (stack_0_cast, stack_1_cast, stack_2_cast)); tensor hn_3_begin_0 = const()[name = tensor("hn_3_begin_0"), val = tensor([0, 0, 0])]; tensor hn_3_end_0 = const()[name = tensor("hn_3_end_0"), val = tensor([1, 1, 64])]; tensor hn_3_end_mask_0 = const()[name = tensor("hn_3_end_mask_0"), val = tensor([false, true, true])]; tensor hn_3_squeeze_mask_0 = const()[name = tensor("hn_3_squeeze_mask_0"), val = tensor([true, false, false])]; tensor hn_3_cast = slice_by_index(begin = hn_3_begin_0, end = hn_3_end_0, end_mask = hn_3_end_mask_0, squeeze_mask = hn_3_squeeze_mask_0, x = hn_1_cast); tensor var_90_axes_0 = const()[name = tensor("op_90_axes_0"), val = tensor([1])]; tensor var_90_cast = expand_dims(axes = var_90_axes_0, x = hn_3_cast); tensor hn_reps_0 = const()[name = tensor("hn_reps_0"), val = tensor([1, 64, 1])]; tensor hn_cast = tile(reps = hn_reps_0, x = var_90_cast); tensor hn_cast_to_fp32_dtype_0 = const()[name = tensor("hn_cast_to_fp32_dtype_0"), val = tensor("fp32")]; tensor var_98 = const()[name = tensor("op_98"), val = tensor(2)]; tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; tensor cast_26 = cast(dtype = lstm_output_batch_first_0_to_fp16_dtype_0, x = lstm_output_batch_first_0); tensor transpose_0 = transpose(perm = lstm_output_perm_0, x = cast_26); tensor input_1_cast = concat(axis = var_98, interleave = input_1_interleave_0, values = (transpose_0, hn_cast)); tensor input_3_cast = relu(x = input_1_cast); tensor attn_weight_to_fp16 = const()[name = tensor("attn_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2692416)))]; tensor attn_bias_to_fp16 = const()[name = tensor("attn_bias_to_fp16"), val = tensor([-0x1.d94p-3])]; tensor input_5_cast = linear(bias = attn_bias_to_fp16, weight = attn_weight_to_fp16, x = input_3_cast); tensor var_104 = const()[name = tensor("op_104"), val = tensor(1)]; tensor attn_weights_cast = softmax(axis = var_104, x = input_5_cast); tensor input_7_cast = mul(x = transpose_0, y = attn_weights_cast); tensor tagged_weight_to_fp16 = const()[name = tensor("tagged_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2692864)))]; tensor tagged_bias_to_fp16 = const()[name = tensor("tagged_bias_to_fp16"), val = tensor([0x1.05cp+1, -0x1.6bp-1, -0x1.744p+0, -0x1.5ep+0, -0x1.5b4p+1, 0x1.38p+0])]; tensor input_cast = linear(bias = tagged_bias_to_fp16, weight = tagged_weight_to_fp16, x = input_7_cast); tensor var_111 = const()[name = tensor("op_111"), val = tensor(2)]; tensor sequence_tags_cast = softmax(axis = var_111, x = input_cast); tensor sequence_tags_cast_to_fp32_dtype_0 = const()[name = tensor("sequence_tags_cast_to_fp32_dtype_0"), val = tensor("fp32")]; tensor reduce_max_0_axes_0 = const()[name = tensor("reduce_max_0_axes_0"), val = tensor([2])]; tensor reduce_max_0_keep_dims_0 = const()[name = tensor("reduce_max_0_keep_dims_0"), val = tensor(false)]; tensor reduce_max_0_cast = reduce_max(axes = reduce_max_0_axes_0, keep_dims = reduce_max_0_keep_dims_0, x = sequence_tags_cast); tensor reduce_max_0_cast_to_fp32_dtype_0 = const()[name = tensor("reduce_max_0_cast_to_fp32_dtype_0"), val = tensor("fp32")]; tensor reduce_argmax_0_axis_0 = const()[name = tensor("reduce_argmax_0_axis_0"), val = tensor(2)]; tensor reduce_argmax_0_keep_dims_0 = const()[name = tensor("reduce_argmax_0_keep_dims_0"), val = tensor(false)]; tensor reduce_argmax_0 = reduce_argmax(axis = reduce_argmax_0_axis_0, keep_dims = reduce_argmax_0_keep_dims_0, x = sequence_tags_cast); tensor reduce_max_0 = cast(dtype = reduce_max_0_cast_to_fp32_dtype_0, x = reduce_max_0_cast); tensor sequence_tags = cast(dtype = sequence_tags_cast_to_fp32_dtype_0, x = sequence_tags_cast); tensor hn = cast(dtype = hn_cast_to_fp32_dtype_0, x = hn_cast); tensor lstm_output = cast(dtype = lstm_output_cast_to_fp32_dtype_0, x = transpose_0); } -> (lstm_output, hn, sequence_tags, reduce_argmax_0, reduce_max_0); }