diff --git a/nncf/onnx/graph/model_transformer.py b/nncf/onnx/graph/model_transformer.py index b6db3d36b0d..877edb6838b 100644 --- a/nncf/onnx/graph/model_transformer.py +++ b/nncf/onnx/graph/model_transformer.py @@ -199,9 +199,10 @@ def _apply_quantizer_insertion_transformations( :return: New model with inserted QuantizeLinear-DequantizeLinear nodes pairs. """ self._added_target_edges = Counter() + node_mapping = get_name_to_node_map(model) + children_node_mapping = get_children_node_mapping(model) for transformation in transformations: - children_node_mapping = get_children_node_mapping(model) - model = self._insert_quantizer_dequantizer(model, transformation, children_node_mapping) + model = self._insert_quantizer_dequantizer(model, transformation, node_mapping, children_node_mapping) return model def _get_quantize_dequantize_nodes( @@ -303,6 +304,7 @@ def _insert_quantizer_dequantizer( self, model: onnx.ModelProto, transformation: ONNXQuantizerInsertionCommand, + node_mapping: Dict[str, onnx.NodeProto], children_node_mapping: Dict[str, List[onnx.ValueInfoProto]], ) -> onnx.ModelProto: """ @@ -310,10 +312,10 @@ def _insert_quantizer_dequantizer( :param model: Model to insert new nodes. :param transformation: QuantizeLinear-DequantizeLinear insertion transformation. + :param node_mapping: Mapping from node name to the node. :param children_node_mapping: Mapping from edge name to nodes which consume this edge as an input. :return: Updated model with inserted QuantizeLinear-DequantizeLinear pair. """ - node_mapping = get_name_to_node_map(model) target_edge_name = self._get_quantizer_dequantizer_edge_name(transformation, node_mapping) quantizer, dequantizer = self._get_quantize_dequantize_nodes(transformation, target_edge_name) onnx_scale_tensor, onnx_zero_point_tensor = ONNXModelTransformer._get_scale_zero_point_tensors( diff --git a/tests/onnx/data/reference_graphs/quantization/MaskRCNN-12.dot b/tests/onnx/data/reference_graphs/quantization/MaskRCNN-12.dot index 0775c236c76..e2a0350705f 100644 --- a/tests/onnx/data/reference_graphs/quantization/MaskRCNN-12.dot +++ b/tests/onnx/data/reference_graphs/quantization/MaskRCNN-12.dot @@ -481,10 +481,10 @@ strict digraph { "479 QuantizeLinear_389_1" [id=479, type=QuantizeLinear]; "480 DequantizeLinear_389_1" [id=480, type=DequantizeLinear]; "481 390" [id=481, type=Conv]; -"482 QuantizeLinear_391_2" [id=482, type=QuantizeLinear]; -"483 DequantizeLinear_391_2" [id=483, type=DequantizeLinear]; -"484 QuantizeLinear_391_1" [id=484, type=QuantizeLinear]; -"485 DequantizeLinear_391_1" [id=485, type=DequantizeLinear]; +"482 QuantizeLinear_391_1" [id=482, type=QuantizeLinear]; +"483 DequantizeLinear_391_1" [id=483, type=DequantizeLinear]; +"484 QuantizeLinear_391_2" [id=484, type=QuantizeLinear]; +"485 DequantizeLinear_391_2" [id=485, type=DequantizeLinear]; "486 487" [id=486, type=MaxPool]; "487 QuantizeLinear_489_1" [id=487, type=QuantizeLinear]; "488 DequantizeLinear_489_1" [id=488, type=DequantizeLinear]; @@ -1749,14 +1749,14 @@ strict digraph { "1747 1172" [id=1747, type=Gather]; "1748 2479" [id=1748, type=Concat]; "1749 2490" [id=1749, type=Gather]; -"1750 QuantizeLinear_2527_4" [id=1750, type=QuantizeLinear]; -"1751 DequantizeLinear_2527_4" [id=1751, type=DequantizeLinear]; -"1752 QuantizeLinear_2527_3" [id=1752, type=QuantizeLinear]; -"1753 DequantizeLinear_2527_3" [id=1753, type=DequantizeLinear]; -"1754 QuantizeLinear_2527_2" [id=1754, type=QuantizeLinear]; -"1755 DequantizeLinear_2527_2" [id=1755, type=DequantizeLinear]; -"1756 QuantizeLinear_2527_1" [id=1756, type=QuantizeLinear]; -"1757 DequantizeLinear_2527_1" [id=1757, type=DequantizeLinear]; +"1750 QuantizeLinear_2527_1" [id=1750, type=QuantizeLinear]; +"1751 DequantizeLinear_2527_1" [id=1751, type=DequantizeLinear]; +"1752 QuantizeLinear_2527_2" [id=1752, type=QuantizeLinear]; +"1753 DequantizeLinear_2527_2" [id=1753, type=DequantizeLinear]; +"1754 QuantizeLinear_2527_3" [id=1754, type=QuantizeLinear]; +"1755 DequantizeLinear_2527_3" [id=1755, type=DequantizeLinear]; +"1756 QuantizeLinear_2527_4" [id=1756, type=QuantizeLinear]; +"1757 DequantizeLinear_2527_4" [id=1757, type=DequantizeLinear]; "1758 2532" [id=1758, type=Slice]; "1759 2534" [id=1759, type=Gather]; "1760 2525" [id=1760, type=Slice]; @@ -3692,14 +3692,14 @@ strict digraph { "3690 3030" [id=3690, type=Gather]; "3691 6518" [id=3691, type=Concat]; "3692 6530" [id=3692, type=Gather]; -"3693 QuantizeLinear_6568_4" [id=3693, type=QuantizeLinear]; -"3694 DequantizeLinear_6568_4" [id=3694, type=DequantizeLinear]; -"3695 QuantizeLinear_6568_3" [id=3695, type=QuantizeLinear]; -"3696 DequantizeLinear_6568_3" [id=3696, type=DequantizeLinear]; -"3697 QuantizeLinear_6568_2" [id=3697, type=QuantizeLinear]; -"3698 DequantizeLinear_6568_2" [id=3698, type=DequantizeLinear]; -"3699 QuantizeLinear_6568_1" [id=3699, type=QuantizeLinear]; -"3700 DequantizeLinear_6568_1" [id=3700, type=DequantizeLinear]; +"3693 QuantizeLinear_6568_1" [id=3693, type=QuantizeLinear]; +"3694 DequantizeLinear_6568_1" [id=3694, type=DequantizeLinear]; +"3695 QuantizeLinear_6568_2" [id=3695, type=QuantizeLinear]; +"3696 DequantizeLinear_6568_2" [id=3696, type=DequantizeLinear]; +"3697 QuantizeLinear_6568_3" [id=3697, type=QuantizeLinear]; +"3698 DequantizeLinear_6568_3" [id=3698, type=DequantizeLinear]; +"3699 QuantizeLinear_6568_4" [id=3699, type=QuantizeLinear]; +"3700 DequantizeLinear_6568_4" [id=3700, type=DequantizeLinear]; "3701 6576" [id=3701, type=Slice]; "3702 6578" [id=3702, type=Gather]; "3703 6569" [id=3703, type=Slice]; @@ -4788,16 +4788,16 @@ strict digraph { "478 DequantizeLinear_388_1" -> "481 390" [label="[1, 256, -1, -1]", style=solid]; "479 QuantizeLinear_389_1" -> "480 DequantizeLinear_389_1" [label="[256, 256, 3, 3]", style=dashed]; "480 DequantizeLinear_389_1" -> "481 390" [label="[256, 256, 3, 3]", style=solid]; -"481 390" -> "482 QuantizeLinear_391_2" [label="[1, 256, -1, -1]", style=solid]; -"481 390" -> "484 QuantizeLinear_391_1" [label="[1, 256, -1, -1]", style=solid]; +"481 390" -> "482 QuantizeLinear_391_1" [label="[1, 256, -1, -1]", style=solid]; +"481 390" -> "484 QuantizeLinear_391_2" [label="[1, 256, -1, -1]", style=solid]; "481 390" -> "784 536" [label="[1, 256, -1, -1]", style=solid]; "481 390" -> "787 533" [label="[1, 256, -1, -1]", style=solid]; "481 390" -> "1929 2620" [label="[1, 256, -1, -1]", style=solid]; "481 390" -> "3872 6664" [label="[1, 256, -1, -1]", style=solid]; -"482 QuantizeLinear_391_2" -> "483 DequantizeLinear_391_2" [label="[1, 256, -1, -1]", style=dashed]; -"483 DequantizeLinear_391_2" -> "722 506" [label="[1, 256, -1, -1]", style=solid]; -"484 QuantizeLinear_391_1" -> "485 DequantizeLinear_391_1" [label="[1, 256, -1, -1]", style=dashed]; -"485 DequantizeLinear_391_1" -> "486 487" [label="[1, 256, -1, -1]", style=solid]; +"482 QuantizeLinear_391_1" -> "483 DequantizeLinear_391_1" [label="[1, 256, -1, -1]", style=dashed]; +"483 DequantizeLinear_391_1" -> "486 487" [label="[1, 256, -1, -1]", style=solid]; +"484 QuantizeLinear_391_2" -> "485 DequantizeLinear_391_2" [label="[1, 256, -1, -1]", style=dashed]; +"485 DequantizeLinear_391_2" -> "722 506" [label="[1, 256, -1, -1]", style=solid]; "486 487" -> "489 510" [label="[1, 256, -1, -1]", style=solid]; "486 487" -> "555 542" [label="[1, 256, -1, -1]", style=solid]; "486 487" -> "558 539" [label="[1, 256, -1, -1]", style=solid]; @@ -6341,21 +6341,21 @@ strict digraph { "1746 1171" -> "1747 1172" [label="[-1]", style=dashed]; "1747 1172" -> "1748 2479" [label="[-1, 4]", style=solid]; "1748 2479" -> "1749 2490" [label="[-1, 4]", style=solid]; -"1749 2490" -> "1750 QuantizeLinear_2527_4" [label="[]", style=solid]; -"1749 2490" -> "1752 QuantizeLinear_2527_3" [label="[]", style=solid]; -"1749 2490" -> "1754 QuantizeLinear_2527_2" [label="[]", style=solid]; -"1749 2490" -> "1756 QuantizeLinear_2527_1" [label="[]", style=solid]; +"1749 2490" -> "1750 QuantizeLinear_2527_1" [label="[]", style=solid]; +"1749 2490" -> "1752 QuantizeLinear_2527_2" [label="[]", style=solid]; +"1749 2490" -> "1754 QuantizeLinear_2527_3" [label="[]", style=solid]; +"1749 2490" -> "1756 QuantizeLinear_2527_4" [label="[]", style=solid]; "1749 2490" -> "1803 2495" [label="[]", style=solid]; "1749 2490" -> "1807 2503" [label="[]", style=solid]; "1749 2490" -> "2009 2775" [label="[]", style=solid]; -"1750 QuantizeLinear_2527_4" -> "1751 DequantizeLinear_2527_4" [label="[]", style=dashed]; -"1751 DequantizeLinear_2527_4" -> "1768 2508" [label="[]", style=solid]; -"1752 QuantizeLinear_2527_3" -> "1753 DequantizeLinear_2527_3" [label="[]", style=dashed]; -"1753 DequantizeLinear_2527_3" -> "1766 2515" [label="[]", style=solid]; -"1754 QuantizeLinear_2527_2" -> "1755 DequantizeLinear_2527_2" [label="[]", style=dashed]; -"1755 DequantizeLinear_2527_2" -> "1760 2525" [label="[]", style=solid]; -"1756 QuantizeLinear_2527_1" -> "1757 DequantizeLinear_2527_1" [label="[]", style=dashed]; -"1757 DequantizeLinear_2527_1" -> "1758 2532" [label="[]", style=solid]; +"1750 QuantizeLinear_2527_1" -> "1751 DequantizeLinear_2527_1" [label="[]", style=dashed]; +"1751 DequantizeLinear_2527_1" -> "1758 2532" [label="[]", style=solid]; +"1752 QuantizeLinear_2527_2" -> "1753 DequantizeLinear_2527_2" [label="[]", style=dashed]; +"1753 DequantizeLinear_2527_2" -> "1760 2525" [label="[]", style=solid]; +"1754 QuantizeLinear_2527_3" -> "1755 DequantizeLinear_2527_3" [label="[]", style=dashed]; +"1755 DequantizeLinear_2527_3" -> "1766 2515" [label="[]", style=solid]; +"1756 QuantizeLinear_2527_4" -> "1757 DequantizeLinear_2527_4" [label="[]", style=dashed]; +"1757 DequantizeLinear_2527_4" -> "1768 2508" [label="[]", style=solid]; "1758 2532" -> "1759 2534" [label="[]", style=solid]; "1759 2534" -> "1762 2535" [label="[]", style=solid]; "1760 2525" -> "1761 2527" [label="[]", style=solid]; @@ -8988,21 +8988,21 @@ strict digraph { "3690 3030" -> "3691 6518" [label="[]", style=solid]; "3690 3030" -> "4259 3037" [label="[]", style=solid]; "3691 6518" -> "3692 6530" [label="[]", style=solid]; -"3692 6530" -> "3693 QuantizeLinear_6568_4" [label="[-1, 4]", style=solid]; -"3692 6530" -> "3695 QuantizeLinear_6568_3" [label="[-1, 4]", style=solid]; -"3692 6530" -> "3697 QuantizeLinear_6568_2" [label="[-1, 4]", style=solid]; -"3692 6530" -> "3699 QuantizeLinear_6568_1" [label="[-1, 4]", style=solid]; +"3692 6530" -> "3693 QuantizeLinear_6568_1" [label="[-1, 4]", style=solid]; +"3692 6530" -> "3695 QuantizeLinear_6568_2" [label="[-1, 4]", style=solid]; +"3692 6530" -> "3697 QuantizeLinear_6568_3" [label="[-1, 4]", style=solid]; +"3692 6530" -> "3699 QuantizeLinear_6568_4" [label="[-1, 4]", style=solid]; "3692 6530" -> "3746 6539" [label="[-1, 4]", style=solid]; "3692 6530" -> "3750 6547" [label="[-1, 4]", style=solid]; "3692 6530" -> "4281 nncf_model_output_0" [label="[-1, 4]", style=solid]; -"3693 QuantizeLinear_6568_4" -> "3694 DequantizeLinear_6568_4" [label="[-1, 4]", style=dashed]; -"3694 DequantizeLinear_6568_4" -> "3711 6552" [label="[-1, 4]", style=solid]; -"3695 QuantizeLinear_6568_3" -> "3696 DequantizeLinear_6568_3" [label="[-1, 4]", style=dashed]; -"3696 DequantizeLinear_6568_3" -> "3709 6559" [label="[-1, 4]", style=solid]; -"3697 QuantizeLinear_6568_2" -> "3698 DequantizeLinear_6568_2" [label="[-1, 4]", style=dashed]; -"3698 DequantizeLinear_6568_2" -> "3703 6569" [label="[-1, 4]", style=solid]; -"3699 QuantizeLinear_6568_1" -> "3700 DequantizeLinear_6568_1" [label="[-1, 4]", style=dashed]; -"3700 DequantizeLinear_6568_1" -> "3701 6576" [label="[-1, 4]", style=solid]; +"3693 QuantizeLinear_6568_1" -> "3694 DequantizeLinear_6568_1" [label="[-1, 4]", style=dashed]; +"3694 DequantizeLinear_6568_1" -> "3701 6576" [label="[-1, 4]", style=solid]; +"3695 QuantizeLinear_6568_2" -> "3696 DequantizeLinear_6568_2" [label="[-1, 4]", style=dashed]; +"3696 DequantizeLinear_6568_2" -> "3703 6569" [label="[-1, 4]", style=solid]; +"3697 QuantizeLinear_6568_3" -> "3698 DequantizeLinear_6568_3" [label="[-1, 4]", style=dashed]; +"3698 DequantizeLinear_6568_3" -> "3709 6559" [label="[-1, 4]", style=solid]; +"3699 QuantizeLinear_6568_4" -> "3700 DequantizeLinear_6568_4" [label="[-1, 4]", style=dashed]; +"3700 DequantizeLinear_6568_4" -> "3711 6552" [label="[-1, 4]", style=solid]; "3701 6576" -> "3702 6578" [label="[-1, 4]", style=solid]; "3702 6578" -> "3705 6579" [label="[-1]", style=solid]; "3703 6569" -> "3704 6571" [label="[-1, 4]", style=solid]; diff --git a/tests/onnx/data/reference_graphs/quantization/bertsquad-12.dot b/tests/onnx/data/reference_graphs/quantization/bertsquad-12.dot index 246765a6663..5e2502ba0c4 100644 --- a/tests/onnx/data/reference_graphs/quantization/bertsquad-12.dot +++ b/tests/onnx/data/reference_graphs/quantization/bertsquad-12.dot @@ -405,14 +405,14 @@ strict digraph { "403 bert/encoder/Reshape_13/shape_Concat__301" [id=403, type=Concat]; "404 bert/encoder/Reshape_13__471" [id=404, type=Cast]; "405 bert/encoder/Reshape_1" [id=405, type=Reshape]; -"406 QuantizeLinear_bert/encoder/Reshape_1^0_3" [id=406, label="406 QuantizeLinear_bert/encoder/Reshape_1:0_3", type=QuantizeLinear]; -"407 DequantizeLinear_bert/encoder/Reshape_1^0_3" [id=407, label="407 DequantizeLinear_bert/encoder/Reshape_1:0_3", type=DequantizeLinear]; -"408 QuantizeLinear_bert/encoder/Reshape_1^0_2" [id=408, label="408 QuantizeLinear_bert/encoder/Reshape_1:0_2", type=QuantizeLinear]; -"409 DequantizeLinear_bert/encoder/Reshape_1^0_2" [id=409, label="409 DequantizeLinear_bert/encoder/Reshape_1:0_2", type=DequantizeLinear]; -"410 QuantizeLinear_bert/encoder/Reshape_1^0_1" [id=410, label="410 QuantizeLinear_bert/encoder/Reshape_1:0_1", type=QuantizeLinear]; -"411 DequantizeLinear_bert/encoder/Reshape_1^0_1" [id=411, label="411 DequantizeLinear_bert/encoder/Reshape_1:0_1", type=DequantizeLinear]; -"412 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [id=412, label="412 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"413 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [id=413, label="413 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"406 QuantizeLinear_bert/encoder/Reshape_1^0_1" [id=406, label="406 QuantizeLinear_bert/encoder/Reshape_1:0_1", type=QuantizeLinear]; +"407 DequantizeLinear_bert/encoder/Reshape_1^0_1" [id=407, label="407 DequantizeLinear_bert/encoder/Reshape_1:0_1", type=DequantizeLinear]; +"408 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [id=408, label="408 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"409 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [id=409, label="409 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"410 QuantizeLinear_bert/encoder/Reshape_1^0_2" [id=410, label="410 QuantizeLinear_bert/encoder/Reshape_1:0_2", type=QuantizeLinear]; +"411 DequantizeLinear_bert/encoder/Reshape_1^0_2" [id=411, label="411 DequantizeLinear_bert/encoder/Reshape_1:0_2", type=DequantizeLinear]; +"412 QuantizeLinear_bert/encoder/Reshape_1^0_3" [id=412, label="412 QuantizeLinear_bert/encoder/Reshape_1:0_3", type=QuantizeLinear]; +"413 DequantizeLinear_bert/encoder/Reshape_1^0_3" [id=413, label="413 DequantizeLinear_bert/encoder/Reshape_1:0_3", type=DequantizeLinear]; "414 bert/encoder/layer_0/attention/self/value/MatMul" [id=414, type=MatMul]; "415 bert/encoder/layer_0/attention/self/value/BiasAdd" [id=415, type=Add]; "416 bert/encoder/layer_0/attention/self/Reshape_2" [id=416, type=Reshape]; @@ -495,14 +495,14 @@ strict digraph { "493 bert/encoder/layer_0/output/LayerNorm/batchnorm/sub" [id=493, type=Sub]; "494 bert/encoder/layer_0/output/LayerNorm/batchnorm/mul_1" [id=494, type=Mul]; "495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" [id=495, type=Add]; -"496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [id=496, label="496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [id=497, label="497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"498 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [id=498, label="498 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"499 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [id=499, label="499 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [id=500, label="500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [id=501, label="501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"502 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [id=502, label="502 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"503 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [id=503, label="503 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [id=496, label="496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [id=497, label="497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"498 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [id=498, label="498 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"499 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [id=499, label="499 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [id=500, label="500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [id=501, label="501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"502 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [id=502, label="502 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"503 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [id=503, label="503 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "504 bert/encoder/layer_1/attention/self/value/MatMul" [id=504, type=MatMul]; "505 bert/encoder/layer_1/attention/self/value/BiasAdd" [id=505, type=Add]; "506 bert/encoder/layer_1/attention/self/Reshape_2" [id=506, type=Reshape]; @@ -585,14 +585,14 @@ strict digraph { "583 bert/encoder/layer_1/output/LayerNorm/batchnorm/sub" [id=583, type=Sub]; "584 bert/encoder/layer_1/output/LayerNorm/batchnorm/mul_1" [id=584, type=Mul]; "585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" [id=585, type=Add]; -"586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [id=586, label="586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [id=587, label="587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"588 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [id=588, label="588 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"589 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [id=589, label="589 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [id=590, label="590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [id=591, label="591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"592 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [id=592, label="592 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"593 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [id=593, label="593 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [id=586, label="586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [id=587, label="587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"588 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [id=588, label="588 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"589 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [id=589, label="589 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [id=590, label="590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [id=591, label="591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"592 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [id=592, label="592 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"593 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [id=593, label="593 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "594 bert/encoder/layer_2/attention/self/value/MatMul" [id=594, type=MatMul]; "595 bert/encoder/layer_2/attention/self/value/BiasAdd" [id=595, type=Add]; "596 bert/encoder/layer_2/attention/self/Reshape_2" [id=596, type=Reshape]; @@ -675,14 +675,14 @@ strict digraph { "673 bert/encoder/layer_2/output/LayerNorm/batchnorm/sub" [id=673, type=Sub]; "674 bert/encoder/layer_2/output/LayerNorm/batchnorm/mul_1" [id=674, type=Mul]; "675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" [id=675, type=Add]; -"676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [id=676, label="676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [id=677, label="677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"678 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [id=678, label="678 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"679 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [id=679, label="679 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [id=680, label="680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [id=681, label="681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"682 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [id=682, label="682 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"683 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [id=683, label="683 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [id=676, label="676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [id=677, label="677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"678 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [id=678, label="678 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"679 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [id=679, label="679 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [id=680, label="680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [id=681, label="681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"682 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [id=682, label="682 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"683 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [id=683, label="683 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "684 bert/encoder/layer_3/attention/self/value/MatMul" [id=684, type=MatMul]; "685 bert/encoder/layer_3/attention/self/value/BiasAdd" [id=685, type=Add]; "686 bert/encoder/layer_3/attention/self/Reshape_2" [id=686, type=Reshape]; @@ -765,14 +765,14 @@ strict digraph { "763 bert/encoder/layer_3/output/LayerNorm/batchnorm/sub" [id=763, type=Sub]; "764 bert/encoder/layer_3/output/LayerNorm/batchnorm/mul_1" [id=764, type=Mul]; "765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" [id=765, type=Add]; -"766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [id=766, label="766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [id=767, label="767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"768 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [id=768, label="768 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"769 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [id=769, label="769 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [id=770, label="770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [id=771, label="771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"772 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [id=772, label="772 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"773 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [id=773, label="773 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [id=766, label="766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [id=767, label="767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"768 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [id=768, label="768 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"769 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [id=769, label="769 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [id=770, label="770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [id=771, label="771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"772 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [id=772, label="772 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"773 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [id=773, label="773 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "774 bert/encoder/layer_4/attention/self/value/MatMul" [id=774, type=MatMul]; "775 bert/encoder/layer_4/attention/self/value/BiasAdd" [id=775, type=Add]; "776 bert/encoder/layer_4/attention/self/Reshape_2" [id=776, type=Reshape]; @@ -855,14 +855,14 @@ strict digraph { "853 bert/encoder/layer_4/output/LayerNorm/batchnorm/sub" [id=853, type=Sub]; "854 bert/encoder/layer_4/output/LayerNorm/batchnorm/mul_1" [id=854, type=Mul]; "855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" [id=855, type=Add]; -"856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [id=856, label="856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [id=857, label="857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"858 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [id=858, label="858 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"859 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [id=859, label="859 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [id=860, label="860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [id=861, label="861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"862 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [id=862, label="862 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"863 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [id=863, label="863 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [id=856, label="856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [id=857, label="857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"858 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [id=858, label="858 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"859 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [id=859, label="859 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [id=860, label="860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [id=861, label="861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"862 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [id=862, label="862 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"863 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [id=863, label="863 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "864 bert/encoder/layer_5/attention/self/value/MatMul" [id=864, type=MatMul]; "865 bert/encoder/layer_5/attention/self/value/BiasAdd" [id=865, type=Add]; "866 bert/encoder/layer_5/attention/self/Reshape_2" [id=866, type=Reshape]; @@ -945,14 +945,14 @@ strict digraph { "943 bert/encoder/layer_5/output/LayerNorm/batchnorm/sub" [id=943, type=Sub]; "944 bert/encoder/layer_5/output/LayerNorm/batchnorm/mul_1" [id=944, type=Mul]; "945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" [id=945, type=Add]; -"946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [id=946, label="946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [id=947, label="947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"948 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [id=948, label="948 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"949 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [id=949, label="949 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [id=950, label="950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [id=951, label="951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"952 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [id=952, label="952 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"953 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [id=953, label="953 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [id=946, label="946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [id=947, label="947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"948 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [id=948, label="948 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"949 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [id=949, label="949 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [id=950, label="950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [id=951, label="951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"952 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [id=952, label="952 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"953 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [id=953, label="953 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "954 bert/encoder/layer_6/attention/self/value/MatMul" [id=954, type=MatMul]; "955 bert/encoder/layer_6/attention/self/value/BiasAdd" [id=955, type=Add]; "956 bert/encoder/layer_6/attention/self/Reshape_2" [id=956, type=Reshape]; @@ -1035,14 +1035,14 @@ strict digraph { "1033 bert/encoder/layer_6/output/LayerNorm/batchnorm/sub" [id=1033, type=Sub]; "1034 bert/encoder/layer_6/output/LayerNorm/batchnorm/mul_1" [id=1034, type=Mul]; "1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" [id=1035, type=Add]; -"1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [id=1036, label="1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [id=1037, label="1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"1038 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [id=1038, label="1038 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"1039 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [id=1039, label="1039 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [id=1040, label="1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [id=1041, label="1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"1042 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [id=1042, label="1042 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"1043 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [id=1043, label="1043 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [id=1036, label="1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [id=1037, label="1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"1038 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [id=1038, label="1038 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"1039 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [id=1039, label="1039 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [id=1040, label="1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [id=1041, label="1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"1042 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [id=1042, label="1042 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"1043 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [id=1043, label="1043 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "1044 bert/encoder/layer_7/attention/self/value/MatMul" [id=1044, type=MatMul]; "1045 bert/encoder/layer_7/attention/self/value/BiasAdd" [id=1045, type=Add]; "1046 bert/encoder/layer_7/attention/self/Reshape_2" [id=1046, type=Reshape]; @@ -1125,14 +1125,14 @@ strict digraph { "1123 bert/encoder/layer_7/output/LayerNorm/batchnorm/sub" [id=1123, type=Sub]; "1124 bert/encoder/layer_7/output/LayerNorm/batchnorm/mul_1" [id=1124, type=Mul]; "1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" [id=1125, type=Add]; -"1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [id=1126, label="1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [id=1127, label="1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"1128 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [id=1128, label="1128 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"1129 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [id=1129, label="1129 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [id=1130, label="1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [id=1131, label="1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"1132 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [id=1132, label="1132 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"1133 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [id=1133, label="1133 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [id=1126, label="1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [id=1127, label="1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"1128 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [id=1128, label="1128 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"1129 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [id=1129, label="1129 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [id=1130, label="1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [id=1131, label="1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"1132 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [id=1132, label="1132 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"1133 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [id=1133, label="1133 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "1134 bert/encoder/layer_8/attention/self/value/MatMul" [id=1134, type=MatMul]; "1135 bert/encoder/layer_8/attention/self/value/BiasAdd" [id=1135, type=Add]; "1136 bert/encoder/layer_8/attention/self/Reshape_2" [id=1136, type=Reshape]; @@ -1215,14 +1215,14 @@ strict digraph { "1213 bert/encoder/layer_8/output/LayerNorm/batchnorm/sub" [id=1213, type=Sub]; "1214 bert/encoder/layer_8/output/LayerNorm/batchnorm/mul_1" [id=1214, type=Mul]; "1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" [id=1215, type=Add]; -"1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [id=1216, label="1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [id=1217, label="1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"1218 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [id=1218, label="1218 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"1219 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [id=1219, label="1219 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [id=1220, label="1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [id=1221, label="1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"1222 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [id=1222, label="1222 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"1223 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [id=1223, label="1223 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [id=1216, label="1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [id=1217, label="1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"1218 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [id=1218, label="1218 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"1219 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [id=1219, label="1219 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [id=1220, label="1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [id=1221, label="1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"1222 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [id=1222, label="1222 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"1223 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [id=1223, label="1223 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "1224 bert/encoder/layer_9/attention/self/value/MatMul" [id=1224, type=MatMul]; "1225 bert/encoder/layer_9/attention/self/value/BiasAdd" [id=1225, type=Add]; "1226 bert/encoder/layer_9/attention/self/Reshape_2" [id=1226, type=Reshape]; @@ -1305,14 +1305,14 @@ strict digraph { "1303 bert/encoder/layer_9/output/LayerNorm/batchnorm/sub" [id=1303, type=Sub]; "1304 bert/encoder/layer_9/output/LayerNorm/batchnorm/mul_1" [id=1304, type=Mul]; "1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" [id=1305, type=Add]; -"1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [id=1306, label="1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [id=1307, label="1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"1308 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [id=1308, label="1308 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"1309 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [id=1309, label="1309 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [id=1310, label="1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [id=1311, label="1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"1312 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [id=1312, label="1312 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"1313 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [id=1313, label="1313 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [id=1306, label="1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [id=1307, label="1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"1308 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [id=1308, label="1308 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"1309 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [id=1309, label="1309 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [id=1310, label="1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [id=1311, label="1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"1312 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [id=1312, label="1312 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"1313 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [id=1313, label="1313 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "1314 bert/encoder/layer_10/attention/self/value/MatMul" [id=1314, type=MatMul]; "1315 bert/encoder/layer_10/attention/self/value/BiasAdd" [id=1315, type=Add]; "1316 bert/encoder/layer_10/attention/self/Reshape_2" [id=1316, type=Reshape]; @@ -1395,14 +1395,14 @@ strict digraph { "1393 bert/encoder/layer_10/output/LayerNorm/batchnorm/sub" [id=1393, type=Sub]; "1394 bert/encoder/layer_10/output/LayerNorm/batchnorm/mul_1" [id=1394, type=Mul]; "1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" [id=1395, type=Add]; -"1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [id=1396, label="1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; -"1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [id=1397, label="1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; -"1398 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [id=1398, label="1398 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; -"1399 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [id=1399, label="1399 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; -"1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [id=1400, label="1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; -"1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [id=1401, label="1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; -"1402 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [id=1402, label="1402 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel:0_1", type=QuantizeLinear]; -"1403 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [id=1403, label="1403 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [id=1396, label="1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_1", type=QuantizeLinear]; +"1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [id=1397, label="1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_1", type=DequantizeLinear]; +"1398 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [id=1398, label="1398 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel:0_1", type=QuantizeLinear]; +"1399 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [id=1399, label="1399 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel:0_1", type=DequantizeLinear]; +"1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [id=1400, label="1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_2", type=QuantizeLinear]; +"1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [id=1401, label="1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_2", type=DequantizeLinear]; +"1402 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [id=1402, label="1402 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_3", type=QuantizeLinear]; +"1403 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [id=1403, label="1403 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1:0_3", type=DequantizeLinear]; "1404 bert/encoder/layer_11/attention/self/value/MatMul" [id=1404, type=MatMul]; "1405 bert/encoder/layer_11/attention/self/value/BiasAdd" [id=1405, type=Add]; "1406 bert/encoder/layer_11/attention/self/Reshape_2" [id=1406, type=Reshape]; @@ -1991,18 +1991,18 @@ strict digraph { "402 bert/encoder/Reshape_13/shape_Unsqueeze__298" -> "403 bert/encoder/Reshape_13/shape_Concat__301" [label="[1]", style=dashed]; "403 bert/encoder/Reshape_13/shape_Concat__301" -> "404 bert/encoder/Reshape_13__471" [label="[3]", style=dashed]; "404 bert/encoder/Reshape_13__471" -> "1488 bert/encoder/Reshape_13" [label="[3]", style=dashed]; -"405 bert/encoder/Reshape_1" -> "406 QuantizeLinear_bert/encoder/Reshape_1^0_3" [label="[]", style=solid]; -"405 bert/encoder/Reshape_1" -> "408 QuantizeLinear_bert/encoder/Reshape_1^0_2" [label="[]", style=solid]; -"405 bert/encoder/Reshape_1" -> "410 QuantizeLinear_bert/encoder/Reshape_1^0_1" [label="[]", style=solid]; +"405 bert/encoder/Reshape_1" -> "406 QuantizeLinear_bert/encoder/Reshape_1^0_1" [label="[]", style=solid]; +"405 bert/encoder/Reshape_1" -> "410 QuantizeLinear_bert/encoder/Reshape_1^0_2" [label="[]", style=solid]; +"405 bert/encoder/Reshape_1" -> "412 QuantizeLinear_bert/encoder/Reshape_1^0_3" [label="[]", style=solid]; "405 bert/encoder/Reshape_1" -> "448 bert/encoder/layer_0/attention/output/add" [label="[]", style=solid]; -"406 QuantizeLinear_bert/encoder/Reshape_1^0_3" -> "407 DequantizeLinear_bert/encoder/Reshape_1^0_3" [label="[]", style=dashed]; -"407 DequantizeLinear_bert/encoder/Reshape_1^0_3" -> "428 bert/encoder/layer_0/attention/self/key/MatMul" [label="[]", style=solid]; -"408 QuantizeLinear_bert/encoder/Reshape_1^0_2" -> "409 DequantizeLinear_bert/encoder/Reshape_1^0_2" [label="[]", style=dashed]; -"409 DequantizeLinear_bert/encoder/Reshape_1^0_2" -> "420 bert/encoder/layer_0/attention/self/query/MatMul" [label="[]", style=solid]; -"410 QuantizeLinear_bert/encoder/Reshape_1^0_1" -> "411 DequantizeLinear_bert/encoder/Reshape_1^0_1" [label="[]", style=dashed]; -"411 DequantizeLinear_bert/encoder/Reshape_1^0_1" -> "414 bert/encoder/layer_0/attention/self/value/MatMul" [label="[]", style=solid]; -"412 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" -> "413 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"413 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" -> "414 bert/encoder/layer_0/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"406 QuantizeLinear_bert/encoder/Reshape_1^0_1" -> "407 DequantizeLinear_bert/encoder/Reshape_1^0_1" [label="[]", style=dashed]; +"407 DequantizeLinear_bert/encoder/Reshape_1^0_1" -> "414 bert/encoder/layer_0/attention/self/value/MatMul" [label="[]", style=solid]; +"408 QuantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" -> "409 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"409 DequantizeLinear_bert/encoder/layer_0/attention/self/value/kernel^0_1" -> "414 bert/encoder/layer_0/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"410 QuantizeLinear_bert/encoder/Reshape_1^0_2" -> "411 DequantizeLinear_bert/encoder/Reshape_1^0_2" [label="[]", style=dashed]; +"411 DequantizeLinear_bert/encoder/Reshape_1^0_2" -> "420 bert/encoder/layer_0/attention/self/query/MatMul" [label="[]", style=solid]; +"412 QuantizeLinear_bert/encoder/Reshape_1^0_3" -> "413 DequantizeLinear_bert/encoder/Reshape_1^0_3" [label="[]", style=dashed]; +"413 DequantizeLinear_bert/encoder/Reshape_1^0_3" -> "428 bert/encoder/layer_0/attention/self/key/MatMul" [label="[]", style=solid]; "414 bert/encoder/layer_0/attention/self/value/MatMul" -> "415 bert/encoder/layer_0/attention/self/value/BiasAdd" [label="[]", style=solid]; "415 bert/encoder/layer_0/attention/self/value/BiasAdd" -> "416 bert/encoder/layer_0/attention/self/Reshape_2" [label="[]", style=solid]; "416 bert/encoder/layer_0/attention/self/Reshape_2" -> "417 bert/encoder/layer_0/attention/self/transpose_2" [label="[]", style=solid]; @@ -2095,18 +2095,18 @@ strict digraph { "492 bert/encoder/layer_0/output/LayerNorm/batchnorm/mul_2" -> "493 bert/encoder/layer_0/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "493 bert/encoder/layer_0/output/LayerNorm/batchnorm/sub" -> "495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "494 bert/encoder/layer_0/output/LayerNorm/batchnorm/mul_1" -> "495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "498 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "502 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "495 bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1" -> "538 bert/encoder/layer_1/attention/output/add" [label="[]", style=solid]; -"496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" -> "497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" -> "518 bert/encoder/layer_1/attention/self/key/MatMul" [label="[]", style=solid]; -"498 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" -> "499 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"499 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" -> "510 bert/encoder/layer_1/attention/self/query/MatMul" [label="[]", style=solid]; -"500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" -> "501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" -> "504 bert/encoder/layer_1/attention/self/value/MatMul" [label="[]", style=solid]; -"502 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" -> "503 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"503 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" -> "504 bert/encoder/layer_1/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"496 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" -> "497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"497 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_1" -> "504 bert/encoder/layer_1/attention/self/value/MatMul" [label="[]", style=solid]; +"498 QuantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" -> "499 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"499 DequantizeLinear_bert/encoder/layer_1/attention/self/value/kernel^0_1" -> "504 bert/encoder/layer_1/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"500 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" -> "501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"501 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_2" -> "510 bert/encoder/layer_1/attention/self/query/MatMul" [label="[]", style=solid]; +"502 QuantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" -> "503 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"503 DequantizeLinear_bert/encoder/layer_0/output/LayerNorm/batchnorm/add_1^0_3" -> "518 bert/encoder/layer_1/attention/self/key/MatMul" [label="[]", style=solid]; "504 bert/encoder/layer_1/attention/self/value/MatMul" -> "505 bert/encoder/layer_1/attention/self/value/BiasAdd" [label="[]", style=solid]; "505 bert/encoder/layer_1/attention/self/value/BiasAdd" -> "506 bert/encoder/layer_1/attention/self/Reshape_2" [label="[]", style=solid]; "506 bert/encoder/layer_1/attention/self/Reshape_2" -> "507 bert/encoder/layer_1/attention/self/transpose_2" [label="[]", style=solid]; @@ -2199,18 +2199,18 @@ strict digraph { "582 bert/encoder/layer_1/output/LayerNorm/batchnorm/mul_2" -> "583 bert/encoder/layer_1/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "583 bert/encoder/layer_1/output/LayerNorm/batchnorm/sub" -> "585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "584 bert/encoder/layer_1/output/LayerNorm/batchnorm/mul_1" -> "585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "588 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "592 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "585 bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1" -> "628 bert/encoder/layer_2/attention/output/add" [label="[]", style=solid]; -"586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" -> "587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" -> "608 bert/encoder/layer_2/attention/self/key/MatMul" [label="[]", style=solid]; -"588 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" -> "589 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"589 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" -> "600 bert/encoder/layer_2/attention/self/query/MatMul" [label="[]", style=solid]; -"590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" -> "591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" -> "594 bert/encoder/layer_2/attention/self/value/MatMul" [label="[]", style=solid]; -"592 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" -> "593 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"593 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" -> "594 bert/encoder/layer_2/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"586 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" -> "587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"587 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_1" -> "594 bert/encoder/layer_2/attention/self/value/MatMul" [label="[]", style=solid]; +"588 QuantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" -> "589 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"589 DequantizeLinear_bert/encoder/layer_2/attention/self/value/kernel^0_1" -> "594 bert/encoder/layer_2/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"590 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" -> "591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"591 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_2" -> "600 bert/encoder/layer_2/attention/self/query/MatMul" [label="[]", style=solid]; +"592 QuantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" -> "593 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"593 DequantizeLinear_bert/encoder/layer_1/output/LayerNorm/batchnorm/add_1^0_3" -> "608 bert/encoder/layer_2/attention/self/key/MatMul" [label="[]", style=solid]; "594 bert/encoder/layer_2/attention/self/value/MatMul" -> "595 bert/encoder/layer_2/attention/self/value/BiasAdd" [label="[]", style=solid]; "595 bert/encoder/layer_2/attention/self/value/BiasAdd" -> "596 bert/encoder/layer_2/attention/self/Reshape_2" [label="[]", style=solid]; "596 bert/encoder/layer_2/attention/self/Reshape_2" -> "597 bert/encoder/layer_2/attention/self/transpose_2" [label="[]", style=solid]; @@ -2303,18 +2303,18 @@ strict digraph { "672 bert/encoder/layer_2/output/LayerNorm/batchnorm/mul_2" -> "673 bert/encoder/layer_2/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "673 bert/encoder/layer_2/output/LayerNorm/batchnorm/sub" -> "675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "674 bert/encoder/layer_2/output/LayerNorm/batchnorm/mul_1" -> "675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "678 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "682 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "675 bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1" -> "718 bert/encoder/layer_3/attention/output/add" [label="[]", style=solid]; -"676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" -> "677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" -> "698 bert/encoder/layer_3/attention/self/key/MatMul" [label="[]", style=solid]; -"678 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" -> "679 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"679 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" -> "690 bert/encoder/layer_3/attention/self/query/MatMul" [label="[]", style=solid]; -"680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" -> "681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" -> "684 bert/encoder/layer_3/attention/self/value/MatMul" [label="[]", style=solid]; -"682 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" -> "683 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"683 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" -> "684 bert/encoder/layer_3/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"676 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" -> "677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"677 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_1" -> "684 bert/encoder/layer_3/attention/self/value/MatMul" [label="[]", style=solid]; +"678 QuantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" -> "679 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"679 DequantizeLinear_bert/encoder/layer_3/attention/self/value/kernel^0_1" -> "684 bert/encoder/layer_3/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"680 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" -> "681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"681 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_2" -> "690 bert/encoder/layer_3/attention/self/query/MatMul" [label="[]", style=solid]; +"682 QuantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" -> "683 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"683 DequantizeLinear_bert/encoder/layer_2/output/LayerNorm/batchnorm/add_1^0_3" -> "698 bert/encoder/layer_3/attention/self/key/MatMul" [label="[]", style=solid]; "684 bert/encoder/layer_3/attention/self/value/MatMul" -> "685 bert/encoder/layer_3/attention/self/value/BiasAdd" [label="[]", style=solid]; "685 bert/encoder/layer_3/attention/self/value/BiasAdd" -> "686 bert/encoder/layer_3/attention/self/Reshape_2" [label="[]", style=solid]; "686 bert/encoder/layer_3/attention/self/Reshape_2" -> "687 bert/encoder/layer_3/attention/self/transpose_2" [label="[]", style=solid]; @@ -2407,18 +2407,18 @@ strict digraph { "762 bert/encoder/layer_3/output/LayerNorm/batchnorm/mul_2" -> "763 bert/encoder/layer_3/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "763 bert/encoder/layer_3/output/LayerNorm/batchnorm/sub" -> "765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "764 bert/encoder/layer_3/output/LayerNorm/batchnorm/mul_1" -> "765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "768 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "772 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "765 bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1" -> "808 bert/encoder/layer_4/attention/output/add" [label="[]", style=solid]; -"766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" -> "767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" -> "788 bert/encoder/layer_4/attention/self/key/MatMul" [label="[]", style=solid]; -"768 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" -> "769 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"769 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" -> "780 bert/encoder/layer_4/attention/self/query/MatMul" [label="[]", style=solid]; -"770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" -> "771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" -> "774 bert/encoder/layer_4/attention/self/value/MatMul" [label="[]", style=solid]; -"772 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" -> "773 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"773 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" -> "774 bert/encoder/layer_4/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"766 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" -> "767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"767 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_1" -> "774 bert/encoder/layer_4/attention/self/value/MatMul" [label="[]", style=solid]; +"768 QuantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" -> "769 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"769 DequantizeLinear_bert/encoder/layer_4/attention/self/value/kernel^0_1" -> "774 bert/encoder/layer_4/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"770 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" -> "771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"771 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_2" -> "780 bert/encoder/layer_4/attention/self/query/MatMul" [label="[]", style=solid]; +"772 QuantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" -> "773 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"773 DequantizeLinear_bert/encoder/layer_3/output/LayerNorm/batchnorm/add_1^0_3" -> "788 bert/encoder/layer_4/attention/self/key/MatMul" [label="[]", style=solid]; "774 bert/encoder/layer_4/attention/self/value/MatMul" -> "775 bert/encoder/layer_4/attention/self/value/BiasAdd" [label="[]", style=solid]; "775 bert/encoder/layer_4/attention/self/value/BiasAdd" -> "776 bert/encoder/layer_4/attention/self/Reshape_2" [label="[]", style=solid]; "776 bert/encoder/layer_4/attention/self/Reshape_2" -> "777 bert/encoder/layer_4/attention/self/transpose_2" [label="[]", style=solid]; @@ -2511,18 +2511,18 @@ strict digraph { "852 bert/encoder/layer_4/output/LayerNorm/batchnorm/mul_2" -> "853 bert/encoder/layer_4/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "853 bert/encoder/layer_4/output/LayerNorm/batchnorm/sub" -> "855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "854 bert/encoder/layer_4/output/LayerNorm/batchnorm/mul_1" -> "855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "858 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "862 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "855 bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1" -> "898 bert/encoder/layer_5/attention/output/add" [label="[]", style=solid]; -"856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" -> "857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" -> "878 bert/encoder/layer_5/attention/self/key/MatMul" [label="[]", style=solid]; -"858 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" -> "859 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"859 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" -> "870 bert/encoder/layer_5/attention/self/query/MatMul" [label="[]", style=solid]; -"860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" -> "861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" -> "864 bert/encoder/layer_5/attention/self/value/MatMul" [label="[]", style=solid]; -"862 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" -> "863 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"863 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" -> "864 bert/encoder/layer_5/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"856 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" -> "857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"857 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_1" -> "864 bert/encoder/layer_5/attention/self/value/MatMul" [label="[]", style=solid]; +"858 QuantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" -> "859 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"859 DequantizeLinear_bert/encoder/layer_5/attention/self/value/kernel^0_1" -> "864 bert/encoder/layer_5/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"860 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" -> "861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"861 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_2" -> "870 bert/encoder/layer_5/attention/self/query/MatMul" [label="[]", style=solid]; +"862 QuantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" -> "863 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"863 DequantizeLinear_bert/encoder/layer_4/output/LayerNorm/batchnorm/add_1^0_3" -> "878 bert/encoder/layer_5/attention/self/key/MatMul" [label="[]", style=solid]; "864 bert/encoder/layer_5/attention/self/value/MatMul" -> "865 bert/encoder/layer_5/attention/self/value/BiasAdd" [label="[]", style=solid]; "865 bert/encoder/layer_5/attention/self/value/BiasAdd" -> "866 bert/encoder/layer_5/attention/self/Reshape_2" [label="[]", style=solid]; "866 bert/encoder/layer_5/attention/self/Reshape_2" -> "867 bert/encoder/layer_5/attention/self/transpose_2" [label="[]", style=solid]; @@ -2615,18 +2615,18 @@ strict digraph { "942 bert/encoder/layer_5/output/LayerNorm/batchnorm/mul_2" -> "943 bert/encoder/layer_5/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "943 bert/encoder/layer_5/output/LayerNorm/batchnorm/sub" -> "945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "944 bert/encoder/layer_5/output/LayerNorm/batchnorm/mul_1" -> "945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "948 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "952 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "945 bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1" -> "988 bert/encoder/layer_6/attention/output/add" [label="[]", style=solid]; -"946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" -> "947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" -> "968 bert/encoder/layer_6/attention/self/key/MatMul" [label="[]", style=solid]; -"948 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" -> "949 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"949 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" -> "960 bert/encoder/layer_6/attention/self/query/MatMul" [label="[]", style=solid]; -"950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" -> "951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" -> "954 bert/encoder/layer_6/attention/self/value/MatMul" [label="[]", style=solid]; -"952 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" -> "953 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"953 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" -> "954 bert/encoder/layer_6/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"946 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" -> "947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"947 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_1" -> "954 bert/encoder/layer_6/attention/self/value/MatMul" [label="[]", style=solid]; +"948 QuantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" -> "949 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"949 DequantizeLinear_bert/encoder/layer_6/attention/self/value/kernel^0_1" -> "954 bert/encoder/layer_6/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"950 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" -> "951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"951 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_2" -> "960 bert/encoder/layer_6/attention/self/query/MatMul" [label="[]", style=solid]; +"952 QuantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" -> "953 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"953 DequantizeLinear_bert/encoder/layer_5/output/LayerNorm/batchnorm/add_1^0_3" -> "968 bert/encoder/layer_6/attention/self/key/MatMul" [label="[]", style=solid]; "954 bert/encoder/layer_6/attention/self/value/MatMul" -> "955 bert/encoder/layer_6/attention/self/value/BiasAdd" [label="[]", style=solid]; "955 bert/encoder/layer_6/attention/self/value/BiasAdd" -> "956 bert/encoder/layer_6/attention/self/Reshape_2" [label="[]", style=solid]; "956 bert/encoder/layer_6/attention/self/Reshape_2" -> "957 bert/encoder/layer_6/attention/self/transpose_2" [label="[]", style=solid]; @@ -2719,18 +2719,18 @@ strict digraph { "1032 bert/encoder/layer_6/output/LayerNorm/batchnorm/mul_2" -> "1033 bert/encoder/layer_6/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "1033 bert/encoder/layer_6/output/LayerNorm/batchnorm/sub" -> "1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "1034 bert/encoder/layer_6/output/LayerNorm/batchnorm/mul_1" -> "1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1038 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1042 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "1035 bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1" -> "1078 bert/encoder/layer_7/attention/output/add" [label="[]", style=solid]; -"1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" -> "1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" -> "1058 bert/encoder/layer_7/attention/self/key/MatMul" [label="[]", style=solid]; -"1038 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" -> "1039 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"1039 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" -> "1050 bert/encoder/layer_7/attention/self/query/MatMul" [label="[]", style=solid]; -"1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" -> "1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" -> "1044 bert/encoder/layer_7/attention/self/value/MatMul" [label="[]", style=solid]; -"1042 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" -> "1043 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"1043 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" -> "1044 bert/encoder/layer_7/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1036 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" -> "1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"1037 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_1" -> "1044 bert/encoder/layer_7/attention/self/value/MatMul" [label="[]", style=solid]; +"1038 QuantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" -> "1039 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"1039 DequantizeLinear_bert/encoder/layer_7/attention/self/value/kernel^0_1" -> "1044 bert/encoder/layer_7/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1040 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" -> "1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"1041 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_2" -> "1050 bert/encoder/layer_7/attention/self/query/MatMul" [label="[]", style=solid]; +"1042 QuantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" -> "1043 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"1043 DequantizeLinear_bert/encoder/layer_6/output/LayerNorm/batchnorm/add_1^0_3" -> "1058 bert/encoder/layer_7/attention/self/key/MatMul" [label="[]", style=solid]; "1044 bert/encoder/layer_7/attention/self/value/MatMul" -> "1045 bert/encoder/layer_7/attention/self/value/BiasAdd" [label="[]", style=solid]; "1045 bert/encoder/layer_7/attention/self/value/BiasAdd" -> "1046 bert/encoder/layer_7/attention/self/Reshape_2" [label="[]", style=solid]; "1046 bert/encoder/layer_7/attention/self/Reshape_2" -> "1047 bert/encoder/layer_7/attention/self/transpose_2" [label="[]", style=solid]; @@ -2823,18 +2823,18 @@ strict digraph { "1122 bert/encoder/layer_7/output/LayerNorm/batchnorm/mul_2" -> "1123 bert/encoder/layer_7/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "1123 bert/encoder/layer_7/output/LayerNorm/batchnorm/sub" -> "1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "1124 bert/encoder/layer_7/output/LayerNorm/batchnorm/mul_1" -> "1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1128 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1132 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "1125 bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1" -> "1168 bert/encoder/layer_8/attention/output/add" [label="[]", style=solid]; -"1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" -> "1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" -> "1148 bert/encoder/layer_8/attention/self/key/MatMul" [label="[]", style=solid]; -"1128 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" -> "1129 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"1129 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" -> "1140 bert/encoder/layer_8/attention/self/query/MatMul" [label="[]", style=solid]; -"1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" -> "1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" -> "1134 bert/encoder/layer_8/attention/self/value/MatMul" [label="[]", style=solid]; -"1132 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" -> "1133 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"1133 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" -> "1134 bert/encoder/layer_8/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1126 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" -> "1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"1127 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_1" -> "1134 bert/encoder/layer_8/attention/self/value/MatMul" [label="[]", style=solid]; +"1128 QuantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" -> "1129 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"1129 DequantizeLinear_bert/encoder/layer_8/attention/self/value/kernel^0_1" -> "1134 bert/encoder/layer_8/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1130 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" -> "1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"1131 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_2" -> "1140 bert/encoder/layer_8/attention/self/query/MatMul" [label="[]", style=solid]; +"1132 QuantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" -> "1133 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"1133 DequantizeLinear_bert/encoder/layer_7/output/LayerNorm/batchnorm/add_1^0_3" -> "1148 bert/encoder/layer_8/attention/self/key/MatMul" [label="[]", style=solid]; "1134 bert/encoder/layer_8/attention/self/value/MatMul" -> "1135 bert/encoder/layer_8/attention/self/value/BiasAdd" [label="[]", style=solid]; "1135 bert/encoder/layer_8/attention/self/value/BiasAdd" -> "1136 bert/encoder/layer_8/attention/self/Reshape_2" [label="[]", style=solid]; "1136 bert/encoder/layer_8/attention/self/Reshape_2" -> "1137 bert/encoder/layer_8/attention/self/transpose_2" [label="[]", style=solid]; @@ -2927,18 +2927,18 @@ strict digraph { "1212 bert/encoder/layer_8/output/LayerNorm/batchnorm/mul_2" -> "1213 bert/encoder/layer_8/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "1213 bert/encoder/layer_8/output/LayerNorm/batchnorm/sub" -> "1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "1214 bert/encoder/layer_8/output/LayerNorm/batchnorm/mul_1" -> "1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1218 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1222 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "1215 bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1" -> "1258 bert/encoder/layer_9/attention/output/add" [label="[]", style=solid]; -"1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" -> "1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" -> "1238 bert/encoder/layer_9/attention/self/key/MatMul" [label="[]", style=solid]; -"1218 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" -> "1219 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"1219 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" -> "1230 bert/encoder/layer_9/attention/self/query/MatMul" [label="[]", style=solid]; -"1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" -> "1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" -> "1224 bert/encoder/layer_9/attention/self/value/MatMul" [label="[]", style=solid]; -"1222 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" -> "1223 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"1223 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" -> "1224 bert/encoder/layer_9/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1216 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" -> "1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"1217 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_1" -> "1224 bert/encoder/layer_9/attention/self/value/MatMul" [label="[]", style=solid]; +"1218 QuantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" -> "1219 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"1219 DequantizeLinear_bert/encoder/layer_9/attention/self/value/kernel^0_1" -> "1224 bert/encoder/layer_9/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1220 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" -> "1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"1221 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_2" -> "1230 bert/encoder/layer_9/attention/self/query/MatMul" [label="[]", style=solid]; +"1222 QuantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" -> "1223 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"1223 DequantizeLinear_bert/encoder/layer_8/output/LayerNorm/batchnorm/add_1^0_3" -> "1238 bert/encoder/layer_9/attention/self/key/MatMul" [label="[]", style=solid]; "1224 bert/encoder/layer_9/attention/self/value/MatMul" -> "1225 bert/encoder/layer_9/attention/self/value/BiasAdd" [label="[]", style=solid]; "1225 bert/encoder/layer_9/attention/self/value/BiasAdd" -> "1226 bert/encoder/layer_9/attention/self/Reshape_2" [label="[]", style=solid]; "1226 bert/encoder/layer_9/attention/self/Reshape_2" -> "1227 bert/encoder/layer_9/attention/self/transpose_2" [label="[]", style=solid]; @@ -3031,18 +3031,18 @@ strict digraph { "1302 bert/encoder/layer_9/output/LayerNorm/batchnorm/mul_2" -> "1303 bert/encoder/layer_9/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "1303 bert/encoder/layer_9/output/LayerNorm/batchnorm/sub" -> "1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "1304 bert/encoder/layer_9/output/LayerNorm/batchnorm/mul_1" -> "1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1308 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1312 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "1305 bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1" -> "1348 bert/encoder/layer_10/attention/output/add" [label="[]", style=solid]; -"1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" -> "1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" -> "1328 bert/encoder/layer_10/attention/self/key/MatMul" [label="[]", style=solid]; -"1308 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" -> "1309 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"1309 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" -> "1320 bert/encoder/layer_10/attention/self/query/MatMul" [label="[]", style=solid]; -"1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" -> "1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" -> "1314 bert/encoder/layer_10/attention/self/value/MatMul" [label="[]", style=solid]; -"1312 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" -> "1313 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"1313 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" -> "1314 bert/encoder/layer_10/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1306 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" -> "1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"1307 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_1" -> "1314 bert/encoder/layer_10/attention/self/value/MatMul" [label="[]", style=solid]; +"1308 QuantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" -> "1309 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"1309 DequantizeLinear_bert/encoder/layer_10/attention/self/value/kernel^0_1" -> "1314 bert/encoder/layer_10/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1310 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" -> "1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"1311 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_2" -> "1320 bert/encoder/layer_10/attention/self/query/MatMul" [label="[]", style=solid]; +"1312 QuantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" -> "1313 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"1313 DequantizeLinear_bert/encoder/layer_9/output/LayerNorm/batchnorm/add_1^0_3" -> "1328 bert/encoder/layer_10/attention/self/key/MatMul" [label="[]", style=solid]; "1314 bert/encoder/layer_10/attention/self/value/MatMul" -> "1315 bert/encoder/layer_10/attention/self/value/BiasAdd" [label="[]", style=solid]; "1315 bert/encoder/layer_10/attention/self/value/BiasAdd" -> "1316 bert/encoder/layer_10/attention/self/Reshape_2" [label="[]", style=solid]; "1316 bert/encoder/layer_10/attention/self/Reshape_2" -> "1317 bert/encoder/layer_10/attention/self/transpose_2" [label="[]", style=solid]; @@ -3135,18 +3135,18 @@ strict digraph { "1392 bert/encoder/layer_10/output/LayerNorm/batchnorm/mul_2" -> "1393 bert/encoder/layer_10/output/LayerNorm/batchnorm/sub" [label="[]", style=solid]; "1393 bert/encoder/layer_10/output/LayerNorm/batchnorm/sub" -> "1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; "1394 bert/encoder/layer_10/output/LayerNorm/batchnorm/mul_1" -> "1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" [label="[]", style=solid]; -"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; -"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1398 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; -"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=solid]; +"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=solid]; +"1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1402 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=solid]; "1395 bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1" -> "1438 bert/encoder/layer_11/attention/output/add" [label="[]", style=solid]; -"1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" -> "1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; -"1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" -> "1418 bert/encoder/layer_11/attention/self/key/MatMul" [label="[]", style=solid]; -"1398 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" -> "1399 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; -"1399 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" -> "1410 bert/encoder/layer_11/attention/self/query/MatMul" [label="[]", style=solid]; -"1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" -> "1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; -"1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" -> "1404 bert/encoder/layer_11/attention/self/value/MatMul" [label="[]", style=solid]; -"1402 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" -> "1403 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; -"1403 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" -> "1404 bert/encoder/layer_11/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1396 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" -> "1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" [label="[]", style=dashed]; +"1397 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_1" -> "1404 bert/encoder/layer_11/attention/self/value/MatMul" [label="[]", style=solid]; +"1398 QuantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" -> "1399 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" [label="[768, 768]", style=dashed]; +"1399 DequantizeLinear_bert/encoder/layer_11/attention/self/value/kernel^0_1" -> "1404 bert/encoder/layer_11/attention/self/value/MatMul" [label="[768, 768]", style=solid]; +"1400 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" -> "1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" [label="[]", style=dashed]; +"1401 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_2" -> "1410 bert/encoder/layer_11/attention/self/query/MatMul" [label="[]", style=solid]; +"1402 QuantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" -> "1403 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" [label="[]", style=dashed]; +"1403 DequantizeLinear_bert/encoder/layer_10/output/LayerNorm/batchnorm/add_1^0_3" -> "1418 bert/encoder/layer_11/attention/self/key/MatMul" [label="[]", style=solid]; "1404 bert/encoder/layer_11/attention/self/value/MatMul" -> "1405 bert/encoder/layer_11/attention/self/value/BiasAdd" [label="[]", style=solid]; "1405 bert/encoder/layer_11/attention/self/value/BiasAdd" -> "1406 bert/encoder/layer_11/attention/self/Reshape_2" [label="[]", style=solid]; "1406 bert/encoder/layer_11/attention/self/Reshape_2" -> "1407 bert/encoder/layer_11/attention/self/transpose_2" [label="[]", style=solid]; diff --git a/tests/onnx/data/reference_graphs/quantization/retinanet-9.dot b/tests/onnx/data/reference_graphs/quantization/retinanet-9.dot index 7aa64281d32..9d2f66780d5 100644 --- a/tests/onnx/data/reference_graphs/quantization/retinanet-9.dot +++ b/tests/onnx/data/reference_graphs/quantization/retinanet-9.dot @@ -855,10 +855,10 @@ strict digraph { "853 QuantizeLinear_backbones.ResNet101FPN.pyramid6.weight_1" [id=853, type=QuantizeLinear]; "854 DequantizeLinear_backbones.ResNet101FPN.pyramid6.weight_1" [id=854, type=DequantizeLinear]; "855 Conv_349" [id=855, type=Conv]; -"856 QuantizeLinear_1028_2" [id=856, type=QuantizeLinear]; -"857 DequantizeLinear_1028_2" [id=857, type=DequantizeLinear]; -"858 QuantizeLinear_1028_1" [id=858, type=QuantizeLinear]; -"859 DequantizeLinear_1028_1" [id=859, type=DequantizeLinear]; +"856 QuantizeLinear_1028_1" [id=856, type=QuantizeLinear]; +"857 DequantizeLinear_1028_1" [id=857, type=DequantizeLinear]; +"858 QuantizeLinear_1028_2" [id=858, type=QuantizeLinear]; +"859 DequantizeLinear_1028_2" [id=859, type=DequantizeLinear]; "860 Relu_350" [id=860, type=Relu]; "861 QuantizeLinear_1029_1" [id=861, type=QuantizeLinear]; "862 DequantizeLinear_1029_1" [id=862, type=DequantizeLinear]; @@ -1983,13 +1983,13 @@ strict digraph { "852 Add_348" -> "866 QuantizeLinear_1027_1" [label="[1, 256, 60, 80]", style=solid]; "853 QuantizeLinear_backbones.ResNet101FPN.pyramid6.weight_1" -> "854 DequantizeLinear_backbones.ResNet101FPN.pyramid6.weight_1" [label="[256, 2048, 3, 3]", style=dashed]; "854 DequantizeLinear_backbones.ResNet101FPN.pyramid6.weight_1" -> "855 Conv_349" [label="[256, 2048, 3, 3]", style=solid]; -"855 Conv_349" -> "856 QuantizeLinear_1028_2" [label="[1, 256, 8, 10]", style=solid]; -"855 Conv_349" -> "858 QuantizeLinear_1028_1" [label="[1, 256, 8, 10]", style=solid]; +"855 Conv_349" -> "856 QuantizeLinear_1028_1" [label="[1, 256, 8, 10]", style=solid]; +"855 Conv_349" -> "858 QuantizeLinear_1028_2" [label="[1, 256, 8, 10]", style=solid]; "855 Conv_349" -> "860 Relu_350" [label="[1, 256, 8, 10]", style=solid]; -"856 QuantizeLinear_1028_2" -> "857 DequantizeLinear_1028_2" [label="[1, 256, 8, 10]", style=dashed]; -"857 DequantizeLinear_1028_2" -> "1041 Conv_427" [label="[1, 256, 8, 10]", style=solid]; -"858 QuantizeLinear_1028_1" -> "859 DequantizeLinear_1028_1" [label="[1, 256, 8, 10]", style=dashed]; -"859 DequantizeLinear_1028_1" -> "944 Conv_382" [label="[1, 256, 8, 10]", style=solid]; +"856 QuantizeLinear_1028_1" -> "857 DequantizeLinear_1028_1" [label="[1, 256, 8, 10]", style=dashed]; +"857 DequantizeLinear_1028_1" -> "944 Conv_382" [label="[1, 256, 8, 10]", style=solid]; +"858 QuantizeLinear_1028_2" -> "859 DequantizeLinear_1028_2" [label="[1, 256, 8, 10]", style=dashed]; +"859 DequantizeLinear_1028_2" -> "1041 Conv_427" [label="[1, 256, 8, 10]", style=solid]; "860 Relu_350" -> "861 QuantizeLinear_1029_1" [label="[1, 256, 8, 10]", style=solid]; "861 QuantizeLinear_1029_1" -> "862 DequantizeLinear_1029_1" [label="[1, 256, 8, 10]", style=dashed]; "862 DequantizeLinear_1029_1" -> "865 Conv_351" [label="[1, 256, 8, 10]", style=solid];