op_name                                              sensitive_type    op_type                                                                           L1  quant_dtype    flops
---------------------------------------------------  ----------------  --------------------------------------------------------------------------  --------  -------------  ---------------------
head.layers.20.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.872621  qint16         0(0%)
backbone.stages.1.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.85173   qint8          506068992(0.02%)
head.anchor_encoder.size_fc.0                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851474  qint8          8306688(0.00%)
backbone.stages.1.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.85144   qint8          506068992(0.02%)
head.anchor_encoder.size_fc.0                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851351  qint8          8306688(0.00%)
head.layers.24.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851315  qint16         0(0%)
head.layers.0.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.85122   qint16         0(0%)
head.layers.9.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851206  qint8          0(0%)
head.layers.3.feat_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851187  qint8          0(0%)
head.layers.38.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851186  qint16         0(0%)
head.layers.41.cls_layers.2.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851161  qint16         0(0%)
head.layers.25.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851148  qint8          3489660928(0.16%)
head.layers.22.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851135  qint8          3489660928(0.16%)
head.layers.22.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851135  qint8          3489660928(0.16%)
head.layers.6.layers.9.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.8.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.41.cls_layers.0                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.18.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.3.camera_encoder.2.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
backbone.stages.2.block.5.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
backbone.stages.1.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.10.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
head.layers.0.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.layers.5                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.40.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.27.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.layers.4.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.anchor_encoder.pos_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.patch_embed.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.851133  qint8          6072827904(0.28%)
head.layers.12.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.1.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.27.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         37486592(0.00%)
head.layers.3.camera_encoder.0                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.20.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.layers.7.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.5.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.1.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.33.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.35.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.9.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.12.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.31.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.39.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.39.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.28.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.15.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
backbone.stage_norm.1                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.851133  qint8          0(0%)
head.layers.21.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.18.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.1.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.5.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.0.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.37.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.33.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.37.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.22.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.27.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.7.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
backbone.stages.1.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.14.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.27.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.layers.35.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.2.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.vel_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.2.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.anchor_encoder.yaw_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.4.pre_norm.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.31.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.31.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.36.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.19.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.34.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
backbone.stage_norm.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.851133  qint8          0(0%)
head.layers.41.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          16613376(0.00%)
head.layers.41.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.25.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.26.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.18.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
backbone.stages.2.block.4.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.22.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
head.layers.7.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.downsample_block.1.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          10796138496(0.49%)
head.layers.31.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.23.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.0                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.38.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.0.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.layers.15.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.2.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.2.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
backbone.stages.2.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.17.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.0.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.39.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.38.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.anchor_encoder.yaw_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.anchor_encoder.yaw_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.39.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.15.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.35.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.pos_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.20.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.quality_layers.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.pos_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.instance_bank.feature_cat                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.35.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.4.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.vel_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.3.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
backbone.stages.3.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.34.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.5.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.0.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
backbone.stages.2.block.0.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.9                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.38.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.short_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.3.weights_fc                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.8.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.29.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.layers.10                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         37486592(0.00%)
head.layers.21.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.5.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.14.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.layers.7                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
backbone.stages.3.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.14.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.15.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.39.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.5.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.anchor_encoder.size_fc.2.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.36.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.36.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.0.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.29.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.10.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.point_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.32.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.anchor_encoder.yaw_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
head.anchor_encoder.yaw_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.41.quality_layers.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.30.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.8.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.24.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.7.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.3.output_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
backbone.stages.2.block.4.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.13.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.cam_add                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.0.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.27.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.quality_layers.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.24.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.33.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         37486592(0.00%)
head.layers.20.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.fc_before                                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          15703474176(0.72%)
head.anchor_encoder.yaw_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.6.layers.2                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.20.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.35.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.31.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.14.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.34.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.0.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.4.pre_norm.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.13.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
backbone.stages.2.block.6.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.35.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.9.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.point_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.anchor_encoder.size_fc.8.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.instance_bank.instance_feature_quant_stub       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint8          0(0%)
head.layers.5.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.36.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.14.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.32.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
backbone.stages.3.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.22.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.1.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.17.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.4.pre_norm.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.2.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.8.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.14.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.weight_softmax.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.downsample_block.0.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
backbone.downsample_block.0.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.anchor_encoder.pos_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.3.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
backbone.stages.0.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.layers.24.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.cls_layers.3                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
backbone.stages.2.block.5.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.6.layers.4.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.7.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.4.pre_norm.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.41.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
backbone.stages.0.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.5.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.36.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.4.layers.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.anchor_encoder.size_fc.5.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
backbone.stages.2.block.1.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.41.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          13631488(0.00%)
head.layers.4.pre_norm.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.7.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.38.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.pos_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.27.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.20.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.3                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.29.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.29.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.39.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.2.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.2.block.6.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.31.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.3.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.23.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.27.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.16.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.38.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.15.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.16.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
neck.conv_add.2                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          28789702656(1.31%)
head.layers.22.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.layers.9.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.31.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.34.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.anchor_encoder.pos_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
head.layers.41.cls_layers.2.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.5.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.32.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.15.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.2.block.7.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.11.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.2.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.37.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.13.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.cls_layers.2.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
head.layers.27.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.15.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.3.kps_generator.keypoints_add            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.layers.5                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.25.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.17.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.2.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.9.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.19.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.32.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.17.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          81788928(0.00%)
backbone.stages.0.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.10.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.22.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.29.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.37.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stage_norm.2                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.851133  qint8          0(0%)
head.layers.3.camera_encoder.5.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.40.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.20.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
neck.conv_add.0                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          5398069248(0.25%)
head.anchor_encoder.size_fc.8.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.35.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.30.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.34.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.24.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.anchor_encoder.size_fc.8.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.3.camera_encoder.0                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.24.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.20.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         37486592(0.00%)
head.anchor_encoder.size_fc.6                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.anchor_encoder.yaw_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.14.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.14.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.20.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.21.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.21.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.24.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.37.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.16.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.pos_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.camera_encoder.2.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.layers.41.quality_layers.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          506068992(0.02%)
backbone.stages.1.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          506068992(0.02%)
head.layers.18.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.7.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.21.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.34.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.10.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.1.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.27.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.2.block.2.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.13.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.anchor_encoder.yaw_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.2.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
backbone.stages.2.block.7.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.41.quality_layers.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.1.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
backbone.stages.2.block.1.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.31.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          5537792(0.00%)
head.anchor_encoder.yaw_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          5537792(0.00%)
head.layers.18.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.24.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.layers.9.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.10.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.1.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
backbone.stages.1.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
head.layers.21.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.38.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.3.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
backbone.stages.2.block.3.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.11.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.31.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.18.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.1.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.3.weight_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.22.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.quality_layers.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.1.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
backbone.stages.1.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.0.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.13.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.11.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.24.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.21.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.0.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.3.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
backbone.stages.3.block.4.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
backbone.stages.3.block.4.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.20.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.33.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.18.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.vel_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.11.scale_quant_stub             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
neck.conv_extract.3.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          2699034624(0.12%)
neck.conv_extract.3.0                                weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          2699034624(0.12%)
head.layers.5.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.14.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.20.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.quality_layers.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.2.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.38.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
backbone.stages.3.block.5.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.1.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.17.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.4.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.26.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.layers.0                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.6.layers.0                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.13.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.2.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.17.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.36.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.36.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.27.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.3.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
backbone.stages.3.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
head.layers.8.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.8.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.2.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.quality_layers.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.34.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.38.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
backbone.stages.2.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.2.block.1.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
backbone.stages.2.block.1.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.38.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.yaw_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          16613376(0.00%)
head.layers.21.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.5.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.11.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.23.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.2.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.10.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.16.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.17.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.21.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.21.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.3.camera_encoder.2.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.11.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.pre_norm.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
head.layers.13.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.20.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.4.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.10.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.10.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.41.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.3.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.24.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.32.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
backbone.stages.2.block.6.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.26.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.8.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.instance_bank.anchor_cat                        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.layers.9.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.33.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.3.camera_encoder.5.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.34.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.19.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.16.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
backbone.stages.3.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.35.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.2.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.36.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.0.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.layers.41.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.10.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.0.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.0.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.41.quality_layers.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.19.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.layers.16.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.37.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.22.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.0.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.20.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.13.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.13.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.34.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.34.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.3.kps_generator.offset                   activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.anchor_encoder.vel_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.3.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
backbone.stages.2.block.3.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.31.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.38.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.31.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.13.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.3.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.25.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.25.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.12.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.28.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.35.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
backbone.stages.1.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
backbone.stages.1.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.38.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.anchor_encoder.vel_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.21.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.36.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.27.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.14.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.25.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.patch_embed.1.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.851133  qint8          32388415488(1.48%)
backbone.patch_embed.1.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.851133  qint8          32388415488(1.48%)
head.layers.13.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.13.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.18.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
backbone.stages.3.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.41.quality_layers.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.15.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.16.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
head.anchor_encoder.vel_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.31.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.16.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.33.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.instance_bank.anchor_quant_stub                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.20.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.10.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.11.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.29.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.28.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.4.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.26.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.9                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.anchor_encoder.vel_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
head.anchor_encoder.vel_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
head.anchor_encoder.vel_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.33.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.short_add                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.13.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.layers.10                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.layers.18.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.13.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.23.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.23.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.40.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.15.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.20.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.4.pre_norm.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.9.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.25.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.15.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.15.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.31.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.36.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.15.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.layers.1                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.4.layers.1                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.10.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.7.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.25.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.25.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.36.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.3                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.anchor_encoder.vel_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
head.layers.7.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
neck.conv_add.1                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          14394851328(0.66%)
neck.conv_add.1                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          14394851328(0.66%)
head.anchor_encoder.pos_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.3.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
backbone.stages.3.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          94887936(0.00%)
head.layers.24.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.3.block.5.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
head.layers.41.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.25.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.35.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.0.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.1.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.cls_layers.2.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.38.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          33226752(0.00%)
backbone.stages.0.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.38.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.31.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.27.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.6.add2                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.quality_layers.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.2.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.11.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.13.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.0.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.17.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.11.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.27.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
head.layers.39.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.37.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.2.block.2.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
backbone.stages.2.block.2.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.38.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.layers.34.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.anchor_encoder.vel_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.12.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.5.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.22.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.27.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.25.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.2.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.17.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.20.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.28.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.33.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.7.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.20.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.24.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          6815744(0.00%)
head.layers.17.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.27.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.31.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.39.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.quality_layers.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.25.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.31.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          81788928(0.00%)
head.anchor_encoder.size_fc.11.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.quality_layers.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.0.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
backbone.stages.0.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.layers.10.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.22.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.24.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.yaw_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.1.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.27.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.13.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.anchor_encoder.vel_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.5.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.26.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.30.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
neck.fpn_conv.2.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          64776830976(2.95%)
neck.fpn_conv.2.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          64776830976(2.95%)
head.layers.6.layers.2                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.9.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.12.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.11.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.4.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.31.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.4.pre_norm.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.13.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.20.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.20.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.36.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.35.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.27.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.27.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.41.cls_layers.5.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.3.camera_encoder.2.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.14.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.38.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.28.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.28.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.41.cls_layers.5.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.downsample_block.2.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.41.quality_layers.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.22.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.2.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.28.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
backbone.stages.1.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          506068992(0.02%)
head.layers.38.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.11.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.2.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.11.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.11.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.20.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
backbone.stages.0.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.anchor_encoder.vel_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.1.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.17.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.8.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.7.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.8.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.5.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.2.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.32.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.32.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.7.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.1.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.41.cls_layers.5.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.17.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.26.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.18.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.19.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.41.quality_layers.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.25.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.5.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.5.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.17.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.41.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.12.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.13.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.15.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.32.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.36.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.17.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.24.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.24.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          479232(0.00%)
head.layers.36.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.layers.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.16.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.cls_layers.2.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.23.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.36.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.40.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.25.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.4.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.size_fc.2.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.34.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.30.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.0.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.13.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.34.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.7.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.41.cls_layers.5.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.40.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.20.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.quality_layers.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.6.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.2.block.6.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.anchor_encoder.pos_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.13.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.5.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.3.point_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.3                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.11.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.10.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
neck.fpn_conv.3.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          16194207744(0.74%)
neck.fpn_conv.3.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          16194207744(0.74%)
head.layers.4.pre_norm.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.35.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.35.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.30.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.38.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.quality_layers.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
neck.fpn_conv.0.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1036429295616(47.24%)
neck.fpn_conv.0.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1036429295616(47.24%)
head.layers.27.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.17.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.0.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
backbone.stages.3.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.41.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.41.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.24.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.8.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.reciprocal_op.reciprocal               activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.10.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.40.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.quality_layers.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.9.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.6.layers.9.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.4.pre_norm.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.2.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.8.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.39.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.30.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.11.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.12.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.7.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.point_quant_stub                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.30.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.34.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          81788928(0.00%)
head.layers.13.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.33.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.23.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.layers.10.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          81788928(0.00%)
head.layers.32.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.30.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          1417674752(0.06%)
head.layers.0.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.9.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.camera_encoder.2.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.10.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.28.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.28.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.10.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.39.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.39.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.38.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.41.cls_layers.5.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.26.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.20.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.1.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.1.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
backbone.stages.2.block.7.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.11.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.14.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.4.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
backbone.stages.3.block.4.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
head.layers.32.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.26.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.28.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.8.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.8.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.6.quality_layers.6                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.2.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.37.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.18.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.10.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.36.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.39.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.35.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.6.add1                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.33.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.31.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.9.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.27.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.layers.17.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.17.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.layers.17.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.35.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.41.cls_layers.5.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.15.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
backbone.stages.3.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.0.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
backbone.stages.0.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
neck.conv_add.0                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          5398069248(0.25%)
head.anchor_encoder.pos_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.patch_embed.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.851133  qint8          6072827904(0.28%)
head.layers.35.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.weight_softmax.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.6                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          88604672(0.00%)
head.layers.10.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.20.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.19.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.layers.8.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.8.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.22.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.24.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.27.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.14.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.31.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.23.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.quality_layers.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.21.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.4.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.15.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.38.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.14.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.15.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.6.layers.4.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.2.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
backbone.stages.3.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
head.layers.20.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
backbone.stages.2.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
backbone.stages.0.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.layers.3.weight_softmax.sum                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.feat_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.size_fc.2.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.5.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.2.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.3.camera_encoder.5.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.39.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.2.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.40.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.3.residual_op                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.31.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.13.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.38.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.11.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.41.quality_layers.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.20.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.17.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.downsample_block.1.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          10796138496(0.49%)
head.layers.13.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.15.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.29.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.2.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.3.kps_generator.offset                   weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          81788928(0.00%)
head.layers.41.cls_layers.2.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.30.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.9.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.8.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.anchor_encoder.vel_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.15.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.21.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          1744830464(0.08%)
head.layers.29.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.24.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.29.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.layers.11.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
backbone.stage_norm.3                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.851133  qint8          0(0%)
head.layers.39.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.39.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          6979321856(0.32%)
head.layers.19.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.9.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.25.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.37.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.40.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.3.weights_fc                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.34.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.13.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
backbone.quant                                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint8          0(0%)
head.anchor_encoder.pos_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.layers.7                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.15.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.34.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.30.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.3                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          10223616(0.00%)
head.anchor_encoder.vel_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          354418688(0.02%)
head.layers.26.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.5.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
backbone.stages.3.block.5.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          12145655808(0.55%)
head.layers.3.weight_softmax.exp                     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          33226752(0.00%)
head.layers.41.quality_layers.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.9.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.5.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.mat_quant_stub                                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.11.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.20.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.14.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.27.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.37.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.20.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.36.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.0.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          105431040(0.00%)
head.layers.24.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.31.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.layers.32.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.32.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.851133  qint8          1744830464(0.08%)
head.layers.26.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.21.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.2.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.5.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.fc_before                                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         15703474176(0.72%)
head.anchor_encoder.vel_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.feat_sum                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.13.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
head.layers.32.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.35.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.24.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.22.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.camera_encoder.5.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.27.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.quality_layers.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.downsample_block.2.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.layers.5.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.1.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          506068992(0.02%)
head.layers.11.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.3.weight_softmax.reciprocal.reciprocal   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.36.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.38.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1744830464(0.08%)
head.layers.24.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.29.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          1308622848(0.06%)
head.layers.34.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
backbone.stages.3.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.22.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.23.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.3.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint8          0(0%)
head.layers.41.cls_layers.2.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.38.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          1012137984(0.05%)
head.anchor_encoder.vel_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.7.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.3                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.29.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.40.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.32.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.28.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.3.point_matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.5.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.6.layers.4.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.5.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.1.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.34.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851133  qint8          872415232(0.04%)
head.layers.38.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          872415232(0.04%)
head.layers.3.camera_encoder.5.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.17.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.3.camera_encoder.2.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.12.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.cls_layers.5.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.anchor_encoder.vel_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.10.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.22.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.0.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          7197425664(0.33%)
head.layers.1.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.vel_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.6.cls_layers.6                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          0(0%)
backbone.stages.0.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851133  qint8          14394851328(0.66%)
head.layers.28.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.36.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint8          0(0%)
head.layers.12.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.34.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         37486592(0.00%)
head.layers.34.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          37486592(0.00%)
head.layers.20.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
neck.conv_add.2                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          28789702656(1.31%)
head.layers.10.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.35.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.41.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.41.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.pos_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.40.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.24.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint16         81788928(0.00%)
head.layers.41.cls_layers.5.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851133  qint16         0(0%)
head.layers.34.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.28.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851133  qint8          3489660928(0.16%)
head.layers.31.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.anchor_encoder.size_fc.11.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
head.layers.8.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.9.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.12.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851133  qint16         0(0%)
head.layers.13.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851133  qint16         0(0%)
backbone.stages.2.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851133  qint8          4048551936(0.18%)
head.layers.31.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint8          0(0%)
backbone.stages.1.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.851132  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.41.cls_layers.2.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851132  qint16         0(0%)
head.anchor_encoder.size_fc.11.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.anchor_encoder.size_fc.2.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851132  qint16         0(0%)
head.layers.22.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.32.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851132  qint16         0(0%)
head.layers.20.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.851132  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.19.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint8          0(0%)
head.layers.29.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          3489660928(0.16%)
head.layers.19.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851132  qint16         0(0%)
head.anchor_encoder.pos_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851132  qint16         0(0%)
head.layers.38.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          81788928(0.00%)
head.layers.19.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.3.output_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          872415232(0.04%)
head.layers.39.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.21.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          1744830464(0.08%)
head.layers.36.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851132  qint16         0(0%)
head.layers.7.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.5.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.41.quality_layers.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.10.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851132  qint16         0(0%)
head.layers.23.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851132  qint16         0(0%)
head.layers.22.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
backbone.stages.3.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851132  qint8          12145655808(0.55%)
head.layers.34.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.6.layers.4.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.35.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          1744830464(0.08%)
head.layers.35.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          1744830464(0.08%)
head.layers.24.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.6.layers.9.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.10.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851132  qint8          1308622848(0.06%)
head.layers.15.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.41.quality_layers.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.2.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851132  qint16         0(0%)
head.anchor_encoder.vel_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.24.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.layers.32.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
head.anchor_encoder.vel_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.851132  qint16         0(0%)
backbone.stages.3.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.851132  qint8          12145655808(0.55%)
head.layers.1.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.851132  qint16         0(0%)
head.layers.16.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.85113   qint16         0(0%)
head.layers.9.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.85113   qint16         0(0%)
head.fc_after                                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851129  qint8          20937965568(0.95%)
head.anchor_encoder.pos_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.851129  qint16         0(0%)
head.layers.25.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851124  qint8          3489660928(0.16%)
head.fc_after                                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.851116  qint16         20937965568(0.95%)
neck.fpn_conv.1.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.850545  qint8          259107323904(11.81%)
backbone.stages.3.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.849331  qint8          0(0%)
backbone.stages.0.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.84876   qint8          1012137984(0.05%)
backbone.stages.0.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.846358  qint8          1012137984(0.05%)
head.layers.38.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.843927  qint16         0(0%)
head.layers.22.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.832647  qint8          3489660928(0.16%)
neck.fpn_conv.1.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.829711  qint8          259107323904(11.81%)
head.layers.10.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.822671  qint16         0(0%)
head.layers.22.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.821725  qint8          3489660928(0.16%)
head.layers.24.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.534028  qint16         0(0%)