op_name                                              sensitive_type    op_type                                                                            L1  quant_dtype    flops
---------------------------------------------------  ----------------  --------------------------------------------------------------------------  ---------  -------------  ---------------------
neck.fpn_conv.1.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.024675   qint8          259107323904(11.81%)
backbone.stages.0.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0245322  qint8          1012137984(0.05%)
neck.fpn_conv.1.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244862  qint8          259107323904(11.81%)
backbone.stages.0.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244647  qint8          1012137984(0.05%)
head.anchor_encoder.yaw_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244289  qint16         0(0%)
head.layers.34.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244289  qint16         0(0%)
head.layers.24.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244289  qint16         0(0%)
head.layers.13.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244288  qint8          872415232(0.04%)
head.layers.20.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244288  qint16         0(0%)
head.layers.11.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244287  qint8          0(0%)
head.layers.13.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244287  qint8          872415232(0.04%)
head.layers.24.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244286  qint16         0(0%)
head.layers.24.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244286  qint16         0(0%)
head.layers.38.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244285  qint16         0(0%)
head.layers.41.quality_layers.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244285  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244284  qint8          0(0%)
head.anchor_encoder.vel_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244284  qint8          354418688(0.02%)
backbone.stages.2.block.4.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244282  qint8          105431040(0.00%)
backbone.stages.2.block.4.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244282  qint8          105431040(0.00%)
head.layers.10.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244281  qint8          10223616(0.00%)
head.anchor_encoder.vel_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244281  qint8          354418688(0.02%)
head.layers.4.pre_norm.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.024428   qint16         0(0%)
head.layers.10.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.024428   qint8          10223616(0.00%)
head.layers.41.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244278  qint16         0(0%)
head.layers.28.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244278  qint8          0(0%)
head.layers.8.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244278  qint16         0(0%)
head.layers.22.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244277  qint8          3489660928(0.16%)
head.layers.24.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244276  qint8          0(0%)
backbone.stages.2.block.3.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244275  qint8          0(0%)
head.layers.32.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244275  qint16         0(0%)
head.layers.24.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244275  qint16         0(0%)
head.anchor_encoder.pos_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244273  qint8          1417674752(0.06%)
head.anchor_encoder.pos_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244271  qint8          1417674752(0.06%)
backbone.stages.2.block.6.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.024427   qint8          0(0%)
head.layers.18.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244268  qint16         0(0%)
head.layers.7.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244268  qint8          0(0%)
head.layers.11.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244265  qint16         0(0%)
head.layers.20.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244265  qint8          0(0%)
head.layers.17.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244263  qint16         0(0%)
head.layers.3.camera_encoder.2.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244261  qint16         0(0%)
head.layers.26.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244261  qint8          0(0%)
head.layers.13.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244257  qint16         0(0%)
head.layers.21.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244254  qint8          1744830464(0.08%)
head.layers.8.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244253  qint8          0(0%)
head.layers.31.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244251  qint16         0(0%)
head.layers.21.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244249  qint8          1744830464(0.08%)
head.layers.41.cls_layers.2.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244247  qint16         0(0%)
head.layers.17.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244246  qint8          0(0%)
head.layers.27.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244241  qint16         0(0%)
backbone.downsample_block.1.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244241  qint8          10796138496(0.49%)
backbone.downsample_block.1.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244237  qint8          10796138496(0.49%)
head.layers.16.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244234  qint16         0(0%)
head.instance_bank.anchor_cat                        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244233  qint16         0(0%)
head.layers.41.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244228  qint8          0(0%)
head.layers.10.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244227  qint8          479232(0.00%)
head.layers.6.layers.9.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244225  qint16         0(0%)
head.layers.41.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244224  qint8          0(0%)
head.layers.10.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244222  qint8          479232(0.00%)
head.layers.33.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244222  qint16         0(0%)
head.layers.3.camera_encoder.5.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024422   qint16         0(0%)
head.layers.13.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.024422   qint16         0(0%)
head.layers.14.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244218  qint8          0(0%)
head.layers.31.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244216  qint8          0(0%)
head.layers.9.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244216  qint16         0(0%)
head.layers.23.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244216  qint16         0(0%)
head.layers.34.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244215  qint8          0(0%)
head.anchor_encoder.yaw_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244213  qint8          0(0%)
head.layers.41.cls_layers.0                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244213  qint8          872415232(0.04%)
head.layers.41.cls_layers.0                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244211  qint8          872415232(0.04%)
head.layers.19.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244209  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244207  qint16         0(0%)
head.anchor_encoder.vel_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244204  qint16         0(0%)
head.layers.27.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244201  qint16         0(0%)
head.anchor_encoder.size_fc.11.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244201  qint16         0(0%)
head.layers.16.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244197  qint16         0(0%)
head.layers.31.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244194  qint16         0(0%)
head.layers.41.cls_layers.2.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244193  qint16         0(0%)
backbone.stages.3.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244192  qint8          94887936(0.00%)
backbone.stages.3.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244187  qint8          94887936(0.00%)
head.anchor_encoder.yaw_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244186  qint16         0(0%)
head.layers.35.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244185  qint16         0(0%)
head.anchor_encoder.size_fc.2.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244184  qint16         0(0%)
head.layers.36.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244181  qint8          0(0%)
head.anchor_encoder.size_fc.8.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244179  qint16         0(0%)
head.layers.38.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244179  qint16         0(0%)
head.layers.31.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244179  qint8          0(0%)
backbone.stages.3.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244175  qint8          0(0%)
head.layers.27.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244175  qint16         0(0%)
head.layers.3.residual_op                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244172  qint8          0(0%)
head.layers.26.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244171  qint16         0(0%)
backbone.stages.0.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.024417   qint8          1012137984(0.05%)
head.layers.32.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244169  qint16         0(0%)
backbone.stages.0.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244168  qint8          1012137984(0.05%)
head.anchor_encoder.yaw_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244165  qint16         0(0%)
head.layers.15.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244163  qint8          3489660928(0.16%)
head.layers.15.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244159  qint8          3489660928(0.16%)
head.layers.41.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244158  qint8          872415232(0.04%)
head.layers.41.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244153  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244152  qint16         0(0%)
head.layers.23.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244149  qint8          0(0%)
head.layers.25.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244146  qint16         0(0%)
head.layers.0.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244146  qint8          0(0%)
head.layers.10.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244146  qint8          0(0%)
backbone.stages.2.block.0.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244145  qint8          0(0%)
head.layers.41.quality_layers.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244138  qint16         0(0%)
head.anchor_encoder.size_fc.2.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244136  qint16         0(0%)
head.anchor_encoder.pos_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244135  qint16         0(0%)
head.anchor_encoder.vel_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244134  qint16         0(0%)
head.layers.3.camera_encoder.2.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244134  qint16         0(0%)
head.layers.41.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244128  qint8          0(0%)
head.layers.19.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244127  qint16         0(0%)
head.layers.40.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244125  qint16         0(0%)
head.layers.2.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244123  qint16         0(0%)
head.layers.29.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244122  qint8          0(0%)
head.layers.11.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244116  qint16         0(0%)
head.layers.41.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244115  qint8          37486592(0.00%)
head.layers.21.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244112  qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.5.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244111  qint16         0(0%)
head.anchor_encoder.vel_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244108  qint8          16613376(0.00%)
head.layers.41.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244107  qint8          37486592(0.00%)
head.layers.16.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244103  qint16         0(0%)
head.anchor_encoder.vel_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.02441    qint8          16613376(0.00%)
head.layers.37.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244098  qint16         0(0%)
head.layers.34.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244091  qint16         0(0%)
backbone.stages.2.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244089  qint8          4048551936(0.18%)
head.layers.3.camera_encoder.2.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244083  qint16         0(0%)
backbone.stages.2.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244082  qint8          4048551936(0.18%)
head.layers.22.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244079  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244073  qint16         0(0%)
head.layers.22.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244072  qint8          3489660928(0.16%)
head.layers.0.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244065  qint8          0(0%)
head.layers.20.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244064  qint16         0(0%)
head.layers.18.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244059  qint8          1744830464(0.08%)
head.layers.24.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244058  qint8          0(0%)
head.layers.18.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244049  qint8          1744830464(0.08%)
head.layers.13.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244045  qint8          872415232(0.04%)
head.anchor_encoder.pos_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244044  qint16         0(0%)
head.layers.13.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244042  qint8          872415232(0.04%)
head.layers.39.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244039  qint16         0(0%)
head.layers.41.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244037  qint16         0(0%)
head.layers.38.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244037  qint16         0(0%)
head.layers.3.camera_encoder.5.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244037  qint16         0(0%)
head.layers.34.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244033  qint8          872415232(0.04%)
backbone.stages.2.block.1.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244032  qint8          105431040(0.00%)
head.layers.13.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244029  qint16         0(0%)
head.layers.5.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244028  qint16         0(0%)
head.layers.38.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244028  qint16         0(0%)
backbone.stages.2.block.1.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244027  qint8          105431040(0.00%)
head.layers.34.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244027  qint8          872415232(0.04%)
head.layers.35.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244026  qint8          1744830464(0.08%)
head.layers.34.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244026  qint16         0(0%)
head.layers.17.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244026  qint8          0(0%)
head.layers.31.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244026  qint8          872415232(0.04%)
head.layers.35.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244026  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244025  qint16         0(0%)
backbone.stages.2.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244025  qint8          4048551936(0.18%)
backbone.stages.2.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244024  qint8          4048551936(0.18%)
head.layers.31.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244024  qint8          872415232(0.04%)
head.anchor_encoder.pos_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244024  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244022  qint16         0(0%)
backbone.stages.0.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244022  qint8          0(0%)
backbone.stage_norm.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0244021  qint8          0(0%)
head.anchor_encoder.pos_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024402   qint16         0(0%)
head.layers.3.kps_generator.offset                   activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244019  qint16         81788928(0.00%)
head.layers.41.quality_layers.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244018  qint16         0(0%)
head.layers.2.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244016  qint8          0(0%)
head.layers.8.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.024401   qint8          3489660928(0.16%)
head.layers.3.kps_generator.offset                   weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.024401   qint8          81788928(0.00%)
head.layers.17.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244009  qint16         0(0%)
head.anchor_encoder.vel_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244006  qint8          0(0%)
head.layers.8.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244004  qint8          3489660928(0.16%)
head.layers.41.quality_layers.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.2.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.quality_layers.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.31.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          81788928(0.00%)
head.layers.41.cls_layers.5.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          12145655808(0.55%)
head.layers.4.pre_norm.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
backbone.stages.2.block.6.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.24.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.39.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.10.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          10223616(0.00%)
head.anchor_encoder.size_fc.11.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.32.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244002  qint8          1744830464(0.08%)
head.layers.38.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.4.layers.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          6979321856(0.32%)
head.layers.29.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.15.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          3489660928(0.16%)
head.layers.5.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.12.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          13631488(0.00%)
head.layers.31.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.4.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          12145655808(0.55%)
backbone.stages.3.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.38.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.28.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.14.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          1744830464(0.08%)
head.layers.8.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.27.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.17.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.28.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.1.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.1.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.7.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.24.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.25.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.5.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.30.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.30.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.41.quality_layers.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
backbone.stages.3.block.5.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.39.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.40.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.38.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.14.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.41.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          6815744(0.00%)
backbone.stages.0.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
backbone.stages.3.block.4.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.3.weight_softmax.reciprocal.reciprocal   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
backbone.stages.3.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.4.pre_norm.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.34.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.38.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.27.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
backbone.stages.3.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.39.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.41.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.39.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.cls_layers.5.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.quality_layers.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.4.pre_norm.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.cls_layers.5.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.28.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          1744830464(0.08%)
head.layers.17.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.pos_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.25.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.26.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.36.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.5.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.anchor_encoder.size_fc.5.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.2.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.22.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          3489660928(0.16%)
backbone.stages.2.block.0.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          105431040(0.00%)
head.layers.13.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          37486592(0.00%)
head.layers.35.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.2.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          479232(0.00%)
head.layers.41.cls_layers.2.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.reciprocal_op.reciprocal               activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.pos_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          1417674752(0.06%)
head.layers.14.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.18.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.11.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.17.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.layers.41.quality_layers.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.28.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.29.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.32.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          6979321856(0.32%)
head.layers.8.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.10.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.20.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
backbone.stages.0.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          7197425664(0.33%)
head.anchor_encoder.size_fc.2.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.0.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.downsample_block.2.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.18.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.23.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.38.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.stages.2.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.27.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.38.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.34.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.6.layers.9.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.13.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.30.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.anchor_encoder.size_fc.8.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.20.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.31.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint16         81788928(0.00%)
backbone.stages.0.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
head.layers.13.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint16         37486592(0.00%)
head.layers.17.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.22.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.8.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.8.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.13.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.camera_encoder.2.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.26.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.13.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
head.layers.24.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          479232(0.00%)
head.layers.12.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.22.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.14.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.pos_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.33.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.13.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.20.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.24.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.24.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.27.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.anchor_encoder.yaw_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          88604672(0.00%)
head.layers.3.camera_encoder.5.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          105431040(0.00%)
head.anchor_encoder.vel_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.35.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.41.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.19.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.32.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.1.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.41.quality_layers.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.28.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.3.cam_add                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.6.layers.2                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.35.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.14.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.13.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.28.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.2.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
head.layers.32.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          6979321856(0.32%)
head.layers.15.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.29.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.6.layers.9.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint16         81788928(0.00%)
head.layers.32.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.29.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.3.point_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.34.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.20.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.22.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.pos_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.38.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          81788928(0.00%)
head.layers.20.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.22.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.16.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.41.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.21.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.6.layers.2                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.24.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.31.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.34.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.4.pre_norm.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.stages.1.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          506068992(0.02%)
head.layers.7.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.7.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.41.cls_layers.5.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.5.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.31.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
neck.conv_add.2                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          28789702656(1.31%)
head.anchor_encoder.pos_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.camera_encoder.2.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.39.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.9.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.31.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.anchor_encoder.vel_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.27.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.41.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.25.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.7.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.20.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.37.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.33.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.7.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.30.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.38.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.size_fc.5.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.camera_encoder.2.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
neck.conv_add.2                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          28789702656(1.31%)
head.layers.32.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.15.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.6.add2                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.15.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.output_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.layers.24.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.21.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.anchor_encoder.yaw_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.9.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.36.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.27.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.28.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
backbone.stages.1.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          506068992(0.02%)
head.layers.17.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          10223616(0.00%)
head.layers.8.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          3489660928(0.16%)
head.layers.5.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.13.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.6.cls_layers.6                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.41.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint16         81788928(0.00%)
head.layers.28.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
backbone.stages.3.block.4.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          12145655808(0.55%)
head.layers.7.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.3.camera_encoder.5.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.quality_layers.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.7.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.29.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.41.cls_layers.5.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.27.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.anchor_encoder.size_fc.11.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.3.weight_softmax.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.downsample_block.0.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
head.layers.11.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.quality_layers.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.10.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.20.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.20.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          12145655808(0.55%)
head.layers.2.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.36.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.mat_quant_stub                                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.35.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.22.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.36.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.40.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
neck.fpn_conv.3.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          16194207744(0.74%)
head.layers.24.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.23.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.20.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.28.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.27.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.29.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.40.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.38.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.7.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.8.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.6.quality_layers.6                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          0(0%)
head.layers.38.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1308622848(0.06%)
head.layers.38.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1308622848(0.06%)
backbone.stages.2.block.2.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
head.layers.29.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.3.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
head.layers.27.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.27.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.24.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.24.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.1.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.pos_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
neck.fpn_conv.2.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          64776830976(2.95%)
head.layers.41.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.24.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          10223616(0.00%)
head.layers.36.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.36.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.23.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.10.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.4.pre_norm.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.31.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.17.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.layers.17.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.22.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.7.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          1744830464(0.08%)
head.layers.7.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.38.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.layers.3.point_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          8097103872(0.37%)
backbone.stages.0.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.27.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.37.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.quality_layers.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.27.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.13.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.17.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
head.layers.27.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.2.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.38.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.stages.0.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          14394851328(0.66%)
head.layers.41.cls_layers.2.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.11.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
backbone.stages.0.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          7197425664(0.33%)
head.layers.14.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.41.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.13.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.32.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.32.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.10.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.27.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
backbone.stages.1.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.24.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          81788928(0.00%)
head.layers.33.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.36.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          3489660928(0.16%)
head.layers.6.layers.4.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.1.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.41.cls_layers.5.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.19.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.41.cls_layers.3                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.41.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.cls_layers.5.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.10.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.16.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.20.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.9.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.35.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.24.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.size_fc.8.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.0.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.38.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.27.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.38.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint16         81788928(0.00%)
head.layers.3.output_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.layers.7.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.24.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.vel_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.12.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.stages.0.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.17.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.5.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.14.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.29.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.31.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint16         0(0%)
head.layers.41.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          872415232(0.04%)
head.layers.34.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
backbone.stages.2.block.6.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244002  qint8          4048551936(0.18%)
neck.fpn_conv.3.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          16194207744(0.74%)
head.layers.10.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          88604672(0.00%)
head.anchor_encoder.pos_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244002  qint8          1417674752(0.06%)
head.layers.1.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.1.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.2.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.37.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
neck.fpn_conv.2.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244002  qint8          64776830976(2.95%)
head.anchor_encoder.vel_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244002  qint16         0(0%)
head.layers.7.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          1744830464(0.08%)
head.layers.8.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244002  qint8          3489660928(0.16%)
head.layers.31.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint8          0(0%)
backbone.stages.3.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244002  qint8          0(0%)
head.layers.38.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.13.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244002  qint16         0(0%)
head.layers.41.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.20.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.20.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.17.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          10223616(0.00%)
head.anchor_encoder.pos_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.4.pre_norm.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.0.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          479232(0.00%)
head.layers.41.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          872415232(0.04%)
head.layers.32.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.29.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.32.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.13.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.40.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.5.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.36.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          8097103872(0.37%)
head.layers.1.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.17.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.fc_before                                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint16         15703474176(0.72%)
head.layers.1.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.2.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.1.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.7.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
neck.fpn_conv.0.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          1036429295616(47.24%)
head.layers.28.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.3.point_quant_stub                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.8.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.6.layers.11.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          1417674752(0.06%)
backbone.stages.2.block.0.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
head.anchor_encoder.size_fc.5.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.fc_before                                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          15703474176(0.72%)
head.layers.3.feat_sum                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.31.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.35.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.38.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.17.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
backbone.stages.3.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint8          0(0%)
backbone.downsample_block.0.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          14394851328(0.66%)
head.layers.29.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.2.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.20.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.28.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
neck.fpn_conv.0.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          1036429295616(47.24%)
head.layers.10.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.34.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          37486592(0.00%)
head.layers.28.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          1744830464(0.08%)
head.layers.37.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.10.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
backbone.stages.0.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244001  qint8          7197425664(0.33%)
head.layers.8.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.13.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.12.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.0.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.38.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.24.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.38.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.24.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          479232(0.00%)
backbone.stages.2.block.1.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint8          0(0%)
head.layers.30.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.38.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          872415232(0.04%)
head.layers.3.camera_encoder.5.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.12.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.38.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.13.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.31.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.27.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          0(0%)
head.anchor_encoder.size_fc.11.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.7.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.8.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
backbone.stages.2.block.7.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
head.anchor_encoder.pos_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          33226752(0.00%)
head.anchor_encoder.size_fc.11.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.39.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.20.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          872415232(0.04%)
head.layers.38.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.31.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.10.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          872415232(0.04%)
head.layers.12.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.10.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.cls_layers.2.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.anchor_encoder.size_fc.5.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.23.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.36.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.10.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.anchor_encoder.pos_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.25.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.3.point_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.vel_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.25.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.3.camera_encoder.5.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.11.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244001  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.41.quality_layers.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.cls_layers.3                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          872415232(0.04%)
head.layers.13.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.3                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          88604672(0.00%)
backbone.stages.2.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244001  qint8          4048551936(0.18%)
head.layers.31.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.2.block.7.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
head.layers.11.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.31.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.2.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244001  qint8          4048551936(0.18%)
head.layers.13.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.39.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.41.cls_layers.2.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.cls_layers.5.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.8.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.11.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
head.layers.5.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.0.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          1012137984(0.05%)
backbone.stages.0.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint8          0(0%)
head.layers.36.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.6.layers.4.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.3                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          88604672(0.00%)
head.layers.11.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.10.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.7.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.40.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.cls_layers.2.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.26.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.4.layers.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          6979321856(0.32%)
head.layers.10.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          81788928(0.00%)
head.layers.38.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.3.camera_encoder.2.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.36.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.14.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.41.quality_layers.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.6.layers.9.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.26.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.1.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.5.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.13.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.41.quality_layers.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.anchor_encoder.vel_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.8.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.30.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.11.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.6.layers.4.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.0.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.34.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          872415232(0.04%)
head.layers.11.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.34.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.34.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint16         37486592(0.00%)
head.layers.26.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.41.quality_layers.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.4.pre_norm.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.22.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.28.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.11.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244001  qint8          1744830464(0.08%)
head.layers.28.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.35.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.16.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.10.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.39.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.5.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.38.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.anchor_encoder.pos_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.31.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.2.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.0.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.0.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          1744830464(0.08%)
head.layers.41.cls_layers.5.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          872415232(0.04%)
head.layers.23.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
backbone.stages.0.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          1012137984(0.05%)
head.anchor_encoder.yaw_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.8.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.pos_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.17.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          1308622848(0.06%)
backbone.stages.2.block.4.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          8097103872(0.37%)
head.anchor_encoder.vel_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.12.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.28.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.41.quality_layers.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.17.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.32.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0244001  qint8          1744830464(0.08%)
head.layers.34.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.instance_bank.instance_feature_quant_stub       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.36.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          33226752(0.00%)
head.layers.20.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.22.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.3.point_matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.35.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
backbone.stages.0.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          14394851328(0.66%)
head.layers.30.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.11.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.41.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.38.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
backbone.stages.3.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244001  qint8          4048551936(0.18%)
head.layers.18.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.35.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          1744830464(0.08%)
head.layers.9.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.downsample_block.2.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.11.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0244001  qint16         0(0%)
head.layers.1.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.anchor_encoder.pos_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244001  qint8          1417674752(0.06%)
head.anchor_encoder.vel_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.3.block.5.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          12145655808(0.55%)
head.layers.32.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.6.layers.9.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.17.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          1308622848(0.06%)
head.layers.3.weight_softmax.exp                     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244001  qint16         0(0%)
head.layers.17.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.anchor_encoder.size_fc.2.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
backbone.stages.3.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          94887936(0.00%)
head.layers.36.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244001  qint8          3489660928(0.16%)
head.layers.10.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.3.camera_encoder.2.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint16         0(0%)
head.layers.24.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
backbone.stages.2.block.5.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
backbone.stages.3.block.5.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          12145655808(0.55%)
backbone.stages.2.block.5.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          105431040(0.00%)
backbone.stages.3.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0244001  qint8          94887936(0.00%)
head.anchor_encoder.pos_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244001  qint8          0(0%)
head.layers.7.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244     qint16         0(0%)
neck.conv_add.1                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244     qint8          14394851328(0.66%)
head.anchor_encoder.pos_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244     qint16         0(0%)
head.anchor_encoder.vel_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244     qint8          354418688(0.02%)
neck.conv_add.1                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0244     qint8          14394851328(0.66%)
head.layers.36.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244     qint16         0(0%)
head.anchor_encoder.vel_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244     qint8          354418688(0.02%)
head.layers.26.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244     qint16         0(0%)
head.layers.25.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244     qint8          6979321856(0.32%)
head.layers.3.camera_encoder.3                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244     qint8          10223616(0.00%)
head.anchor_encoder.pos_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0244     qint16         0(0%)
head.layers.31.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244     qint16         0(0%)
head.layers.7.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0244     qint8          0(0%)
head.layers.3.camera_encoder.3                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0244     qint8          10223616(0.00%)
head.layers.4.layers.1                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0244     qint8          3489660928(0.16%)
head.layers.20.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243999  qint16         0(0%)
head.layers.25.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243999  qint8          6979321856(0.32%)
head.layers.10.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243999  qint8          0(0%)
head.layers.39.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243999  qint16         0(0%)
head.layers.15.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243999  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243999  qint16         0(0%)
head.layers.19.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243999  qint16         0(0%)
backbone.stages.2.block.3.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243999  qint8          105431040(0.00%)
head.layers.34.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243999  qint8          872415232(0.04%)
head.layers.1.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243999  qint8          3489660928(0.16%)
head.layers.15.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243999  qint8          3489660928(0.16%)
head.layers.4.layers.1                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243999  qint8          3489660928(0.16%)
backbone.stages.2.block.3.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243998  qint8          105431040(0.00%)
head.layers.30.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.36.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.27.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.34.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243998  qint8          872415232(0.04%)
head.layers.31.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.14.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.34.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243998  qint16         0(0%)
backbone.stages.3.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243998  qint8          94887936(0.00%)
head.layers.31.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
head.layers.15.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243998  qint8          3489660928(0.16%)
head.layers.15.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243998  qint16         0(0%)
backbone.stages.3.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243998  qint8          94887936(0.00%)
head.layers.25.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243997  qint16         0(0%)
head.layers.15.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243997  qint8          3489660928(0.16%)
head.layers.20.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243997  qint8          0(0%)
head.layers.6.layers.9.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243997  qint16         0(0%)
head.layers.15.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243997  qint8          0(0%)
head.layers.4.pre_norm.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243997  qint16         0(0%)
backbone.stages.3.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243996  qint8          12145655808(0.55%)
head.layers.23.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243996  qint16         0(0%)
head.layers.40.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243996  qint16         0(0%)
head.anchor_encoder.pos_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243996  qint16         0(0%)
head.layers.6.layers.7                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243996  qint8          872415232(0.04%)
head.layers.6.layers.7                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243996  qint8          872415232(0.04%)
head.layers.14.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243996  qint8          3489660928(0.16%)
head.layers.23.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243996  qint16         0(0%)
head.layers.14.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243996  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243995  qint16         0(0%)
head.layers.3.camera_encoder.2.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243995  qint16         0(0%)
head.layers.24.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243995  qint16         0(0%)
head.layers.21.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243995  qint16         0(0%)
head.layers.3.camera_encoder.5.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243995  qint16         0(0%)
head.anchor_encoder.size_fc.0                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243995  qint8          8306688(0.00%)
head.layers.29.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243995  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243994  qint16         0(0%)
head.layers.18.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243994  qint16         0(0%)
head.anchor_encoder.pos_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243994  qint16         0(0%)
backbone.quant                                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243994  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243994  qint16         0(0%)
head.layers.13.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243994  qint16         0(0%)
head.layers.6.layers.10                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243994  qint8          37486592(0.00%)
head.layers.29.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243994  qint8          3489660928(0.16%)
head.layers.21.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243994  qint16         0(0%)
head.layers.13.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243994  qint16         0(0%)
head.layers.6.layers.10                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243993  qint16         37486592(0.00%)
head.layers.8.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243993  qint8          3489660928(0.16%)
head.layers.29.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243993  qint8          3489660928(0.16%)
head.layers.3.weights_fc                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243993  qint8          1308622848(0.06%)
head.layers.3.weights_fc                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243993  qint8          1308622848(0.06%)
head.layers.27.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243993  qint16         0(0%)
head.layers.40.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243993  qint16         0(0%)
head.layers.34.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243993  qint16         0(0%)
head.layers.37.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243993  qint16         0(0%)
head.anchor_encoder.pos_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243992  qint16         0(0%)
head.layers.13.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.27.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.4.short_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243992  qint8          1744830464(0.08%)
head.layers.25.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243992  qint16         0(0%)
head.anchor_encoder.size_fc.11.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243992  qint16         0(0%)
head.layers.38.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.38.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.4.short_add                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243992  qint8          1744830464(0.08%)
head.anchor_encoder.yaw_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.33.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243992  qint16         0(0%)
head.layers.21.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243992  qint16         0(0%)
head.layers.17.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243991  qint16         0(0%)
head.layers.0.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243991  qint16         0(0%)
head.layers.6.layers.9.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243991  qint16         0(0%)
head.anchor_encoder.vel_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243991  qint16         0(0%)
head.layers.24.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243991  qint16         0(0%)
head.anchor_encoder.vel_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243991  qint8          354418688(0.02%)
head.anchor_encoder.size_fc.9                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243991  qint8          88604672(0.00%)
head.layers.0.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024399   qint16         0(0%)
head.anchor_encoder.vel_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.024399   qint8          354418688(0.02%)
head.anchor_encoder.size_fc.9                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.024399   qint8          88604672(0.00%)
head.layers.19.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024399   qint16         0(0%)
head.layers.26.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024399   qint16         0(0%)
head.layers.24.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.024399   qint16         0(0%)
head.anchor_encoder.vel_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024399   qint8          0(0%)
head.anchor_encoder.vel_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024399   qint8          0(0%)
head.layers.6.layers.4.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243989  qint16         0(0%)
head.layers.39.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243989  qint8          6979321856(0.32%)
head.layers.39.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243989  qint8          6979321856(0.32%)
backbone.stages.2.block.0.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243989  qint8          8097103872(0.37%)
head.layers.28.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243989  qint16         0(0%)
head.layers.11.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243989  qint8          6979321856(0.32%)
backbone.stages.2.block.0.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243989  qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.5.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243989  qint16         0(0%)
head.layers.30.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243989  qint16         0(0%)
head.layers.29.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243988  qint8          3489660928(0.16%)
head.layers.29.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243988  qint8          3489660928(0.16%)
head.layers.11.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243988  qint8          6979321856(0.32%)
backbone.stage_norm.3                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0243988  qint8          0(0%)
backbone.stages.3.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243988  qint8          94887936(0.00%)
head.anchor_encoder.yaw_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243988  qint16         0(0%)
backbone.stages.3.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243988  qint8          94887936(0.00%)
head.layers.19.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243988  qint8          0(0%)
head.layers.41.cls_layers.2.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243987  qint16         0(0%)
head.layers.34.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243987  qint16         0(0%)
head.layers.10.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243987  qint8          1308622848(0.06%)
head.anchor_encoder.vel_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243987  qint8          0(0%)
head.layers.24.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243987  qint8          0(0%)
head.anchor_encoder.yaw_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243987  qint16         0(0%)
head.layers.20.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243987  qint16         0(0%)
head.layers.10.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243987  qint8          1308622848(0.06%)
backbone.stages.3.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243986  qint8          12145655808(0.55%)
head.layers.10.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243986  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243986  qint16         0(0%)
head.instance_bank.anchor_quant_stub                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243986  qint16         0(0%)
head.layers.29.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243986  qint8          3489660928(0.16%)
head.layers.16.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243986  qint16         0(0%)
head.layers.33.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243986  qint8          0(0%)
head.anchor_encoder.vel_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243985  qint16         0(0%)
head.layers.31.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243985  qint16         0(0%)
head.layers.6.layers.4.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243985  qint16         0(0%)
head.layers.35.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243985  qint16         0(0%)
head.anchor_encoder.pos_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243985  qint16         0(0%)
backbone.stages.1.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243985  qint8          7197425664(0.33%)
backbone.stages.1.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243984  qint8          7197425664(0.33%)
head.layers.5.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243984  qint16         0(0%)
head.layers.16.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243984  qint16         0(0%)
head.instance_bank.feature_cat                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243984  qint8          0(0%)
head.anchor_encoder.pos_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243984  qint16         0(0%)
head.layers.41.quality_layers.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243983  qint8          872415232(0.04%)
head.layers.24.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243983  qint16         0(0%)
head.layers.15.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243983  qint16         0(0%)
head.layers.41.quality_layers.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243983  qint8          872415232(0.04%)
backbone.stages.3.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243983  qint8          4048551936(0.18%)
head.layers.38.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243983  qint16         0(0%)
head.anchor_encoder.vel_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243983  qint16         0(0%)
backbone.stages.3.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243982  qint8          4048551936(0.18%)
head.layers.20.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243982  qint8          872415232(0.04%)
head.layers.20.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243982  qint8          872415232(0.04%)
head.layers.18.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243982  qint8          6979321856(0.32%)
head.layers.18.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243982  qint8          6979321856(0.32%)
head.layers.13.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243982  qint8          872415232(0.04%)
head.layers.31.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243981  qint8          10223616(0.00%)
backbone.patch_embed.1.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0243981  qint8          32388415488(1.48%)
head.layers.13.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243981  qint8          872415232(0.04%)
head.layers.41.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243981  qint16         0(0%)
head.layers.29.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243981  qint16         0(0%)
backbone.patch_embed.1.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0243981  qint8          32388415488(1.48%)
head.anchor_encoder.yaw_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024398   qint16         0(0%)
head.layers.21.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024398   qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024398   qint16         0(0%)
head.layers.35.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024398   qint8          0(0%)
head.layers.32.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243979  qint16         0(0%)
head.layers.6.layers.9.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243979  qint16         0(0%)
head.layers.15.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243979  qint16         0(0%)
head.layers.10.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243979  qint16         0(0%)
head.layers.27.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243979  qint8          0(0%)
head.layers.24.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243979  qint8          1308622848(0.06%)
head.layers.25.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243978  qint16         0(0%)
head.layers.31.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243978  qint8          10223616(0.00%)
head.layers.14.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243978  qint16         0(0%)
head.layers.15.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243978  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243978  qint16         0(0%)
head.layers.13.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243978  qint16         0(0%)
backbone.stages.3.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243978  qint8          4048551936(0.18%)
head.layers.31.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243977  qint16         0(0%)
head.layers.41.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243977  qint16         0(0%)
head.layers.24.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243977  qint8          1308622848(0.06%)
head.layers.21.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243977  qint16         0(0%)
backbone.stages.3.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243976  qint8          4048551936(0.18%)
head.layers.36.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.layers.27.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.layers.34.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.anchor_encoder.vel_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.layers.39.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.layers.8.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243976  qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243975  qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243975  qint16         0(0%)
head.anchor_encoder.vel_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243975  qint16         0(0%)
head.layers.25.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243975  qint8          1744830464(0.08%)
head.layers.25.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243975  qint8          1744830464(0.08%)
head.layers.38.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243975  qint8          479232(0.00%)
head.anchor_encoder.yaw_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243975  qint8          88604672(0.00%)
head.anchor_encoder.yaw_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243974  qint8          88604672(0.00%)
head.layers.13.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243974  qint16         0(0%)
head.layers.34.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243974  qint8          872415232(0.04%)
head.layers.6.layers.9.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243974  qint16         0(0%)
backbone.stages.1.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243973  qint8          7197425664(0.33%)
head.layers.38.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243973  qint8          479232(0.00%)
head.layers.35.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243973  qint16         0(0%)
head.layers.0.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243972  qint16         0(0%)
head.layers.34.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243972  qint8          872415232(0.04%)
backbone.stages.1.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243972  qint8          7197425664(0.33%)
head.anchor_encoder.size_fc.8.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243971  qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243971  qint8          105431040(0.00%)
head.layers.17.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243971  qint8          0(0%)
head.layers.12.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024397   qint16         0(0%)
head.layers.39.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024397   qint8          0(0%)
head.layers.9.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024397   qint16         0(0%)
head.layers.28.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024397   qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.024397   qint8          105431040(0.00%)
head.layers.41.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243969  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243969  qint16         0(0%)
head.layers.6.layers.4.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243963  qint16         0(0%)
head.layers.27.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243963  qint16         0(0%)
head.layers.36.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243956  qint8          3489660928(0.16%)
head.layers.36.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243949  qint8          3489660928(0.16%)
head.layers.27.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243944  qint8          0(0%)
head.layers.24.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243937  qint16         0(0%)
head.anchor_encoder.vel_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243929  qint16         0(0%)
head.fc_after                                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243923  qint8          20937965568(0.95%)
head.layers.25.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243922  qint8          3489660928(0.16%)
head.layers.27.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024392   qint16         0(0%)
head.layers.25.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243918  qint8          3489660928(0.16%)
head.layers.17.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243912  qint16         0(0%)
head.anchor_encoder.size_fc.2.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243904  qint16         0(0%)
backbone.stages.3.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243902  qint8          12145655808(0.55%)
head.layers.19.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243896  qint16         0(0%)
head.layers.36.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024389   qint8          3489660928(0.16%)
head.layers.31.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243889  qint8          479232(0.00%)
head.layers.31.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243886  qint8          479232(0.00%)
head.layers.20.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243882  qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243877  qint16         0(0%)
head.layers.13.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243871  qint16         0(0%)
head.layers.4.pre_norm.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243864  qint16         0(0%)
backbone.stages.2.block.5.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243862  qint8          0(0%)
head.layers.6.layers.0                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243858  qint8          872415232(0.04%)
head.layers.6.layers.0                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243852  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243844  qint8          0(0%)
head.layers.26.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243838  qint16         0(0%)
backbone.stages.1.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243835  qint8          506068992(0.02%)
head.layers.6.layers.4.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243832  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243823  qint16         0(0%)
head.anchor_encoder.size_fc.2.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243818  qint16         0(0%)
head.layers.17.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243814  qint16         0(0%)
backbone.stages.2.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243806  qint8          4048551936(0.18%)
backbone.stages.2.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243798  qint8          4048551936(0.18%)
head.layers.22.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243797  qint8          3489660928(0.16%)
head.layers.3.feat_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243792  qint8          0(0%)
head.anchor_encoder.vel_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243787  qint16         0(0%)
head.layers.1.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243781  qint8          0(0%)
head.layers.3.weight_softmax.sum                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243776  qint16         0(0%)
head.layers.9.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243776  qint8          0(0%)
head.anchor_encoder.size_fc.2.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243772  qint8          0(0%)
head.layers.35.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243767  qint16         0(0%)
backbone.stages.0.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243762  qint8          1012137984(0.05%)
backbone.stages.0.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243758  qint8          1012137984(0.05%)
head.anchor_encoder.yaw_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243757  qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243753  qint8          94887936(0.00%)
backbone.stages.3.block.5.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243748  qint8          94887936(0.00%)
head.layers.27.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243744  qint16         0(0%)
head.layers.38.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243741  qint16         0(0%)
backbone.stages.2.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243737  qint8          4048551936(0.18%)
head.layers.7.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243735  qint8          1744830464(0.08%)
backbone.stages.3.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243733  qint8          12145655808(0.55%)
head.anchor_encoder.size_fc.8.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243733  qint16         0(0%)
backbone.stages.1.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243732  qint8          7197425664(0.33%)
backbone.stages.1.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243731  qint8          7197425664(0.33%)
head.layers.41.quality_layers.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243731  qint16         0(0%)
head.layers.5.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243731  qint8          0(0%)
head.layers.0.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243731  qint16         0(0%)
backbone.stages.2.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243731  qint8          4048551936(0.18%)
head.layers.31.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024373   qint16         0(0%)
backbone.stages.3.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.024373   qint8          12145655808(0.55%)
head.layers.22.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.024373   qint16         0(0%)
head.layers.0.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243729  qint16         0(0%)
head.anchor_encoder.pos_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243728  qint16         0(0%)
backbone.stages.3.block.4.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243728  qint8          94887936(0.00%)
head.layers.34.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243728  qint16         0(0%)
head.layers.22.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243728  qint8          0(0%)
head.layers.41.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243728  qint16         0(0%)
head.layers.27.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243727  qint16         0(0%)
head.layers.20.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243727  qint16         0(0%)
head.layers.21.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243726  qint8          3489660928(0.16%)
head.layers.10.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243726  qint8          0(0%)
head.layers.18.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243726  qint16         0(0%)
head.layers.0.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243725  qint8          1744830464(0.08%)
head.layers.21.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243725  qint8          3489660928(0.16%)
head.layers.33.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243725  qint16         0(0%)
head.layers.10.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243725  qint8          0(0%)
head.layers.24.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243725  qint8          0(0%)
head.anchor_encoder.size_fc.2.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243725  qint16         0(0%)
backbone.stages.3.block.4.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243724  qint8          94887936(0.00%)
head.layers.13.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243724  qint8          872415232(0.04%)
head.layers.0.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243724  qint8          1744830464(0.08%)
head.layers.41.quality_layers.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243724  qint16         0(0%)
head.layers.37.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243724  qint16         0(0%)
head.layers.13.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243723  qint8          872415232(0.04%)
head.layers.7.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243723  qint16         0(0%)
head.layers.38.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243721  qint16         0(0%)
head.layers.3.weight_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024372   qint8          0(0%)
head.anchor_encoder.yaw_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.024372   qint16         0(0%)
head.layers.17.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024372   qint16         0(0%)
head.layers.20.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024372   qint8          0(0%)
head.layers.37.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024372   qint16         0(0%)
head.layers.11.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243718  qint16         0(0%)
head.layers.20.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243718  qint8          872415232(0.04%)
head.layers.14.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243716  qint8          1744830464(0.08%)
head.layers.10.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243716  qint16         0(0%)
head.layers.14.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243715  qint8          1744830464(0.08%)
head.layers.20.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243714  qint8          872415232(0.04%)
head.layers.5.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243713  qint16         0(0%)
head.layers.3.camera_encoder.5.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243713  qint16         0(0%)
backbone.stages.1.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243713  qint8          0(0%)
head.layers.33.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243712  qint16         0(0%)
neck.conv_extract.3.0                                weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.024371   qint8          2699034624(0.12%)
head.layers.1.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243709  qint8          3489660928(0.16%)
head.layers.18.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243708  qint16         0(0%)
head.anchor_encoder.vel_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243707  qint16         0(0%)
head.layers.8.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243706  qint8          0(0%)
neck.conv_extract.3.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243705  qint8          2699034624(0.12%)
head.layers.6.layers.11.scale_quant_stub             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243705  qint16         0(0%)
head.layers.1.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243704  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.02437    qint16         0(0%)
head.layers.34.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243697  qint16         0(0%)
head.layers.18.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243697  qint8          3489660928(0.16%)
head.layers.18.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243691  qint8          3489660928(0.16%)
head.layers.6.layers.4.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243686  qint16         0(0%)
head.layers.34.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243683  qint8          0(0%)
head.layers.31.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243678  qint16         0(0%)
head.layers.10.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243675  qint16         0(0%)
head.layers.24.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243672  qint16         0(0%)
head.anchor_encoder.pos_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243672  qint16         0(0%)
head.layers.31.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243671  qint16         0(0%)
head.layers.33.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243671  qint16         0(0%)
head.layers.17.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243671  qint16         0(0%)
head.layers.1.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243671  qint8          3489660928(0.16%)
head.layers.24.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024367   qint16         0(0%)
head.layers.24.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024367   qint16         0(0%)
head.layers.17.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.024367   qint16         0(0%)
head.layers.5.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.024367   qint16         0(0%)
head.layers.14.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243669  qint8          1744830464(0.08%)
head.layers.35.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243668  qint16         0(0%)
head.layers.27.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243668  qint8          0(0%)
head.layers.11.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243667  qint16         0(0%)
head.layers.1.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243665  qint16         0(0%)
head.layers.27.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243665  qint16         0(0%)
head.layers.3.camera_encoder.5.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243664  qint16         0(0%)
head.layers.24.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243662  qint8          872415232(0.04%)
head.layers.9.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243662  qint16         0(0%)
backbone.stages.2.block.3.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243661  qint8          8097103872(0.37%)
head.layers.31.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243659  qint16         0(0%)
head.layers.24.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243658  qint8          872415232(0.04%)
backbone.stages.2.block.3.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243657  qint8          8097103872(0.37%)
backbone.stages.2.block.2.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243657  qint8          0(0%)
head.layers.10.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243655  qint16         0(0%)
head.layers.17.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243655  qint16         0(0%)
head.layers.22.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243655  qint16         0(0%)
head.layers.12.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243655  qint16         0(0%)
head.layers.13.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243654  qint16         0(0%)
head.layers.17.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243653  qint8          0(0%)
head.layers.18.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243652  qint8          0(0%)
head.layers.20.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.024365   qint8          0(0%)
head.layers.38.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243648  qint8          0(0%)
head.layers.34.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243647  qint16         0(0%)
head.layers.23.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243646  qint16         0(0%)
head.anchor_encoder.yaw_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243646  qint8          88604672(0.00%)
head.layers.21.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243645  qint16         0(0%)
head.anchor_encoder.yaw_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243644  qint8          88604672(0.00%)
head.layers.15.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243644  qint16         0(0%)
head.layers.7.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243641  qint8          3489660928(0.16%)
head.layers.14.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.024364   qint8          1744830464(0.08%)
head.layers.2.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024364   qint16         0(0%)
head.layers.14.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243637  qint8          1744830464(0.08%)
head.layers.31.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243637  qint16         0(0%)
head.anchor_encoder.cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243637  qint8          0(0%)
head.layers.21.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243636  qint16         0(0%)
head.layers.7.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243635  qint8          3489660928(0.16%)
backbone.stages.1.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243633  qint8          7197425664(0.33%)
backbone.stages.2.block.7.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243633  qint8          8097103872(0.37%)
head.layers.17.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024363   qint8          0(0%)
backbone.stages.1.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243629  qint8          7197425664(0.33%)
head.layers.38.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243628  qint16         0(0%)
head.anchor_encoder.pos_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243628  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243627  qint8          8097103872(0.37%)
head.layers.21.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243627  qint8          0(0%)
head.layers.41.quality_layers.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243627  qint8          0(0%)
backbone.stages.2.block.1.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243626  qint8          8097103872(0.37%)
head.layers.39.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243624  qint8          1744830464(0.08%)
head.layers.18.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243623  qint16         0(0%)
backbone.stage_norm.1                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0243623  qint8          0(0%)
head.layers.18.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243622  qint16         0(0%)
head.layers.28.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243622  qint16         0(0%)
head.layers.39.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0243622  qint8          1744830464(0.08%)
backbone.stages.2.block.1.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243621  qint8          8097103872(0.37%)
head.anchor_encoder.pos_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243621  qint16         0(0%)
head.layers.38.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243621  qint8          0(0%)
head.layers.24.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024362   qint8          0(0%)
head.layers.31.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243619  qint16         0(0%)
head.anchor_encoder.yaw_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243619  qint8          5537792(0.00%)
backbone.stages.1.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243619  qint8          506068992(0.02%)
head.anchor_encoder.yaw_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243618  qint8          5537792(0.00%)
head.anchor_encoder.pos_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243617  qint16         0(0%)
backbone.stages.1.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243616  qint8          506068992(0.02%)
head.layers.10.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243615  qint8          0(0%)
head.layers.6.layers.9.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243614  qint16         0(0%)
head.layers.6.layers.4.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243614  qint8          0(0%)
head.layers.41.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243614  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243614  qint8          0(0%)
head.layers.15.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243614  qint8          3489660928(0.16%)
head.layers.15.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243614  qint8          3489660928(0.16%)
head.layers.31.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243613  qint16         0(0%)
head.layers.21.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243612  qint16         0(0%)
head.layers.41.quality_layers.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243612  qint16         0(0%)
head.layers.41.quality_layers.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243612  qint16         0(0%)
head.layers.8.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243606  qint8          3489660928(0.16%)
head.layers.8.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243602  qint8          3489660928(0.16%)
backbone.stages.0.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243598  qint8          14394851328(0.66%)
head.layers.3.feat_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243594  qint8          0(0%)
backbone.stages.0.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243592  qint8          14394851328(0.66%)
head.layers.19.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243588  qint16         0(0%)
head.layers.3.camera_encoder.2.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243583  qint16         0(0%)
head.layers.1.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243579  qint8          0(0%)
head.anchor_encoder.yaw_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243573  qint16         0(0%)
head.layers.1.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243571  qint16         0(0%)
head.layers.20.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024357   qint16         0(0%)
head.anchor_encoder.pos_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243565  qint16         0(0%)
head.layers.16.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024356   qint8          0(0%)
head.anchor_encoder.vel_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243555  qint16         0(0%)
head.layers.29.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243551  qint16         0(0%)
head.layers.20.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243546  qint16         0(0%)
head.layers.10.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243541  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243539  qint8          7197425664(0.33%)
head.layers.37.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243538  qint16         0(0%)
head.layers.24.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243533  qint16         0(0%)
head.layers.21.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243527  qint8          3489660928(0.16%)
head.layers.21.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243524  qint8          3489660928(0.16%)
head.layers.20.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024352   qint8          0(0%)
head.layers.33.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243515  qint16         0(0%)
head.layers.34.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243513  qint16         0(0%)
head.layers.14.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243513  qint8          3489660928(0.16%)
head.layers.14.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243508  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243504  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243498  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243491  qint8          7197425664(0.33%)
head.anchor_encoder.size_fc.6                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243491  qint8          88604672(0.00%)
head.anchor_encoder.size_fc.6                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243485  qint8          88604672(0.00%)
head.anchor_encoder.vel_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243484  qint16         0(0%)
head.layers.20.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243478  qint8          37486592(0.00%)
head.layers.20.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243472  qint16         37486592(0.00%)
head.layers.3.weight_softmax.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243465  qint16         0(0%)
head.layers.24.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243459  qint8          0(0%)
head.layers.3.camera_encoder.0                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243453  qint8          479232(0.00%)
head.layers.3.camera_encoder.0                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243447  qint8          479232(0.00%)
backbone.stages.1.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243442  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243441  qint8          0(0%)
head.layers.0.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024344   qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.8.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243435  qint16         0(0%)
head.layers.15.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243434  qint16         0(0%)
head.layers.27.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243428  qint8          37486592(0.00%)
head.layers.27.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243422  qint16         37486592(0.00%)
head.layers.28.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243417  qint16         0(0%)
backbone.stages.1.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243411  qint8          0(0%)
head.layers.24.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243404  qint16         0(0%)
head.layers.34.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243398  qint16         0(0%)
backbone.stages.0.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243392  qint8          7197425664(0.33%)
head.layers.22.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243391  qint8          3489660928(0.16%)
backbone.stages.0.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243386  qint8          7197425664(0.33%)
head.layers.30.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243381  qint16         0(0%)
head.layers.35.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243374  qint8          1744830464(0.08%)
head.layers.0.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243371  qint8          3489660928(0.16%)
head.layers.35.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243368  qint8          1744830464(0.08%)
head.layers.31.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243363  qint8          1308622848(0.06%)
head.layers.12.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243362  qint16         0(0%)
head.layers.0.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243361  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243356  qint16         0(0%)
head.anchor_encoder.size_fc.8.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024335   qint16         0(0%)
backbone.patch_embed.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0243346  qint8          6072827904(0.28%)
backbone.patch_embed.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0243341  qint8          6072827904(0.28%)
head.layers.38.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243337  qint16         0(0%)
head.anchor_encoder.pos_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243333  qint16         0(0%)
head.layers.31.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243331  qint8          1308622848(0.06%)
neck.conv_add.0                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243327  qint8          5398069248(0.25%)
head.anchor_encoder.size_fc.0                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243326  qint8          8306688(0.00%)
neck.conv_add.0                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243321  qint8          5398069248(0.25%)
head.layers.20.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243318  qint16         0(0%)
backbone.stages.0.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243313  qint8          7197425664(0.33%)
head.layers.6.layers.5                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243312  qint8          872415232(0.04%)
backbone.stages.1.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243306  qint8          7197425664(0.33%)
backbone.stages.0.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243305  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243304  qint16         0(0%)
head.layers.40.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243301  qint16         0(0%)
head.layers.9.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.02433    qint16         0(0%)
backbone.stages.3.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.024329   qint8          4048551936(0.18%)
head.layers.10.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243288  qint16         0(0%)
backbone.stages.3.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0243287  qint8          4048551936(0.18%)
head.layers.3.camera_encoder.5.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243286  qint16         0(0%)
head.anchor_encoder.vel_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243277  qint16         0(0%)
backbone.stage_norm.2                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0243271  qint8          0(0%)
head.layers.10.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024327   qint16         0(0%)
head.layers.31.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243265  qint8          0(0%)
head.anchor_encoder.yaw_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243261  qint16         0(0%)
head.layers.16.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243259  qint16         0(0%)
head.layers.31.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243258  qint8          0(0%)
head.layers.10.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243255  qint16         0(0%)
head.layers.6.layers.4.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243251  qint16         0(0%)
head.layers.17.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243249  qint16         0(0%)
head.layers.15.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243247  qint8          3489660928(0.16%)
head.layers.8.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243244  qint16         0(0%)
head.layers.3.kps_generator.keypoints_add            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243238  qint16         0(0%)
head.layers.27.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243238  qint16         0(0%)
head.layers.31.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243232  qint8          0(0%)
head.layers.41.cls_layers.5.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243229  qint16         0(0%)
head.layers.21.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243228  qint8          0(0%)
head.layers.6.layers.5                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243223  qint8          872415232(0.04%)
head.layers.35.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243215  qint8          1744830464(0.08%)
head.layers.17.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243214  qint16         0(0%)
head.layers.17.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243214  qint16         0(0%)
head.layers.27.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243212  qint16         0(0%)
head.layers.29.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243207  qint8          0(0%)
head.layers.37.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243206  qint8          0(0%)
head.layers.17.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243206  qint16         0(0%)
head.layers.17.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243202  qint16         0(0%)
head.layers.22.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243201  qint16         0(0%)
head.layers.17.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243198  qint16         81788928(0.00%)
head.layers.9.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243198  qint16         0(0%)
head.layers.40.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243193  qint16         0(0%)
head.layers.21.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243193  qint8          0(0%)
head.layers.38.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243191  qint8          10223616(0.00%)
head.layers.25.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243191  qint8          0(0%)
head.layers.27.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.024319   qint16         0(0%)
head.layers.10.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243187  qint8          0(0%)
backbone.stages.0.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243186  qint8          0(0%)
head.layers.17.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243184  qint16         0(0%)
head.layers.17.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243183  qint8          81788928(0.00%)
head.layers.10.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243181  qint8          0(0%)
head.anchor_encoder.yaw_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243175  qint16         0(0%)
head.layers.38.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0243175  qint8          10223616(0.00%)
head.layers.32.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243173  qint16         0(0%)
head.layers.13.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243165  qint16         0(0%)
head.layers.17.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243155  qint16         0(0%)
head.layers.19.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243152  qint16         0(0%)
head.layers.41.cls_layers.2.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243143  qint16         0(0%)
head.layers.6.add1                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0243127  qint8          0(0%)
head.layers.24.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0243097  qint16         0(0%)
head.layers.22.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243093  qint8          3489660928(0.16%)
backbone.stages.3.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243093  qint8          12145655808(0.55%)
backbone.stages.1.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0243081  qint8          506068992(0.02%)
head.fc_after                                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0243063  qint16         20937965568(0.95%)
head.layers.41.cls_layers.2.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0243005  qint16         0(0%)
head.layers.0.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0242981  qint16         0(0%)
head.layers.38.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0242762  qint16         0(0%)
head.layers.38.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0242455  qint16         0(0%)
head.layers.20.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.023862   qint16         0(0%)
head.layers.22.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0237513  qint8          3489660928(0.16%)
head.layers.24.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.023419   qint16         0(0%)