op_name                                              sensitive_type    op_type                                                                            L1  quant_dtype    flops
---------------------------------------------------  ----------------  --------------------------------------------------------------------------  ---------  -------------  ---------------------
head.layers.41.cls_layers.0                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.113567   qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.113354   qint16         0(0%)
head.layers.41.cls_layers.3                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0582398  qint8          872415232(0.04%)
head.anchor_encoder.size_fc.11.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0415663  qint16         0(0%)
head.layers.26.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.030678   qint16         0(0%)
head.layers.33.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0303168  qint16         0(0%)
head.anchor_encoder.size_fc.0                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0230101  qint8          8306688(0.00%)
head.anchor_encoder.pos_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0228263  qint16         0(0%)
head.anchor_encoder.pos_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0227795  qint8          1417674752(0.06%)
head.layers.6.layers.10                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.022647   qint8          37486592(0.00%)
head.anchor_encoder.vel_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0225465  qint8          354418688(0.02%)
head.layers.6.layers.2                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0223917  qint8          872415232(0.04%)
head.layers.6.layers.0                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0223137  qint8          872415232(0.04%)
head.anchor_encoder.size_fc.3                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0220172  qint8          88604672(0.00%)
head.layers.6.layers.9.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0219466  qint16         0(0%)
head.layers.6.layers.9.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0218813  qint16         0(0%)
head.anchor_encoder.yaw_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0216615  qint8          88604672(0.00%)
head.layers.6.layers.11.scale_quant_stub             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0216061  qint16         0(0%)
head.layers.6.layers.7                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0215407  qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0214342  qint8          88604672(0.00%)
head.anchor_encoder.vel_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0214208  qint8          354418688(0.02%)
backbone.stages.3.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0213243  qint8          0(0%)
backbone.stages.1.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0212002  qint8          7197425664(0.33%)
head.layers.34.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.021047   qint16         0(0%)
head.layers.40.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0210336  qint16         0(0%)
head.layers.25.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0209776  qint8          1744830464(0.08%)
head.layers.5.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0209502  qint16         0(0%)
head.layers.27.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0209195  qint16         0(0%)
head.fc_after                                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0208835  qint8          20937965568(0.95%)
head.layers.10.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0208459  qint16         0(0%)
backbone.downsample_block.1.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0207965  qint8          10796138496(0.49%)
head.anchor_encoder.yaw_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0207803  qint16         0(0%)
head.anchor_encoder.cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0207708  qint8          0(0%)
head.layers.6.layers.9.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0207679  qint16         0(0%)
head.anchor_encoder.size_fc.8.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0207569  qint8          0(0%)
head.layers.6.layers.9.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0207554  qint8          0(0%)
backbone.downsample_block.0.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.020683   qint8          14394851328(0.66%)
backbone.stages.1.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0206188  qint8          7197425664(0.33%)
head.layers.24.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0206069  qint16         0(0%)
head.layers.17.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0206062  qint16         0(0%)
head.layers.32.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0205533  qint8          1744830464(0.08%)
head.layers.6.layers.5                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0205469  qint8          872415232(0.04%)
head.layers.11.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0205304  qint8          1744830464(0.08%)
backbone.stages.0.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0205256  qint8          7197425664(0.33%)
backbone.stages.2.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0204865  qint8          4048551936(0.18%)
backbone.stages.1.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0204803  qint8          7197425664(0.33%)
backbone.stages.0.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.020475   qint8          7197425664(0.33%)
backbone.patch_embed.1.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0204516  qint8          32388415488(1.48%)
head.mat_quant_stub                                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0204214  qint16         0(0%)
head.layers.27.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0203888  qint16         0(0%)
backbone.stages.0.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.020384   qint8          7197425664(0.33%)
backbone.stages.0.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0203811  qint8          7197425664(0.33%)
backbone.stages.2.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0203796  qint8          4048551936(0.18%)
head.layers.22.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0203622  qint16         0(0%)
backbone.quant                                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0203612  qint8          0(0%)
backbone.stages.2.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0203544  qint8          4048551936(0.18%)
head.anchor_encoder.yaw_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0203416  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0203304  qint16         0(0%)
head.anchor_encoder.vel_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0202657  qint16         0(0%)
head.anchor_encoder.size_fc.6                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.020265   qint8          88604672(0.00%)
backbone.stages.2.block.6.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0202224  qint8          4048551936(0.18%)
backbone.stages.0.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0201986  qint8          7197425664(0.33%)
backbone.stages.2.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0201933  qint8          4048551936(0.18%)
head.anchor_encoder.pos_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0201716  qint16         0(0%)
backbone.patch_embed.1.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.020164   qint8          32388415488(1.48%)
head.anchor_encoder.pos_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0201094  qint16         0(0%)
backbone.stages.1.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0201066  qint8          7197425664(0.33%)
backbone.patch_embed.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0200708  qint8          6072827904(0.28%)
backbone.patch_embed.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                   0.0200373  qint8          6072827904(0.28%)
backbone.stages.1.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0200362  qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0200284  qint16         0(0%)
head.layers.39.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0200283  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.020016   qint8          0(0%)
neck.fpn_conv.2.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0200079  qint8          64776830976(2.95%)
head.fc_before                                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0199978  qint8          15703474176(0.72%)
head.anchor_encoder.pos_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0199613  qint16         0(0%)
head.anchor_encoder.yaw_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0199447  qint8          5537792(0.00%)
backbone.downsample_block.1.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0199401  qint8          10796138496(0.49%)
head.layers.19.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0199395  qint16         0(0%)
head.layers.41.cls_layers.5.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.019935   qint16         0(0%)
head.layers.24.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0199339  qint8          0(0%)
backbone.stages.2.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0198912  qint8          4048551936(0.18%)
head.anchor_encoder.size_fc.2.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0198359  qint16         0(0%)
head.layers.41.cls_layers.2.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0198311  qint16         0(0%)
head.layers.6.layers.10                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0198111  qint16         37486592(0.00%)
head.layers.30.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0197643  qint16         0(0%)
backbone.stages.1.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0197528  qint8          7197425664(0.33%)
head.anchor_encoder.yaw_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0197296  qint16         0(0%)
backbone.stages.0.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0197223  qint8          0(0%)
backbone.downsample_block.0.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0197145  qint8          14394851328(0.66%)
neck.conv_add.0                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0197136  qint8          5398069248(0.25%)
head.layers.18.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0196871  qint8          1744830464(0.08%)
head.layers.41.cls_layers.2.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0196652  qint8          0(0%)
head.anchor_encoder.size_fc.0                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0196622  qint8          8306688(0.00%)
head.layers.2.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0196584  qint16         0(0%)
backbone.stages.2.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0196363  qint8          4048551936(0.18%)
head.layers.31.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0196021  qint16         0(0%)
head.layers.37.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0196012  qint16         0(0%)
backbone.stages.1.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0195985  qint8          0(0%)
backbone.stages.2.block.7.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0195839  qint8          4048551936(0.18%)
head.anchor_encoder.vel_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0195437  qint16         0(0%)
backbone.stages.2.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0195155  qint8          4048551936(0.18%)
head.anchor_encoder.pos_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01946    qint16         0(0%)
backbone.stages.2.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0194226  qint8          4048551936(0.18%)
backbone.downsample_block.2.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0194059  qint8          8097103872(0.37%)
backbone.stages.1.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.019332   qint8          7197425664(0.33%)
head.anchor_encoder.pos_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0193021  qint8          0(0%)
head.layers.34.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0192671  qint16         0(0%)
backbone.stages.1.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0192663  qint8          7197425664(0.33%)
head.layers.41.cls_layers.2.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0192589  qint16         0(0%)
head.layers.20.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0192473  qint16         0(0%)
head.anchor_encoder.pos_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0192431  qint16         0(0%)
head.anchor_encoder.pos_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.01924    qint8          1417674752(0.06%)
backbone.stages.2.block.1.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0192181  qint8          0(0%)
backbone.stages.3.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0191669  qint8          4048551936(0.18%)
head.layers.23.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0191566  qint16         0(0%)
head.anchor_encoder.pos_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0191313  qint8          1417674752(0.06%)
head.layers.27.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0191225  qint16         0(0%)
head.layers.24.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0191048  qint16         0(0%)
backbone.stages.1.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0190931  qint8          0(0%)
head.layers.3.weight_softmax.reciprocal.reciprocal   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0190894  qint16         0(0%)
head.anchor_encoder.size_fc.9                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0190797  qint8          88604672(0.00%)
backbone.stages.0.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.019077   qint8          7197425664(0.33%)
backbone.stages.2.block.1.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0190662  qint8          8097103872(0.37%)
head.anchor_encoder.pos_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0190622  qint16         0(0%)
head.anchor_encoder.size_fc.2.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.019053   qint16         0(0%)
head.anchor_encoder.vel_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0190491  qint8          16613376(0.00%)
head.layers.10.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0190342  qint16         0(0%)
head.anchor_encoder.vel_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0190293  qint16         0(0%)
head.anchor_encoder.pos_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0190188  qint16         0(0%)
backbone.stages.2.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0190067  qint8          4048551936(0.18%)
head.layers.10.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0189969  qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0189949  qint16         0(0%)
head.anchor_encoder.pos_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0189419  qint16         0(0%)
backbone.stages.3.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0189399  qint8          4048551936(0.18%)
head.layers.35.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0189078  qint8          3489660928(0.16%)
head.layers.4.short_add                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0188943  qint8          1744830464(0.08%)
backbone.stages.2.block.4.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0188928  qint8          0(0%)
backbone.stages.2.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.018892   qint8          4048551936(0.18%)
head.layers.29.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018891   qint8          3489660928(0.16%)
head.layers.22.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0188839  qint8          3489660928(0.16%)
backbone.stages.2.block.2.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0188754  qint8          8097103872(0.37%)
head.anchor_encoder.pos_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188713  qint16         0(0%)
head.anchor_encoder.size_fc.8.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188698  qint16         0(0%)
backbone.stages.3.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0188664  qint8          4048551936(0.18%)
head.anchor_encoder.yaw_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188627  qint16         0(0%)
head.anchor_encoder.pos_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0188608  qint8          33226752(0.00%)
backbone.stages.2.block.0.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0188564  qint8          0(0%)
head.anchor_encoder.pos_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188531  qint8          0(0%)
backbone.stages.3.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0188483  qint8          4048551936(0.18%)
head.anchor_encoder.vel_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0188376  qint8          354418688(0.02%)
head.anchor_encoder.size_fc.8.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018835   qint16         0(0%)
backbone.stages.3.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0188331  qint8          4048551936(0.18%)
head.anchor_encoder.yaw_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188315  qint8          0(0%)
head.layers.7.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188273  qint8          1744830464(0.08%)
backbone.stages.0.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0188223  qint8          14394851328(0.66%)
head.layers.28.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0188177  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188168  qint16         0(0%)
head.layers.16.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0188146  qint16         0(0%)
head.anchor_encoder.yaw_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0188097  qint8          88604672(0.00%)
head.layers.29.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0188096  qint8          0(0%)
backbone.downsample_block.2.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0188059  qint8          8097103872(0.37%)
backbone.stages.3.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0188051  qint8          0(0%)
head.anchor_encoder.pos_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0188046  qint8          1417674752(0.06%)
head.layers.29.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0188023  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.9                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0187906  qint8          88604672(0.00%)
backbone.stages.3.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0187819  qint8          4048551936(0.18%)
head.layers.10.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187817  qint8          0(0%)
backbone.stages.2.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.018778   qint8          4048551936(0.18%)
backbone.stages.0.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0187766  qint8          0(0%)
backbone.stages.3.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0187703  qint8          12145655808(0.55%)
head.anchor_encoder.vel_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187604  qint16         0(0%)
head.layers.22.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0187589  qint16         0(0%)
head.layers.41.cls_layers.5.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187573  qint16         0(0%)
backbone.stages.1.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0187569  qint8          7197425664(0.33%)
backbone.stages.0.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0187526  qint8          1012137984(0.05%)
head.anchor_encoder.pos_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0187409  qint8          1417674752(0.06%)
head.layers.3.weight_softmax.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187395  qint8          0(0%)
backbone.stages.3.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0187385  qint8          94887936(0.00%)
backbone.stages.3.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0187375  qint8          4048551936(0.18%)
backbone.stages.3.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.018733   qint8          4048551936(0.18%)
backbone.stages.0.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0187284  qint8          7197425664(0.33%)
head.anchor_encoder.yaw_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187253  qint16         0(0%)
head.layers.29.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0187185  qint16         0(0%)
backbone.stages.2.block.3.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0187154  qint8          0(0%)
head.layers.22.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0187144  qint8          3489660928(0.16%)
head.layers.11.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0187095  qint16         0(0%)
head.layers.29.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0187089  qint8          3489660928(0.16%)
head.layers.17.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018707   qint8          0(0%)
head.layers.39.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0187061  qint8          6979321856(0.32%)
head.layers.19.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0187059  qint16         0(0%)
head.anchor_encoder.vel_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187059  qint16         0(0%)
head.layers.3.kps_generator.offset                   weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0187047  qint8          81788928(0.00%)
head.layers.13.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187044  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0187036  qint16         0(0%)
head.layers.36.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186987  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0186985  qint8          5537792(0.00%)
backbone.stages.2.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.018698   qint8          4048551936(0.18%)
head.layers.3.feat_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186978  qint8          0(0%)
backbone.stages.0.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186896  qint8          14394851328(0.66%)
backbone.stages.3.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186849  qint8          12145655808(0.55%)
backbone.stages.1.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0186829  qint8          0(0%)
head.layers.20.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186799  qint16         0(0%)
head.layers.32.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.018675   qint8          1744830464(0.08%)
head.layers.24.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186745  qint8          0(0%)
head.layers.36.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186704  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186671  qint16         0(0%)
head.layers.20.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186662  qint16         0(0%)
backbone.stages.0.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0186602  qint8          7197425664(0.33%)
head.layers.6.layers.11.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01866    qint16         0(0%)
backbone.stages.2.block.6.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0186594  qint8          4048551936(0.18%)
backbone.stages.2.block.0.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186562  qint8          105431040(0.00%)
head.layers.36.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018652   qint16         0(0%)
head.layers.32.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186514  qint16         0(0%)
head.layers.10.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186464  qint8          872415232(0.04%)
head.layers.6.layers.4.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186418  qint16         0(0%)
head.layers.13.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186417  qint16         0(0%)
head.layers.31.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186394  qint16         0(0%)
head.layers.20.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018639   qint16         0(0%)
head.instance_bank.feature_cat                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018636   qint8          0(0%)
backbone.stages.2.block.5.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0186326  qint8          0(0%)
neck.conv_extract.3.0                                weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186324  qint8          2699034624(0.12%)
head.layers.3.camera_encoder.5.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186318  qint16         0(0%)
head.layers.20.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0186317  qint16         0(0%)
head.layers.41.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01863    qint16         0(0%)
head.layers.24.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186277  qint8          1308622848(0.06%)
head.layers.24.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186276  qint8          872415232(0.04%)
head.layers.38.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186251  qint8          1308622848(0.06%)
head.anchor_encoder.yaw_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0186238  qint8          88604672(0.00%)
backbone.stages.2.block.5.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186236  qint8          8097103872(0.37%)
backbone.stages.2.block.5.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186224  qint8          105431040(0.00%)
head.layers.32.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186198  qint16         0(0%)
head.layers.38.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186182  qint16         0(0%)
head.layers.24.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186159  qint8          0(0%)
head.anchor_encoder.size_fc.8.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186111  qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0186104  qint8          14394851328(0.66%)
head.layers.41.cls_layers.5.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186098  qint8          0(0%)
head.layers.10.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186071  qint8          1308622848(0.06%)
head.layers.15.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0186055  qint8          3489660928(0.16%)
head.layers.13.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0186031  qint8          872415232(0.04%)
head.layers.41.cls_layers.5.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0186015  qint16         0(0%)
head.layers.7.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185997  qint8          0(0%)
head.layers.36.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185945  qint8          0(0%)
head.layers.3.weights_fc                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185943  qint8          1308622848(0.06%)
head.layers.32.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185942  qint16         0(0%)
head.layers.22.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185934  qint16         0(0%)
head.layers.13.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018593   qint16         0(0%)
head.layers.41.quality_layers.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185914  qint16         0(0%)
head.layers.12.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185913  qint16         0(0%)
backbone.stages.2.block.0.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185905  qint8          8097103872(0.37%)
head.layers.6.layers.9.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185894  qint16         0(0%)
backbone.stages.3.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0185881  qint8          4048551936(0.18%)
head.layers.29.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185858  qint16         0(0%)
head.layers.24.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185857  qint16         0(0%)
head.anchor_encoder.size_fc.8.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185853  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185846  qint16         0(0%)
head.layers.25.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185837  qint8          3489660928(0.16%)
backbone.stages.3.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0185827  qint8          4048551936(0.18%)
head.layers.41.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185827  qint16         0(0%)
head.layers.36.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185825  qint16         0(0%)
backbone.stages.3.block.4.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185762  qint8          12145655808(0.55%)
head.layers.31.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185744  qint8          0(0%)
head.layers.31.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185733  qint16         0(0%)
head.layers.22.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185688  qint8          0(0%)
head.layers.17.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018568   qint8          0(0%)
head.layers.31.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185627  qint16         0(0%)
head.layers.41.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185625  qint16         0(0%)
backbone.stages.0.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185624  qint8          1012137984(0.05%)
head.layers.15.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185618  qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185612  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185606  qint8          7197425664(0.33%)
head.anchor_encoder.size_fc.2.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185591  qint16         0(0%)
head.layers.24.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185585  qint16         0(0%)
head.layers.36.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185545  qint8          3489660928(0.16%)
backbone.stages.3.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018554   qint8          0(0%)
head.anchor_encoder.yaw_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0185529  qint8          88604672(0.00%)
backbone.stages.3.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.018552   qint8          4048551936(0.18%)
head.layers.10.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185478  qint8          0(0%)
head.layers.3.kps_generator.offset                   activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185467  qint16         81788928(0.00%)
head.layers.6.layers.9.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185464  qint16         0(0%)
head.layers.8.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185459  qint8          3489660928(0.16%)
head.layers.36.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185456  qint8          0(0%)
head.layers.15.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185438  qint8          0(0%)
head.layers.39.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185421  qint16         0(0%)
head.layers.29.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185415  qint8          3489660928(0.16%)
head.layers.1.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185414  qint8          3489660928(0.16%)
head.layers.41.quality_layers.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185414  qint16         0(0%)
head.layers.24.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185383  qint8          81788928(0.00%)
head.layers.22.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185348  qint16         0(0%)
head.layers.8.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185344  qint8          3489660928(0.16%)
head.layers.6.add2                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185339  qint16         0(0%)
head.anchor_encoder.pos_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185337  qint16         0(0%)
backbone.stages.1.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185334  qint8          7197425664(0.33%)
head.layers.25.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0185313  qint8          1744830464(0.08%)
head.layers.22.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185311  qint8          3489660928(0.16%)
head.layers.13.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185301  qint8          0(0%)
head.anchor_encoder.size_fc.11.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185296  qint16         0(0%)
backbone.stages.3.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018527   qint8          12145655808(0.55%)
head.layers.9.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185261  qint16         0(0%)
head.layers.14.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185221  qint8          1744830464(0.08%)
head.layers.38.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185198  qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185176  qint16         0(0%)
backbone.stages.2.block.7.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185174  qint8          105431040(0.00%)
head.layers.24.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0185163  qint16         0(0%)
backbone.stages.2.block.7.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0185155  qint8          4048551936(0.18%)
head.anchor_encoder.size_fc.2.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018511   qint16         0(0%)
head.layers.38.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185108  qint8          0(0%)
head.layers.3.output_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0185097  qint8          872415232(0.04%)
head.layers.40.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185097  qint16         0(0%)
head.layers.26.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0185084  qint16         0(0%)
head.layers.21.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185084  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185041  qint16         0(0%)
head.layers.22.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018504   qint8          0(0%)
head.anchor_encoder.yaw_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185035  qint8          0(0%)
backbone.stages.0.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0185022  qint8          14394851328(0.66%)
head.layers.27.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0185013  qint16         0(0%)
head.anchor_encoder.pos_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184993  qint16         0(0%)
head.layers.34.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018498   qint8          0(0%)
head.layers.27.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184974  qint8          0(0%)
head.layers.13.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184971  qint16         37486592(0.00%)
head.layers.41.cls_layers.5.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184957  qint16         0(0%)
head.layers.41.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184957  qint16         0(0%)
head.layers.13.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184941  qint8          872415232(0.04%)
head.fc_after                                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184932  qint16         20937965568(0.95%)
head.layers.41.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184929  qint16         0(0%)
head.layers.39.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184903  qint16         0(0%)
head.layers.20.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184902  qint8          872415232(0.04%)
head.layers.33.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184902  qint16         0(0%)
backbone.stages.3.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184898  qint8          0(0%)
head.layers.13.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184885  qint16         0(0%)
backbone.stages.2.block.5.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184871  qint8          8097103872(0.37%)
head.layers.38.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184809  qint16         0(0%)
head.layers.3.weight_softmax.exp                     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184793  qint16         0(0%)
head.layers.13.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184751  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184751  qint8          8097103872(0.37%)
head.layers.37.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184747  qint16         0(0%)
head.layers.13.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184744  qint16         0(0%)
head.layers.23.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184737  qint16         0(0%)
head.layers.38.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184728  qint8          0(0%)
head.layers.34.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184722  qint16         0(0%)
head.layers.31.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184715  qint8          1308622848(0.06%)
head.anchor_encoder.size_fc.2.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184708  qint16         0(0%)
head.anchor_encoder.size_fc.6                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184707  qint8          88604672(0.00%)
head.layers.40.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184707  qint16         0(0%)
head.layers.32.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184702  qint8          6979321856(0.32%)
head.layers.3.camera_encoder.2.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184675  qint16         0(0%)
head.anchor_encoder.vel_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184663  qint8          354418688(0.02%)
head.layers.1.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184659  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.3                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184631  qint8          10223616(0.00%)
head.layers.27.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184614  qint16         0(0%)
head.layers.0.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184611  qint16         0(0%)
neck.conv_extract.3.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184595  qint8          2699034624(0.12%)
backbone.stages.3.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184595  qint8          94887936(0.00%)
head.anchor_encoder.vel_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184577  qint16         0(0%)
head.layers.31.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184576  qint16         81788928(0.00%)
head.layers.24.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184571  qint16         0(0%)
head.layers.7.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184563  qint8          1744830464(0.08%)
head.layers.27.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.018456   qint8          872415232(0.04%)
head.layers.34.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184556  qint8          872415232(0.04%)
head.layers.4.layers.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184555  qint8          6979321856(0.32%)
backbone.stages.0.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184544  qint8          0(0%)
head.layers.24.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184541  qint16         0(0%)
backbone.stages.3.block.4.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184541  qint8          94887936(0.00%)
head.layers.6.layers.4.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184537  qint16         0(0%)
head.layers.10.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184528  qint8          0(0%)
backbone.stage_norm.1                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0184524  qint8          0(0%)
head.layers.24.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184522  qint8          0(0%)
head.layers.26.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184518  qint16         0(0%)
head.layers.3.feat_sum                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184517  qint8          0(0%)
head.layers.3.camera_encoder.5.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184505  qint16         0(0%)
head.layers.6.layers.4.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184505  qint16         0(0%)
head.layers.10.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184503  qint16         0(0%)
head.layers.7.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184502  qint8          3489660928(0.16%)
head.layers.41.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184501  qint8          0(0%)
head.layers.38.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184499  qint8          0(0%)
head.layers.1.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184489  qint8          3489660928(0.16%)
head.layers.22.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184485  qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.2.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184481  qint16         0(0%)
head.layers.10.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184476  qint8          81788928(0.00%)
head.layers.38.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184473  qint16         0(0%)
head.layers.20.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184473  qint16         0(0%)
head.anchor_encoder.size_fc.11.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018446   qint16         0(0%)
head.layers.28.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184458  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184451  qint8          10223616(0.00%)
head.layers.20.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184451  qint16         0(0%)
head.layers.18.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0184451  qint8          1744830464(0.08%)
head.layers.41.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184438  qint16         0(0%)
head.layers.28.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184429  qint8          1744830464(0.08%)
head.layers.3.camera_encoder.2.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184425  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184424  qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184418  qint16         0(0%)
head.layers.41.quality_layers.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01844    qint8          0(0%)
head.layers.41.quality_layers.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184383  qint16         0(0%)
head.layers.31.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184375  qint8          0(0%)
head.anchor_encoder.yaw_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184373  qint16         0(0%)
head.layers.31.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184372  qint8          872415232(0.04%)
head.layers.20.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184371  qint16         0(0%)
head.layers.15.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184362  qint8          3489660928(0.16%)
head.layers.17.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184349  qint8          0(0%)
backbone.stages.0.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184342  qint8          1012137984(0.05%)
head.layers.17.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184334  qint8          479232(0.00%)
head.layers.10.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184327  qint16         0(0%)
head.layers.13.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184319  qint16         0(0%)
head.layers.10.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184318  qint16         0(0%)
head.layers.13.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184314  qint8          872415232(0.04%)
head.layers.8.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184307  qint8          3489660928(0.16%)
head.layers.9.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184295  qint16         0(0%)
head.layers.28.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184294  qint8          1744830464(0.08%)
backbone.stages.0.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184281  qint8          14394851328(0.66%)
head.layers.41.cls_layers.5.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184275  qint16         0(0%)
backbone.stages.2.block.2.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184268  qint8          8097103872(0.37%)
head.layers.21.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184267  qint16         0(0%)
head.layers.34.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184251  qint16         0(0%)
neck.fpn_conv.1.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184243  qint8          259107323904(11.81%)
head.layers.17.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184231  qint16         0(0%)
head.layers.37.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184229  qint8          0(0%)
head.anchor_encoder.vel_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184226  qint16         0(0%)
head.anchor_encoder.yaw_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184225  qint8          88604672(0.00%)
backbone.stages.3.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0184222  qint8          12145655808(0.55%)
head.layers.4.layers.1                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018421   qint8          3489660928(0.16%)
head.layers.10.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184208  qint16         0(0%)
head.anchor_encoder.vel_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184205  qint16         0(0%)
head.anchor_encoder.vel_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184202  qint16         0(0%)
head.layers.41.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184202  qint16         0(0%)
head.layers.17.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01842    qint16         0(0%)
head.layers.28.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018419   qint8          0(0%)
head.layers.27.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184183  qint8          872415232(0.04%)
head.layers.17.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184173  qint16         0(0%)
head.anchor_encoder.size_fc.5.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0184164  qint16         0(0%)
head.layers.17.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184154  qint8          0(0%)
head.layers.26.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184152  qint16         0(0%)
head.layers.1.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184144  qint8          0(0%)
head.layers.10.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184134  qint8          0(0%)
head.layers.11.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184128  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.2.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184122  qint16         0(0%)
head.layers.13.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0184121  qint16         0(0%)
head.layers.17.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0184096  qint8          81788928(0.00%)
head.anchor_encoder.vel_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184085  qint16         0(0%)
head.layers.41.cls_layers.5.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184081  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018407   qint16         0(0%)
head.layers.28.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184067  qint16         0(0%)
backbone.stages.0.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018406   qint8          1012137984(0.05%)
head.layers.38.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184059  qint8          0(0%)
head.layers.21.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184059  qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184053  qint16         0(0%)
head.layers.24.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018405   qint16         0(0%)
head.layers.31.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018405   qint16         0(0%)
head.anchor_encoder.pos_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184043  qint16         0(0%)
head.layers.30.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018404   qint16         0(0%)
head.layers.13.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0184039  qint8          872415232(0.04%)
head.layers.5.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184035  qint16         0(0%)
head.layers.18.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184019  qint16         0(0%)
head.layers.3.point_matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184014  qint16         0(0%)
head.layers.3.camera_encoder.2.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184013  qint16         0(0%)
head.layers.3.camera_encoder.2.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0184     qint16         0(0%)
head.layers.0.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183996  qint8          0(0%)
head.layers.14.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018397   qint8          1744830464(0.08%)
backbone.stages.0.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183968  qint8          1012137984(0.05%)
head.layers.9.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183963  qint16         0(0%)
backbone.stages.3.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018396   qint8          12145655808(0.55%)
head.anchor_encoder.yaw_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183954  qint16         0(0%)
backbone.stages.3.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183949  qint8          12145655808(0.55%)
backbone.stages.3.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183948  qint8          94887936(0.00%)
backbone.stages.3.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183947  qint8          94887936(0.00%)
head.anchor_encoder.yaw_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183946  qint16         0(0%)
head.layers.8.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183946  qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183946  qint8          16613376(0.00%)
head.layers.20.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183945  qint8          0(0%)
head.layers.8.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183943  qint16         0(0%)
head.instance_bank.instance_feature_quant_stub       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183943  qint8          0(0%)
head.layers.6.layers.0                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.018394   qint8          872415232(0.04%)
head.layers.18.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183938  qint16         0(0%)
backbone.stages.1.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183936  qint8          506068992(0.02%)
head.layers.27.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183933  qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183931  qint8          8097103872(0.37%)
head.anchor_encoder.vel_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.018393   qint8          354418688(0.02%)
head.layers.16.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183928  qint16         0(0%)
head.layers.14.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183919  qint8          1744830464(0.08%)
head.layers.38.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183915  qint16         0(0%)
head.layers.40.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183914  qint8          0(0%)
head.layers.3.camera_encoder.2.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183904  qint16         0(0%)
head.layers.14.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183896  qint8          1744830464(0.08%)
head.layers.14.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183873  qint8          0(0%)
head.layers.27.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183871  qint16         37486592(0.00%)
head.layers.41.cls_layers.0                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183866  qint8          872415232(0.04%)
head.layers.2.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183862  qint16         0(0%)
head.layers.1.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183856  qint16         0(0%)
head.layers.35.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183853  qint8          0(0%)
head.layers.18.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183845  qint16         0(0%)
head.layers.12.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183842  qint16         0(0%)
head.layers.21.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183828  qint16         0(0%)
head.layers.10.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183827  qint16         0(0%)
head.layers.6.quality_layers.6                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183823  qint8          0(0%)
head.layers.10.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183822  qint8          479232(0.00%)
head.layers.40.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183821  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018382   qint16         0(0%)
head.layers.5.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183816  qint16         0(0%)
head.layers.15.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183808  qint8          3489660928(0.16%)
head.layers.6.layers.4.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183808  qint16         0(0%)
head.anchor_encoder.pos_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183801  qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183798  qint8          8097103872(0.37%)
head.layers.17.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183794  qint16         0(0%)
head.layers.7.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183792  qint16         0(0%)
head.layers.37.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183791  qint16         0(0%)
head.layers.22.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183789  qint8          3489660928(0.16%)
head.layers.8.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183788  qint8          0(0%)
head.layers.34.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183781  qint8          872415232(0.04%)
head.layers.23.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018378   qint16         0(0%)
head.layers.5.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183776  qint16         0(0%)
head.anchor_encoder.vel_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183774  qint16         0(0%)
head.layers.41.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183768  qint8          37486592(0.00%)
head.layers.35.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183764  qint8          1744830464(0.08%)
head.layers.13.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183764  qint16         0(0%)
head.layers.34.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183757  qint8          872415232(0.04%)
head.layers.15.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183755  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.5.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183748  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183746  qint16         0(0%)
head.layers.36.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183737  qint8          3489660928(0.16%)
head.layers.2.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183732  qint16         0(0%)
head.layers.26.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183732  qint16         0(0%)
head.layers.13.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183728  qint8          0(0%)
head.layers.30.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183721  qint16         0(0%)
head.layers.3.weight_softmax.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183721  qint16         0(0%)
head.layers.31.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018372   qint16         0(0%)
head.layers.41.quality_layers.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183717  qint16         0(0%)
head.layers.6.layers.4.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183716  qint8          0(0%)
head.layers.10.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183709  qint16         0(0%)
backbone.stages.3.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183707  qint8          94887936(0.00%)
head.layers.35.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183706  qint16         0(0%)
head.layers.38.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183703  qint8          872415232(0.04%)
head.layers.20.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183703  qint16         0(0%)
head.layers.29.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183696  qint16         0(0%)
head.layers.41.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183696  qint16         0(0%)
head.layers.20.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183696  qint16         0(0%)
head.layers.27.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183695  qint16         0(0%)
head.layers.36.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183693  qint8          3489660928(0.16%)
head.layers.38.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183692  qint16         0(0%)
head.layers.27.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018368   qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183672  qint16         0(0%)
head.layers.24.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183671  qint16         0(0%)
head.layers.0.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183667  qint16         0(0%)
head.layers.28.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183665  qint16         0(0%)
head.layers.22.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183665  qint8          0(0%)
head.layers.15.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183664  qint16         0(0%)
head.layers.34.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183663  qint8          872415232(0.04%)
head.layers.40.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183657  qint16         0(0%)
head.layers.22.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183657  qint8          3489660928(0.16%)
head.layers.7.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183656  qint8          3489660928(0.16%)
head.layers.27.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183643  qint8          0(0%)
head.layers.29.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183642  qint8          3489660928(0.16%)
backbone.stages.2.block.3.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183641  qint8          105431040(0.00%)
backbone.stages.3.block.4.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183639  qint8          0(0%)
head.layers.30.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183633  qint16         0(0%)
head.layers.14.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183632  qint8          1744830464(0.08%)
backbone.stages.2.block.3.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183632  qint8          8097103872(0.37%)
head.layers.38.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183631  qint16         0(0%)
head.anchor_encoder.vel_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183624  qint16         0(0%)
head.layers.6.layers.9.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183622  qint16         0(0%)
head.layers.35.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183619  qint8          1744830464(0.08%)
head.layers.24.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183615  qint16         0(0%)
backbone.stages.3.block.4.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018361   qint8          94887936(0.00%)
head.layers.36.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183609  qint8          3489660928(0.16%)
head.layers.34.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183605  qint16         0(0%)
head.layers.41.quality_layers.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183601  qint8          872415232(0.04%)
head.layers.21.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183599  qint16         0(0%)
head.layers.17.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183599  qint8          0(0%)
head.layers.17.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183599  qint16         0(0%)
head.layers.36.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183596  qint8          3489660928(0.16%)
head.layers.24.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183596  qint16         0(0%)
head.layers.20.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183594  qint8          872415232(0.04%)
head.layers.13.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183593  qint16         0(0%)
head.layers.8.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018359   qint8          3489660928(0.16%)
head.layers.27.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183582  qint16         0(0%)
head.layers.15.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183579  qint16         0(0%)
head.layers.34.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183576  qint16         0(0%)
head.layers.41.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183573  qint8          37486592(0.00%)
head.anchor_encoder.vel_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183565  qint16         0(0%)
head.layers.17.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183561  qint16         0(0%)
head.layers.4.layers.1                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183557  qint8          3489660928(0.16%)
head.layers.1.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183554  qint16         0(0%)
head.layers.25.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183544  qint16         0(0%)
head.layers.20.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183541  qint8          0(0%)
head.layers.15.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018354   qint16         0(0%)
head.layers.9.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183539  qint16         0(0%)
head.layers.13.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183537  qint16         0(0%)
head.layers.41.cls_layers.2.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183534  qint16         0(0%)
head.layers.5.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183532  qint16         0(0%)
head.layers.28.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183532  qint16         0(0%)
head.layers.11.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183531  qint16         0(0%)
head.layers.27.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018353   qint8          0(0%)
head.layers.3.camera_encoder.5.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183528  qint8          0(0%)
head.layers.34.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018352   qint16         0(0%)
head.anchor_encoder.yaw_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183519  qint16         0(0%)
head.layers.1.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183513  qint8          0(0%)
head.layers.38.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183507  qint16         0(0%)
head.layers.34.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183505  qint16         0(0%)
head.layers.6.layers.4.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183501  qint16         0(0%)
head.layers.17.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183499  qint16         0(0%)
head.layers.31.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183498  qint8          81788928(0.00%)
head.layers.8.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183496  qint8          3489660928(0.16%)
head.layers.10.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183494  qint16         0(0%)
head.layers.41.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183494  qint16         0(0%)
head.anchor_encoder.vel_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183492  qint16         0(0%)
head.layers.37.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183484  qint16         0(0%)
head.layers.31.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183484  qint16         0(0%)
head.layers.21.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183481  qint16         0(0%)
head.layers.31.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183481  qint16         0(0%)
head.layers.24.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183478  qint16         0(0%)
head.layers.41.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183475  qint16         0(0%)
head.layers.39.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183473  qint16         0(0%)
head.layers.10.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183466  qint16         0(0%)
head.layers.16.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183465  qint16         0(0%)
head.instance_bank.anchor_quant_stub                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183465  qint16         0(0%)
head.layers.27.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183461  qint16         0(0%)
head.layers.33.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183458  qint16         0(0%)
head.layers.34.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183457  qint16         0(0%)
head.layers.19.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183453  qint16         0(0%)
head.layers.3.camera_encoder.5.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183449  qint16         0(0%)
head.instance_bank.anchor_cat                        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183438  qint16         0(0%)
head.layers.38.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183433  qint16         0(0%)
head.layers.34.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183427  qint8          872415232(0.04%)
head.layers.14.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183425  qint16         0(0%)
head.layers.14.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183424  qint8          3489660928(0.16%)
head.layers.15.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018342   qint8          3489660928(0.16%)
head.layers.29.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183414  qint8          0(0%)
head.layers.3.camera_encoder.2.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183412  qint16         0(0%)
head.layers.17.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183407  qint16         0(0%)
head.anchor_encoder.pos_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183405  qint16         0(0%)
head.anchor_encoder.vel_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183405  qint16         0(0%)
neck.conv_add.1                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0183404  qint8          14394851328(0.66%)
head.layers.11.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183401  qint16         0(0%)
head.anchor_encoder.pos_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01834    qint16         0(0%)
head.layers.4.short_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0183397  qint8          1744830464(0.08%)
head.layers.37.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183396  qint16         0(0%)
head.layers.28.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183392  qint16         0(0%)
head.layers.39.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183384  qint16         0(0%)
head.layers.15.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018338   qint8          3489660928(0.16%)
head.layers.35.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183378  qint16         0(0%)
head.layers.7.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183377  qint8          3489660928(0.16%)
head.layers.38.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183376  qint8          0(0%)
head.layers.32.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183374  qint8          3489660928(0.16%)
head.layers.31.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018337   qint16         0(0%)
head.layers.12.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183367  qint16         0(0%)
head.layers.3.camera_encoder.0                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183362  qint8          479232(0.00%)
head.anchor_encoder.size_fc.8.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183358  qint16         0(0%)
head.layers.27.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183356  qint8          0(0%)
head.layers.34.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183353  qint8          0(0%)
head.layers.4.pre_norm.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183353  qint16         0(0%)
head.layers.41.quality_layers.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183352  qint16         0(0%)
head.layers.26.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018335   qint16         0(0%)
backbone.stages.3.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018335   qint8          94887936(0.00%)
head.layers.31.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183349  qint16         0(0%)
head.layers.41.quality_layers.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183349  qint8          0(0%)
head.layers.8.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183348  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183347  qint16         0(0%)
head.layers.17.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183346  qint8          479232(0.00%)
head.layers.27.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183345  qint8          872415232(0.04%)
head.layers.38.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183342  qint16         0(0%)
head.layers.22.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183339  qint8          3489660928(0.16%)
head.layers.41.cls_layers.5.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183336  qint16         0(0%)
head.layers.39.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183333  qint16         0(0%)
backbone.stages.2.block.3.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183332  qint8          8097103872(0.37%)
backbone.stages.3.block.4.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183331  qint8          12145655808(0.55%)
head.anchor_encoder.vel_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183328  qint16         0(0%)
head.layers.36.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183324  qint16         0(0%)
head.layers.7.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183324  qint16         0(0%)
head.layers.6.layers.2                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.018332   qint8          872415232(0.04%)
head.layers.1.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183319  qint16         0(0%)
head.layers.3.camera_encoder.2.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183317  qint16         0(0%)
backbone.stages.0.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183317  qint8          14394851328(0.66%)
head.layers.35.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183314  qint8          3489660928(0.16%)
head.layers.13.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183313  qint16         0(0%)
backbone.stage_norm.3                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0183312  qint8          0(0%)
head.layers.39.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183312  qint16         0(0%)
head.layers.14.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018331   qint16         0(0%)
head.layers.27.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183307  qint16         0(0%)
head.layers.39.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183306  qint16         0(0%)
head.layers.31.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01833    qint16         0(0%)
head.layers.0.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183299  qint8          3489660928(0.16%)
head.layers.41.cls_layers.5.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183298  qint16         0(0%)
head.layers.29.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183297  qint16         0(0%)
head.layers.6.add1                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183296  qint8          0(0%)
head.layers.34.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183295  qint16         0(0%)
head.layers.41.quality_layers.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183293  qint16         0(0%)
head.layers.38.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018329   qint8          0(0%)
head.layers.34.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183289  qint16         0(0%)
head.layers.41.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183288  qint16         0(0%)
head.layers.13.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183287  qint16         0(0%)
head.layers.20.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183287  qint8          872415232(0.04%)
head.layers.7.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183281  qint8          3489660928(0.16%)
head.layers.21.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183278  qint8          0(0%)
head.layers.10.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183271  qint16         0(0%)
head.layers.34.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183268  qint16         37486592(0.00%)
head.layers.34.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183266  qint16         0(0%)
head.layers.38.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183265  qint8          0(0%)
head.anchor_encoder.size_fc.8.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183259  qint16         0(0%)
head.layers.34.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183257  qint8          872415232(0.04%)
head.layers.13.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183251  qint16         0(0%)
head.layers.27.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018325   qint16         0(0%)
head.layers.22.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183249  qint16         0(0%)
head.layers.30.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183246  qint16         0(0%)
head.layers.10.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183243  qint16         0(0%)
head.layers.41.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183231  qint16         0(0%)
head.layers.31.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183231  qint8          0(0%)
head.anchor_encoder.yaw_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183231  qint16         0(0%)
head.layers.35.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183225  qint8          1744830464(0.08%)
head.layers.28.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183224  qint8          1744830464(0.08%)
head.layers.36.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183223  qint8          3489660928(0.16%)
head.layers.38.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183214  qint16         0(0%)
head.anchor_encoder.pos_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183214  qint16         0(0%)
head.anchor_encoder.size_fc.8.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183214  qint16         0(0%)
head.layers.21.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183207  qint16         0(0%)
head.layers.41.quality_layers.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183198  qint16         0(0%)
head.layers.3.weight_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183197  qint8          0(0%)
neck.fpn_conv.0.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183194  qint8          1036429295616(47.24%)
head.layers.38.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183194  qint16         0(0%)
head.layers.38.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183188  qint8          1308622848(0.06%)
head.anchor_encoder.vel_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183185  qint16         0(0%)
head.layers.17.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183184  qint16         0(0%)
head.layers.32.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183182  qint16         0(0%)
head.layers.41.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183181  qint8          872415232(0.04%)
head.layers.41.quality_layers.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183179  qint16         0(0%)
head.layers.34.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183176  qint16         0(0%)
head.layers.34.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183173  qint8          872415232(0.04%)
head.layers.19.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183172  qint16         0(0%)
head.layers.38.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018317   qint16         0(0%)
head.layers.14.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183169  qint16         0(0%)
backbone.stages.2.block.7.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183168  qint8          105431040(0.00%)
head.layers.3.camera_encoder.0                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183163  qint8          479232(0.00%)
head.layers.24.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018316   qint16         0(0%)
head.layers.17.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183155  qint16         0(0%)
head.layers.1.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018315   qint16         0(0%)
head.layers.24.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018315   qint16         0(0%)
head.layers.15.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183149  qint16         0(0%)
head.anchor_encoder.size_fc.5.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183148  qint16         0(0%)
head.layers.28.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183148  qint16         0(0%)
head.layers.14.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183145  qint16         0(0%)
head.layers.8.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018314   qint16         0(0%)
backbone.stages.2.block.2.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018314   qint8          0(0%)
head.layers.28.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183138  qint8          3489660928(0.16%)
head.layers.31.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183138  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183126  qint16         0(0%)
head.layers.34.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183125  qint16         0(0%)
head.layers.24.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183122  qint16         0(0%)
head.layers.29.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183121  qint16         0(0%)
head.layers.27.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183117  qint16         0(0%)
head.layers.24.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183113  qint8          0(0%)
head.layers.17.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183113  qint16         0(0%)
backbone.stages.0.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183108  qint8          0(0%)
head.layers.40.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183103  qint16         0(0%)
head.layers.35.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183102  qint16         0(0%)
head.layers.41.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183096  qint8          0(0%)
head.layers.41.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018309   qint16         0(0%)
head.layers.3.point_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018309   qint16         0(0%)
head.layers.36.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183089  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183084  qint16         0(0%)
backbone.stages.2.block.0.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183082  qint8          8097103872(0.37%)
backbone.stages.2.block.4.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183081  qint8          105431040(0.00%)
head.anchor_encoder.pos_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183078  qint16         0(0%)
head.layers.10.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183074  qint8          0(0%)
head.layers.24.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183067  qint8          0(0%)
backbone.stages.3.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183066  qint8          12145655808(0.55%)
backbone.stages.3.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183066  qint8          94887936(0.00%)
head.layers.10.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183061  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018306   qint16         0(0%)
head.layers.27.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183058  qint16         0(0%)
head.layers.25.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183056  qint16         0(0%)
head.layers.6.layers.4.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183053  qint16         0(0%)
head.layers.17.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183052  qint16         0(0%)
head.layers.3.point_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183041  qint16         0(0%)
head.layers.20.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183038  qint16         0(0%)
head.layers.31.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183036  qint8          0(0%)
head.layers.23.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183036  qint16         0(0%)
head.layers.38.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183034  qint16         0(0%)
head.layers.41.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183034  qint8          0(0%)
backbone.stages.0.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0183027  qint8          1012137984(0.05%)
head.layers.13.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0183026  qint8          872415232(0.04%)
head.layers.8.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183026  qint8          0(0%)
head.layers.3.kps_generator.keypoints_add            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183022  qint16         0(0%)
head.layers.28.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183019  qint16         0(0%)
head.layers.38.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183017  qint16         0(0%)
head.layers.22.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183017  qint16         0(0%)
head.layers.1.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183014  qint8          3489660928(0.16%)
head.layers.20.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018301   qint16         0(0%)
head.fc_before                                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0183009  qint16         15703474176(0.72%)
head.layers.41.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0183006  qint16         0(0%)
head.layers.27.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183005  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183005  qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0183001  qint8          0(0%)
head.layers.27.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0183001  qint16         0(0%)
head.layers.4.layers.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182987  qint8          6979321856(0.32%)
head.layers.38.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182985  qint16         0(0%)
head.layers.24.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182984  qint8          479232(0.00%)
head.layers.14.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182983  qint8          3489660928(0.16%)
head.layers.35.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182983  qint16         0(0%)
head.layers.35.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182981  qint16         0(0%)
head.layers.8.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182978  qint16         0(0%)
head.layers.0.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182974  qint8          1744830464(0.08%)
head.layers.17.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182973  qint8          0(0%)
head.layers.38.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182969  qint16         0(0%)
head.layers.38.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182968  qint16         0(0%)
head.layers.35.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182964  qint8          0(0%)
head.anchor_encoder.pos_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182963  qint8          1417674752(0.06%)
backbone.stage_norm.2                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.018296   qint8          0(0%)
head.layers.41.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182959  qint8          872415232(0.04%)
head.layers.29.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182958  qint16         0(0%)
head.anchor_encoder.vel_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182951  qint8          0(0%)
head.layers.20.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182951  qint16         0(0%)
head.layers.13.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018295   qint16         0(0%)
head.layers.27.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182949  qint16         0(0%)
head.anchor_encoder.vel_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182949  qint16         0(0%)
head.layers.38.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182945  qint8          0(0%)
head.layers.31.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182942  qint16         0(0%)
head.layers.20.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182942  qint16         37486592(0.00%)
head.layers.17.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182941  qint16         0(0%)
head.anchor_encoder.size_fc.5.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182937  qint16         0(0%)
backbone.stages.1.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182936  qint8          506068992(0.02%)
head.anchor_encoder.yaw_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182934  qint16         0(0%)
head.layers.3.camera_encoder.2.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182934  qint8          0(0%)
head.layers.2.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182933  qint16         0(0%)
head.anchor_encoder.pos_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182926  qint16         0(0%)
head.layers.31.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182916  qint16         0(0%)
head.layers.34.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182915  qint8          0(0%)
head.layers.36.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182914  qint16         0(0%)
head.anchor_encoder.size_fc.5.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182913  qint16         0(0%)
head.layers.27.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018291   qint8          37486592(0.00%)
neck.fpn_conv.2.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182909  qint8          64776830976(2.95%)
head.anchor_encoder.size_fc.8.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182902  qint16         0(0%)
head.layers.3.camera_encoder.5.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182902  qint16         0(0%)
head.layers.15.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182899  qint8          0(0%)
head.layers.35.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182893  qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182891  qint16         0(0%)
head.layers.35.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018289   qint8          1744830464(0.08%)
head.layers.17.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018289   qint16         81788928(0.00%)
head.layers.34.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182885  qint8          37486592(0.00%)
head.layers.38.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182885  qint16         0(0%)
head.layers.10.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182879  qint16         0(0%)
head.layers.17.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182879  qint16         0(0%)
head.layers.41.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182876  qint16         0(0%)
head.layers.31.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182864  qint16         0(0%)
head.layers.3.output_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182861  qint8          872415232(0.04%)
head.layers.31.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182855  qint8          1308622848(0.06%)
head.layers.3.camera_encoder.5.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182852  qint16         0(0%)
head.layers.24.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182848  qint8          10223616(0.00%)
head.anchor_encoder.yaw_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182847  qint16         0(0%)
head.layers.15.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182847  qint16         0(0%)
head.anchor_encoder.vel_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018283   qint8          0(0%)
head.layers.18.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182829  qint8          3489660928(0.16%)
head.layers.41.cls_layers.2.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182828  qint16         0(0%)
head.layers.24.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182826  qint16         0(0%)
head.layers.4.pre_norm.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182823  qint16         0(0%)
head.anchor_encoder.size_fc.5.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182819  qint8          0(0%)
head.layers.20.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182819  qint8          872415232(0.04%)
head.layers.41.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182815  qint8          872415232(0.04%)
head.layers.15.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182813  qint8          0(0%)
head.layers.41.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182806  qint16         0(0%)
head.layers.38.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182804  qint16         0(0%)
head.anchor_encoder.pos_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182804  qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182803  qint8          94887936(0.00%)
head.layers.28.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182801  qint8          0(0%)
backbone.stages.1.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182793  qint8          506068992(0.02%)
head.layers.10.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182791  qint8          0(0%)
backbone.stages.3.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0182785  qint8          4048551936(0.18%)
head.layers.24.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182784  qint16         81788928(0.00%)
head.layers.17.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182777  qint16         0(0%)
head.layers.36.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182775  qint8          0(0%)
head.layers.20.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018277   qint16         0(0%)
head.layers.24.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182766  qint16         0(0%)
head.layers.24.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182761  qint16         0(0%)
head.layers.17.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182757  qint16         0(0%)
head.layers.34.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182756  qint16         0(0%)
head.anchor_encoder.pos_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182755  qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182754  qint16         0(0%)
head.layers.17.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182746  qint8          0(0%)
head.layers.8.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182741  qint8          3489660928(0.16%)
head.layers.0.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182741  qint16         0(0%)
head.layers.35.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182737  qint8          3489660928(0.16%)
head.layers.20.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182735  qint16         0(0%)
head.layers.19.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182731  qint16         0(0%)
head.layers.35.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182729  qint16         0(0%)
head.layers.20.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182727  qint16         0(0%)
head.layers.38.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182725  qint16         0(0%)
head.layers.38.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182725  qint16         0(0%)
head.layers.5.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182723  qint16         0(0%)
head.anchor_encoder.pos_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182719  qint16         0(0%)
head.layers.3.feat_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182716  qint8          0(0%)
head.layers.24.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182713  qint16         0(0%)
backbone.stages.3.block.5.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182712  qint8          12145655808(0.55%)
head.layers.3.camera_encoder.3                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182711  qint8          10223616(0.00%)
head.layers.8.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018271   qint8          3489660928(0.16%)
head.layers.32.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182709  qint16         0(0%)
head.layers.17.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182708  qint8          0(0%)
backbone.stages.3.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182705  qint8          94887936(0.00%)
head.anchor_encoder.yaw_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182697  qint16         0(0%)
head.anchor_encoder.pos_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182696  qint16         0(0%)
head.layers.21.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182696  qint8          1744830464(0.08%)
neck.conv_add.1                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0182685  qint8          14394851328(0.66%)
head.layers.3.weights_fc                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182683  qint8          1308622848(0.06%)
head.anchor_encoder.pos_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182678  qint16         0(0%)
head.layers.31.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182676  qint8          10223616(0.00%)
head.layers.41.quality_layers.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182675  qint8          872415232(0.04%)
head.layers.20.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182674  qint16         0(0%)
head.layers.17.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182669  qint8          1308622848(0.06%)
head.layers.8.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182668  qint16         0(0%)
head.layers.38.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182663  qint16         0(0%)
head.layers.20.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182662  qint8          872415232(0.04%)
backbone.stages.0.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182661  qint8          1012137984(0.05%)
head.layers.23.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018266   qint16         0(0%)
head.layers.28.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182658  qint16         0(0%)
head.layers.27.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182656  qint8          872415232(0.04%)
head.layers.17.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018265   qint16         0(0%)
head.layers.10.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182649  qint16         0(0%)
head.anchor_encoder.pos_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182647  qint16         0(0%)
neck.fpn_conv.0.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182629  qint8          1036429295616(47.24%)
head.anchor_encoder.vel_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182625  qint8          0(0%)
head.layers.21.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182623  qint16         0(0%)
head.layers.24.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182621  qint8          10223616(0.00%)
backbone.stages.0.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182621  qint8          1012137984(0.05%)
head.anchor_encoder.vel_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182616  qint16         0(0%)
head.layers.14.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182611  qint8          3489660928(0.16%)
head.layers.6.layers.9.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182611  qint16         0(0%)
head.anchor_encoder.pos_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182611  qint8          33226752(0.00%)
head.layers.20.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182606  qint8          37486592(0.00%)
head.layers.20.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182604  qint16         0(0%)
head.layers.41.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182603  qint16         0(0%)
head.layers.31.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182601  qint8          10223616(0.00%)
head.layers.21.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182601  qint8          1744830464(0.08%)
head.layers.34.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182601  qint8          0(0%)
head.layers.31.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182594  qint16         0(0%)
head.layers.10.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182592  qint16         0(0%)
head.layers.21.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182583  qint8          3489660928(0.16%)
head.layers.41.cls_layers.2.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182578  qint16         0(0%)
head.layers.13.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182578  qint8          872415232(0.04%)
head.layers.34.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182577  qint8          872415232(0.04%)
head.layers.23.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182575  qint16         0(0%)
head.layers.17.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182573  qint16         0(0%)
backbone.stages.3.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182565  qint8          12145655808(0.55%)
head.layers.29.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182554  qint8          3489660928(0.16%)
head.layers.8.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182553  qint8          3489660928(0.16%)
head.layers.41.quality_layers.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182548  qint16         0(0%)
head.layers.10.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182543  qint16         0(0%)
head.layers.27.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182541  qint16         0(0%)
head.layers.9.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182539  qint16         0(0%)
head.layers.41.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182539  qint16         0(0%)
head.layers.13.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182538  qint16         0(0%)
head.layers.10.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182535  qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018253   qint8          105431040(0.00%)
head.layers.27.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182518  qint16         0(0%)
head.layers.17.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182518  qint16         0(0%)
head.layers.21.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182511  qint8          0(0%)
head.layers.31.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182507  qint16         0(0%)
head.layers.37.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182505  qint16         0(0%)
head.layers.7.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182501  qint8          0(0%)
head.anchor_encoder.size_fc.3                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182497  qint8          88604672(0.00%)
head.layers.12.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182495  qint16         0(0%)
head.layers.37.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182489  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182487  qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182483  qint16         0(0%)
head.layers.41.quality_layers.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182482  qint16         0(0%)
head.layers.10.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182478  qint16         0(0%)
head.layers.14.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182474  qint8          1744830464(0.08%)
head.layers.31.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182473  qint16         0(0%)
head.layers.25.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182469  qint8          6979321856(0.32%)
backbone.stage_norm.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>               0.0182469  qint8          0(0%)
head.layers.34.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182466  qint16         0(0%)
head.anchor_encoder.vel_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182465  qint16         0(0%)
head.layers.40.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182462  qint16         0(0%)
head.layers.29.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018246   qint8          0(0%)
head.layers.6.layers.9.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018245   qint16         0(0%)
head.layers.15.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182448  qint16         0(0%)
head.layers.17.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182446  qint16         0(0%)
head.layers.38.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182445  qint16         0(0%)
head.layers.41.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182443  qint8          872415232(0.04%)
head.layers.5.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182443  qint16         0(0%)
head.layers.8.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182442  qint8          0(0%)
head.layers.9.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182441  qint16         0(0%)
head.layers.7.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182436  qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182434  qint8          105431040(0.00%)
head.layers.28.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182434  qint8          1744830464(0.08%)
head.layers.41.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182429  qint8          872415232(0.04%)
head.layers.24.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182426  qint8          1308622848(0.06%)
head.layers.7.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182418  qint16         0(0%)
head.layers.28.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182415  qint8          0(0%)
head.layers.33.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182414  qint16         0(0%)
head.layers.20.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182406  qint8          0(0%)
head.anchor_encoder.pos_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.01824    qint16         0(0%)
head.layers.24.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182387  qint16         0(0%)
head.layers.6.layers.9.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182387  qint16         0(0%)
head.layers.0.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182385  qint8          1744830464(0.08%)
head.layers.30.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182385  qint16         0(0%)
backbone.stages.0.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182379  qint8          14394851328(0.66%)
head.layers.24.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182368  qint16         0(0%)
head.layers.18.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182365  qint8          6979321856(0.32%)
head.layers.3.weight_softmax.sum                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182362  qint16         0(0%)
head.layers.18.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182361  qint16         0(0%)
head.layers.39.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018236   qint16         0(0%)
neck.conv_add.2                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0182357  qint8          28789702656(1.31%)
head.layers.27.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182357  qint8          0(0%)
head.layers.38.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182356  qint16         0(0%)
head.layers.32.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182353  qint16         0(0%)
head.layers.10.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182348  qint16         0(0%)
head.layers.34.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182338  qint16         0(0%)
neck.conv_add.0                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0182336  qint8          5398069248(0.25%)
head.layers.17.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182331  qint16         0(0%)
head.layers.33.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182329  qint16         0(0%)
head.layers.17.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182324  qint16         0(0%)
head.layers.6.layers.4.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182321  qint16         0(0%)
head.layers.41.quality_layers.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182321  qint16         0(0%)
head.layers.13.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182319  qint8          37486592(0.00%)
head.layers.14.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182317  qint8          0(0%)
neck.fpn_conv.3.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182307  qint8          16194207744(0.74%)
head.layers.31.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182301  qint16         0(0%)
head.layers.41.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182299  qint8          872415232(0.04%)
head.layers.41.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182287  qint8          872415232(0.04%)
head.layers.2.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182285  qint16         0(0%)
head.layers.19.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182281  qint16         0(0%)
head.anchor_encoder.size_fc.5.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182278  qint16         0(0%)
head.layers.24.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182275  qint16         0(0%)
backbone.stages.3.block.5.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.018227   qint8          0(0%)
head.layers.41.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018227   qint16         0(0%)
head.layers.7.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182268  qint16         0(0%)
head.layers.36.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182268  qint16         0(0%)
head.layers.27.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182262  qint8          872415232(0.04%)
head.layers.29.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182262  qint8          3489660928(0.16%)
head.layers.20.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182247  qint8          0(0%)
head.layers.22.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182239  qint16         0(0%)
head.layers.17.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182238  qint8          10223616(0.00%)
head.anchor_encoder.vel_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182237  qint16         0(0%)
head.layers.4.pre_norm.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182236  qint16         0(0%)
head.layers.8.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182232  qint16         0(0%)
head.layers.17.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182229  qint16         0(0%)
head.layers.38.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182225  qint8          0(0%)
head.anchor_encoder.pos_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182211  qint16         0(0%)
head.layers.24.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182202  qint16         0(0%)
head.anchor_encoder.vel_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01822    qint8          0(0%)
head.layers.6.cls_layers.6                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018219   qint8          0(0%)
backbone.stages.2.block.6.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182187  qint8          0(0%)
head.layers.27.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182186  qint8          872415232(0.04%)
head.anchor_encoder.size_fc.8.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182179  qint16         0(0%)
head.layers.14.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182178  qint16         0(0%)
head.layers.29.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182175  qint16         0(0%)
head.layers.31.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018215   qint8          0(0%)
head.layers.3.point_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182147  qint16         0(0%)
backbone.stages.2.block.1.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182143  qint8          105431040(0.00%)
head.layers.40.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182143  qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182142  qint8          94887936(0.00%)
head.layers.41.cls_layers.2.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182141  qint16         0(0%)
head.layers.31.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182138  qint16         0(0%)
head.layers.41.quality_layers.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018213   qint16         0(0%)
head.layers.3.reciprocal_op.reciprocal               activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182126  qint16         0(0%)
head.layers.38.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182125  qint16         0(0%)
backbone.stages.0.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182124  qint8          14394851328(0.66%)
head.layers.41.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182123  qint8          13631488(0.00%)
head.anchor_encoder.vel_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182121  qint16         0(0%)
head.layers.41.cls_layers.5.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182117  qint16         0(0%)
head.layers.31.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182112  qint16         0(0%)
head.layers.20.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182111  qint8          0(0%)
head.layers.7.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182109  qint8          0(0%)
head.layers.34.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182108  qint8          0(0%)
head.layers.31.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182102  qint16         0(0%)
head.layers.4.pre_norm.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182102  qint16         0(0%)
head.layers.38.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182089  qint8          10223616(0.00%)
head.layers.14.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182074  qint16         0(0%)
backbone.stages.1.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0182071  qint8          506068992(0.02%)
neck.conv_add.2                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                    0.0182069  qint8          28789702656(1.31%)
head.layers.34.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182067  qint16         0(0%)
head.layers.0.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182066  qint16         0(0%)
head.layers.41.quality_layers.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182064  qint16         0(0%)
head.layers.8.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182061  qint16         0(0%)
head.layers.10.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182059  qint16         0(0%)
head.anchor_encoder.size_fc.2.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182057  qint8          0(0%)
head.layers.0.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182043  qint8          0(0%)
head.anchor_encoder.yaw_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182041  qint16         0(0%)
head.layers.31.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182039  qint16         0(0%)
head.anchor_encoder.vel_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182038  qint16         0(0%)
head.layers.10.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182036  qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0182034  qint16         0(0%)
head.layers.24.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182033  qint8          0(0%)
head.layers.39.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0182031  qint8          1744830464(0.08%)
head.layers.16.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182031  qint16         0(0%)
head.layers.31.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018203   qint8          0(0%)
head.layers.27.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182025  qint8          872415232(0.04%)
head.layers.24.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182022  qint16         0(0%)
head.layers.31.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182015  qint8          872415232(0.04%)
head.layers.17.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0182012  qint16         0(0%)
head.layers.35.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0182009  qint8          3489660928(0.16%)
head.layers.17.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0182006  qint16         0(0%)
head.layers.31.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0182006  qint8          479232(0.00%)
backbone.stages.3.block.5.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181984  qint8          12145655808(0.55%)
head.layers.2.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181981  qint16         0(0%)
head.layers.41.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181976  qint8          872415232(0.04%)
backbone.stages.1.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181973  qint8          506068992(0.02%)
backbone.stages.2.block.1.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.018197   qint8          105431040(0.00%)
head.layers.27.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181959  qint8          872415232(0.04%)
head.layers.17.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181953  qint16         0(0%)
head.layers.6.layers.4.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181941  qint16         0(0%)
backbone.stages.2.block.0.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181941  qint8          105431040(0.00%)
head.layers.3.camera_encoder.5.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181932  qint16         0(0%)
head.layers.22.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018193   qint8          3489660928(0.16%)
head.layers.38.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181923  qint16         0(0%)
head.layers.33.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181917  qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181911  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181907  qint8          8097103872(0.37%)
head.layers.10.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181907  qint16         0(0%)
head.layers.24.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181905  qint8          479232(0.00%)
head.layers.15.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181903  qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181897  qint16         0(0%)
head.layers.14.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181897  qint16         0(0%)
head.layers.0.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181897  qint8          3489660928(0.16%)
head.layers.2.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181895  qint16         0(0%)
head.layers.7.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181885  qint16         0(0%)
head.layers.21.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018188   qint8          0(0%)
head.layers.24.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018188   qint16         0(0%)
head.layers.6.layers.5                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181877  qint8          872415232(0.04%)
head.layers.26.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181875  qint16         0(0%)
head.layers.35.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181873  qint8          0(0%)
head.layers.24.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018187   qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181865  qint16         0(0%)
head.layers.30.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181863  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181861  qint16         0(0%)
head.layers.31.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181855  qint16         0(0%)
head.layers.34.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181849  qint16         0(0%)
head.layers.26.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181847  qint8          0(0%)
head.layers.1.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181845  qint16         0(0%)
head.layers.35.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181843  qint8          1744830464(0.08%)
head.layers.31.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181842  qint16         0(0%)
head.layers.14.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181832  qint8          3489660928(0.16%)
head.layers.7.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181827  qint8          1744830464(0.08%)
head.layers.34.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181821  qint16         0(0%)
head.layers.17.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181813  qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181808  qint16         0(0%)
backbone.stages.2.block.1.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181803  qint8          8097103872(0.37%)
head.anchor_encoder.yaw_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181798  qint16         0(0%)
head.layers.0.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181797  qint8          3489660928(0.16%)
head.layers.27.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181796  qint16         0(0%)
head.layers.38.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181795  qint16         0(0%)
head.layers.12.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181791  qint16         0(0%)
head.layers.12.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181787  qint16         0(0%)
head.layers.25.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181783  qint16         0(0%)
head.layers.41.quality_layers.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181779  qint16         0(0%)
head.layers.8.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181779  qint8          3489660928(0.16%)
head.layers.6.layers.4.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181774  qint16         0(0%)
head.layers.13.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181765  qint16         0(0%)
neck.fpn_conv.1.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181763  qint8          259107323904(11.81%)
backbone.stages.2.block.7.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181753  qint8          0(0%)
head.layers.13.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181753  qint8          872415232(0.04%)
neck.fpn_conv.3.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181752  qint8          16194207744(0.74%)
head.layers.13.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181752  qint16         0(0%)
head.layers.24.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018175   qint16         0(0%)
head.layers.28.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181747  qint8          1744830464(0.08%)
backbone.stages.1.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181745  qint8          7197425664(0.33%)
head.anchor_encoder.size_fc.11.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181741  qint16         0(0%)
head.layers.16.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181737  qint16         0(0%)
head.layers.13.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181736  qint8          0(0%)
head.layers.20.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181732  qint16         0(0%)
head.layers.24.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181723  qint8          0(0%)
head.anchor_encoder.yaw_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181716  qint16         0(0%)
head.layers.0.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181715  qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.11.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181714  qint16         0(0%)
head.layers.1.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181714  qint16         0(0%)
head.layers.3.camera_encoder.5.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181706  qint16         0(0%)
head.layers.31.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181703  qint8          0(0%)
head.layers.36.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.01817    qint16         0(0%)
head.layers.19.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181681  qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181675  qint16         0(0%)
head.layers.3.camera_encoder.5.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181668  qint16         0(0%)
head.anchor_encoder.yaw_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181662  qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181655  qint8          105431040(0.00%)
head.layers.10.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181652  qint16         81788928(0.00%)
head.layers.27.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181652  qint16         0(0%)
head.layers.4.pre_norm.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181649  qint16         0(0%)
head.layers.7.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181643  qint16         0(0%)
head.anchor_encoder.vel_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181642  qint8          354418688(0.02%)
head.layers.38.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181631  qint16         81788928(0.00%)
head.layers.31.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181629  qint16         0(0%)
backbone.stages.2.block.7.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181626  qint8          8097103872(0.37%)
head.layers.41.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181621  qint16         0(0%)
head.layers.41.cls_layers.2.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181618  qint16         0(0%)
head.layers.38.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181595  qint8          10223616(0.00%)
head.layers.17.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181588  qint16         0(0%)
head.layers.33.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018158   qint16         0(0%)
head.layers.0.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181576  qint16         0(0%)
backbone.stages.2.block.3.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181574  qint8          105431040(0.00%)
backbone.stages.1.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181552  qint8          506068992(0.02%)
head.layers.31.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181551  qint8          479232(0.00%)
head.layers.21.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018154   qint8          3489660928(0.16%)
head.layers.0.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181538  qint8          1744830464(0.08%)
head.layers.30.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181537  qint16         0(0%)
head.layers.19.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181535  qint16         0(0%)
head.layers.0.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0181501  qint16         0(0%)
head.layers.24.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181491  qint16         0(0%)
head.layers.13.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181478  qint8          0(0%)
head.layers.34.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181468  qint16         0(0%)
head.layers.21.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181455  qint8          3489660928(0.16%)
head.layers.26.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181452  qint16         0(0%)
head.layers.17.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181441  qint8          1308622848(0.06%)
head.layers.38.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018144   qint8          81788928(0.00%)
head.layers.15.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181437  qint16         0(0%)
head.layers.28.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181437  qint8          3489660928(0.16%)
head.layers.3.camera_encoder.2.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181424  qint16         0(0%)
head.layers.10.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181423  qint16         0(0%)
head.layers.3.point_quant_stub                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181399  qint16         0(0%)
head.layers.41.cls_layers.2.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181384  qint16         0(0%)
head.layers.38.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181383  qint8          479232(0.00%)
head.layers.10.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181375  qint16         0(0%)
head.layers.38.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181365  qint8          479232(0.00%)
head.layers.10.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181361  qint8          0(0%)
head.layers.23.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018136   qint16         0(0%)
head.layers.12.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181331  qint16         0(0%)
head.layers.6.layers.7                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181326  qint8          872415232(0.04%)
head.layers.26.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018132   qint16         0(0%)
head.layers.19.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181309  qint8          0(0%)
head.layers.18.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181306  qint16         0(0%)
head.layers.20.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181293  qint8          872415232(0.04%)
backbone.stages.2.block.4.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181278  qint8          105431040(0.00%)
head.layers.1.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181251  qint8          3489660928(0.16%)
backbone.stages.2.block.5.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181247  qint8          105431040(0.00%)
head.layers.0.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181246  qint16         0(0%)
head.layers.19.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181213  qint16         0(0%)
head.layers.5.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181206  qint16         0(0%)
head.layers.10.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181204  qint8          10223616(0.00%)
head.layers.0.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.018118   qint8          1744830464(0.08%)
head.layers.31.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181168  qint16         0(0%)
head.layers.0.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181145  qint8          1744830464(0.08%)
head.layers.39.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018111   qint8          0(0%)
head.layers.41.quality_layers.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181107  qint8          872415232(0.04%)
head.layers.33.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181103  qint16         0(0%)
head.anchor_encoder.size_fc.2.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181097  qint16         0(0%)
head.layers.1.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181094  qint8          3489660928(0.16%)
head.layers.33.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0181081  qint16         0(0%)
head.layers.13.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181077  qint8          872415232(0.04%)
head.layers.20.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181058  qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0181056  qint8          8097103872(0.37%)
head.layers.41.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0181051  qint8          6815744(0.00%)
head.layers.13.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181036  qint16         0(0%)
head.layers.9.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181036  qint16         0(0%)
head.layers.12.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181005  qint16         0(0%)
head.layers.31.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0181005  qint16         0(0%)
head.layers.20.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0181002  qint8          872415232(0.04%)
head.layers.21.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180997  qint8          1744830464(0.08%)
head.anchor_encoder.pos_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180949  qint16         0(0%)
head.layers.11.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0180922  qint8          6979321856(0.32%)
head.layers.41.cls_layers.2.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180889  qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180855  qint16         0(0%)
head.layers.7.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180826  qint8          1744830464(0.08%)
head.layers.24.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180815  qint8          872415232(0.04%)
head.layers.10.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180814  qint16         0(0%)
head.layers.24.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180805  qint8          0(0%)
head.layers.5.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180778  qint8          0(0%)
head.layers.21.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180756  qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180723  qint16         0(0%)
head.layers.24.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180706  qint16         0(0%)
head.layers.13.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180698  qint16         0(0%)
head.layers.41.quality_layers.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018069   qint16         0(0%)
head.layers.21.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180682  qint8          1744830464(0.08%)
head.layers.0.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180638  qint8          3489660928(0.16%)
head.layers.3.cam_add                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180614  qint8          0(0%)
backbone.stages.2.block.2.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                       0.0180613  qint8          105431040(0.00%)
head.anchor_encoder.size_fc.11.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180572  qint16         0(0%)
head.layers.31.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180525  qint16         0(0%)
head.layers.10.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0180524  qint8          10223616(0.00%)
head.layers.12.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018051   qint8          0(0%)
head.layers.5.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180492  qint16         0(0%)
head.layers.25.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018048   qint16         0(0%)
head.layers.37.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180444  qint16         0(0%)
head.layers.10.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180439  qint8          1308622848(0.06%)
head.layers.41.cls_layers.3                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0180347  qint8          872415232(0.04%)
head.layers.20.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180342  qint16         0(0%)
head.layers.33.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180341  qint8          0(0%)
head.anchor_encoder.yaw_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180339  qint8          0(0%)
head.layers.40.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.018033   qint16         0(0%)
head.layers.20.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.018032   qint8          872415232(0.04%)
head.layers.2.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180308  qint16         0(0%)
head.layers.13.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180306  qint16         0(0%)
head.layers.41.quality_layers.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180306  qint16         0(0%)
head.layers.0.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180293  qint8          0(0%)
head.anchor_encoder.size_fc.2.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180284  qint16         0(0%)
head.layers.25.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180264  qint16         0(0%)
head.layers.25.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180264  qint16         0(0%)
head.layers.41.quality_layers.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180255  qint16         0(0%)
head.layers.39.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180233  qint8          3489660928(0.16%)
head.layers.30.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.018023   qint16         0(0%)
head.layers.1.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180161  qint8          0(0%)
head.layers.1.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180145  qint8          3489660928(0.16%)
head.layers.4.pre_norm.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180144  qint16         0(0%)
head.layers.28.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0180117  qint8          3489660928(0.16%)
head.layers.37.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180075  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0180047  qint16         0(0%)
head.layers.41.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0180026  qint16         0(0%)
head.layers.41.quality_layers.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.017997   qint8          872415232(0.04%)
head.layers.11.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179925  qint16         0(0%)
head.layers.16.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0179923  qint16         0(0%)
head.layers.9.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0179893  qint16         0(0%)
head.layers.10.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179825  qint16         0(0%)
head.layers.32.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179789  qint16         0(0%)
head.layers.9.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179756  qint8          0(0%)
head.anchor_encoder.size_fc.2.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179742  qint16         0(0%)
head.layers.10.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0179734  qint8          872415232(0.04%)
head.layers.1.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0179721  qint8          3489660928(0.16%)
head.layers.23.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.017972   qint8          0(0%)
head.anchor_encoder.size_fc.5.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179712  qint16         0(0%)
head.anchor_encoder.pos_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179646  qint8          0(0%)
head.layers.18.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0179644  qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179559  qint8          0(0%)
head.layers.12.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0179557  qint16         0(0%)
head.anchor_encoder.pos_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0179494  qint16         0(0%)
head.layers.41.quality_layers.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179479  qint16         0(0%)
head.anchor_encoder.size_fc.5.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179474  qint16         0(0%)
head.layers.7.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0179331  qint8          1744830464(0.08%)
head.layers.25.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.017924   qint8          0(0%)
head.layers.4.pre_norm.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179073  qint16         0(0%)
head.layers.22.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0179041  qint8          3489660928(0.16%)
head.layers.17.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.017904   qint8          872415232(0.04%)
head.layers.25.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0179024  qint16         0(0%)
head.layers.21.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0179004  qint8          3489660928(0.16%)
head.layers.10.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0179003  qint8          479232(0.00%)
head.layers.31.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178959  qint8          0(0%)
head.layers.14.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178923  qint8          0(0%)
head.anchor_encoder.pos_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178864  qint16         0(0%)
head.layers.1.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0178748  qint16         0(0%)
head.layers.11.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                    0.0178631  qint8          1744830464(0.08%)
head.layers.1.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0178548  qint8          3489660928(0.16%)
head.layers.9.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178431  qint16         0(0%)
head.layers.16.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178368  qint16         0(0%)
head.layers.13.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0178367  qint8          0(0%)
head.layers.4.pre_norm.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178366  qint16         0(0%)
head.layers.7.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0178268  qint8          1744830464(0.08%)
head.layers.2.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178221  qint8          0(0%)
head.layers.23.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178159  qint16         0(0%)
head.layers.18.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0178047  qint8          0(0%)
head.layers.29.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0177964  qint8          3489660928(0.16%)
head.layers.32.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177956  qint16         0(0%)
head.layers.1.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0177854  qint8          3489660928(0.16%)
head.layers.3.residual_op                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177837  qint8          0(0%)
head.layers.11.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.017782   qint16         0(0%)
head.layers.4.pre_norm.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177776  qint16         0(0%)
head.anchor_encoder.pos_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0177726  qint16         0(0%)
head.anchor_encoder.vel_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177721  qint16         0(0%)
head.layers.25.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177655  qint16         0(0%)
head.layers.22.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0177631  qint8          3489660928(0.16%)
head.layers.39.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0177623  qint16         0(0%)
head.layers.25.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0177547  qint8          6979321856(0.32%)
head.layers.18.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177504  qint16         0(0%)
head.layers.4.pre_norm.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177437  qint8          0(0%)
head.layers.25.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0177177  qint8          3489660928(0.16%)
head.layers.35.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177124  qint16         0(0%)
head.layers.30.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0177069  qint8          0(0%)
head.layers.23.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176965  qint16         0(0%)
head.layers.32.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176885  qint8          0(0%)
head.anchor_encoder.yaw_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.017685   qint16         0(0%)
head.anchor_encoder.size_fc.11.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176702  qint16         0(0%)
head.layers.16.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176644  qint16         0(0%)
head.layers.11.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176623  qint16         0(0%)
head.anchor_encoder.pos_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176611  qint16         0(0%)
head.anchor_encoder.vel_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176609  qint16         0(0%)
head.layers.10.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0176227  qint16         0(0%)
head.layers.2.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0176172  qint16         0(0%)
head.layers.36.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.017558   qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.11.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>              0.0175392  qint16         0(0%)
head.layers.17.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0175386  qint8          872415232(0.04%)
head.layers.18.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0175224  qint16         0(0%)
head.layers.10.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0175181  qint8          0(0%)
head.layers.25.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0175103  qint16         0(0%)
head.anchor_encoder.yaw_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0175088  qint16         0(0%)
head.layers.11.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0174742  qint16         0(0%)
head.layers.32.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0174368  qint16         0(0%)
head.layers.15.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0173735  qint8          3489660928(0.16%)
head.layers.36.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0173291  qint8          3489660928(0.16%)
head.layers.11.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0172802  qint8          0(0%)
head.layers.15.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0172581  qint8          3489660928(0.16%)
head.layers.16.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0172534  qint16         0(0%)
head.layers.18.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.017198   qint8          3489660928(0.16%)
head.layers.18.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0171838  qint8          6979321856(0.32%)
head.layers.31.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0171039  qint8          0(0%)
head.layers.11.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                     0.0170831  qint16         0(0%)
head.layers.11.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.017001   qint8          6979321856(0.32%)
head.layers.11.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0169381  qint8          3489660928(0.16%)
head.layers.29.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0167388  qint8          3489660928(0.16%)
head.layers.29.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0165583  qint8          3489660928(0.16%)
head.layers.39.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0164307  qint8          6979321856(0.32%)
head.layers.18.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0163143  qint16         0(0%)
head.layers.39.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.016236   qint8          3489660928(0.16%)
head.layers.16.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0162289  qint8          0(0%)
head.layers.32.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                   0.0161704  qint8          6979321856(0.32%)
head.layers.32.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                       0.0161243  qint8          3489660928(0.16%)
head.layers.17.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0160271  qint8          0(0%)
head.layers.10.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0160088  qint8          0(0%)
head.layers.11.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  0.0159789  qint16         0(0%)