op_name                                              sensitive_type    op_type                                                                                   L1  quant_dtype    flops
---------------------------------------------------  ----------------  --------------------------------------------------------------------------  ----------------  -------------  ---------------------
head.layers.20.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.68007e+15  qint16         0(0%)
head.layers.27.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       6.18158e+14  qint16         0(0%)
head.layers.13.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.68797e+10  qint16         0(0%)
head.layers.34.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>  932325            qint16         0(0%)
head.layers.41.layers.9.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>      54.4943       qint16         0(0%)
head.anchor_encoder.size_fc.11.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       4.26802      qint16         0(0%)
head.anchor_encoder.yaw_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       4.10661      qint16         0(0%)
head.anchor_encoder.pos_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   4.0926       qint16         0(0%)
head.anchor_encoder.pos_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.92391      qint8          1417674752(0.06%)
head.anchor_encoder.yaw_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.75886      qint8          88604672(0.00%)
head.layers.6.layers.5                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.7523       qint8          872415232(0.04%)
head.layers.6.layers.0                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.7209       qint8          872415232(0.04%)
head.anchor_encoder.pos_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.65641      qint8          1417674752(0.06%)
head.layers.6.layers.2                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.56498      qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.48403      qint8          88604672(0.00%)
head.layers.6.layers.10                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            3.40521      qint8          37486592(0.00%)
head.anchor_encoder.pos_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.33374      qint8          1417674752(0.06%)
head.layers.6.layers.7                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        3.28406      qint8          872415232(0.04%)
head.anchor_encoder.vel_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       2.9902       qint16         0(0%)
head.layers.6.layers.9.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   2.94678      qint16         0(0%)
head.anchor_encoder.vel_fc.6                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        2.90851      qint8          354418688(0.02%)
head.fc_after                                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            2.66136      qint8          20937965568(0.95%)
head.anchor_encoder.pos_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        2.42667      qint8          33226752(0.00%)
head.layers.2.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   2.40429      qint16         0(0%)
head.anchor_encoder.vel_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        2.11516      qint8          354418688(0.02%)
head.anchor_encoder.size_fc.3                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.69039      qint8          88604672(0.00%)
head.layers.5.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          1.63638      qint16         0(0%)
head.anchor_encoder.pos_fc.11.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.61378      qint16         0(0%)
head.anchor_encoder.yaw_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.60616      qint8          5537792(0.00%)
head.anchor_encoder.yaw_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.50357      qint16         0(0%)
head.layers.4.short_add                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         1.45047      qint8          1744830464(0.08%)
head.anchor_encoder.size_fc.0                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.44499      qint8          8306688(0.00%)
head.anchor_encoder.yaw_fc.9                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.39962      qint8          88604672(0.00%)
head.anchor_encoder.pos_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.3336       qint16         0(0%)
head.anchor_encoder.size_fc.0                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.29664      qint8          8306688(0.00%)
head.layers.13.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.29148      qint16         0(0%)
head.anchor_encoder.pos_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.2694       qint16         0(0%)
head.layers.41.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.235        qint16         0(0%)
head.layers.34.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.23266      qint16         0(0%)
head.layers.27.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.21715      qint16         0(0%)
head.layers.20.add2                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.21606      qint16         0(0%)
head.anchor_encoder.pos_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.1965       qint16         0(0%)
head.anchor_encoder.pos_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.18867      qint8          0(0%)
head.layers.6.layers.11.scale_quant_stub             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          1.17619      qint16         0(0%)
head.anchor_encoder.pos_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.14107      qint8          1417674752(0.06%)
head.layers.5.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   1.13242      qint16         0(0%)
head.anchor_encoder.vel_fc.3                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.12768      qint8          354418688(0.02%)
head.anchor_encoder.size_fc.6                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.11993      qint8          88604672(0.00%)
head.anchor_encoder.size_fc.2.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.10517      qint16         0(0%)
head.layers.4.pre_norm.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.09784      qint8          0(0%)
head.anchor_encoder.vel_fc.0                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        1.07847      qint8          16613376(0.00%)
head.anchor_encoder.pos_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.05961      qint16         0(0%)
head.anchor_encoder.vel_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   1.05564      qint16         0(0%)
head.layers.6.add2                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.04994      qint16         0(0%)
head.layers.11.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         1.04869      qint8          1744830464(0.08%)
backbone.downsample_block.1.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            1.03982      qint8          10796138496(0.49%)
head.anchor_encoder.pos_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.01148      qint8          0(0%)
head.anchor_encoder.size_fc.2.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.00896      qint16         0(0%)
backbone.stages.0.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         1.00502      qint8          7197425664(0.33%)
head.layers.6.layers.11.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.0038       qint16         0(0%)
head.anchor_encoder.pos_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       1.00217      qint16         0(0%)
backbone.downsample_block.0.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.99184      qint8          14394851328(0.66%)
head.anchor_encoder.size_fc.2.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.982766     qint16         0(0%)
head.anchor_encoder.size_fc.9                        weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.980153     qint8          88604672(0.00%)
backbone.stages.0.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.977908     qint8          7197425664(0.33%)
head.mat_quant_stub                                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.977322     qint16         0(0%)
head.anchor_encoder.size_fc.2.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.970628     qint16         0(0%)
head.anchor_encoder.size_fc.2.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.970571     qint8          0(0%)
head.layers.6.layers.9.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.969199     qint8          0(0%)
backbone.stages.1.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.966473     qint8          7197425664(0.33%)
head.layers.3.weight_softmax.reciprocal.reciprocal   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.960355     qint16         0(0%)
head.layers.22.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.959696     qint16         0(0%)
backbone.patch_embed.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                        0.958349     qint8          6072827904(0.28%)
backbone.stages.2.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.957793     qint8          4048551936(0.18%)
backbone.stages.1.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.955784     qint8          7197425664(0.33%)
backbone.quant                                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.954763     qint8          0(0%)
head.layers.6.layers.9.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.953265     qint16         0(0%)
backbone.stages.0.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.952659     qint8          7197425664(0.33%)
backbone.stages.2.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.952369     qint8          4048551936(0.18%)
backbone.downsample_block.1.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.94502      qint8          10796138496(0.49%)
backbone.stages.0.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.944861     qint8          7197425664(0.33%)
neck.fpn_conv.2.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.942437     qint8          64776830976(2.95%)
backbone.patch_embed.1.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                        0.939967     qint8          32388415488(1.48%)
head.anchor_encoder.yaw_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.938339     qint8          5537792(0.00%)
backbone.downsample_block.0.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.934592     qint8          14394851328(0.66%)
head.anchor_encoder.size_fc.5.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.93273      qint16         0(0%)
backbone.patch_embed.1.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                        0.932722     qint8          32388415488(1.48%)
backbone.stages.1.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.932659     qint8          7197425664(0.33%)
head.layers.3.weight_softmax.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.930995     qint8          0(0%)
backbone.stages.2.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.93088      qint8          4048551936(0.18%)
head.anchor_encoder.pos_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.925715     qint16         0(0%)
head.layers.4.layers.1                               weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.923664     qint8          3489660928(0.16%)
head.layers.5.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.923221     qint8          0(0%)
head.layers.3.output_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.923076     qint8          872415232(0.04%)
backbone.stages.2.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.921544     qint8          4048551936(0.18%)
head.layers.6.layers.9.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.91924      qint16         0(0%)
head.anchor_encoder.pos_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.918377     qint16         0(0%)
backbone.stages.1.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.916252     qint8          7197425664(0.33%)
head.anchor_encoder.yaw_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.915966     qint16         0(0%)
head.layers.6.layers.10                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.914587     qint16         37486592(0.00%)
head.layers.3.output_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.913312     qint8          872415232(0.04%)
head.anchor_encoder.cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.911432     qint8          0(0%)
head.anchor_encoder.pos_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.911334     qint8          0(0%)
head.layers.3.kps_generator.offset                   weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.909906     qint8          81788928(0.00%)
head.layers.20.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.906491     qint16         0(0%)
head.layers.4.layers.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.904357     qint8          6979321856(0.32%)
head.layers.27.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.904338     qint16         0(0%)
head.layers.41.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.903299     qint8          872415232(0.04%)
backbone.stages.2.block.6.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.902868     qint8          4048551936(0.18%)
head.layers.5.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.902038     qint16         0(0%)
backbone.patch_embed.0.0                             weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvReLU2d'>                        0.901915     qint8          6072827904(0.28%)
backbone.stages.1.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.901313     qint8          7197425664(0.33%)
head.layers.13.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.900862     qint16         0(0%)
head.layers.20.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.899712     qint8          0(0%)
head.layers.3.feat_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.899653     qint8          0(0%)
head.layers.6.layers.4.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.898557     qint16         0(0%)
head.layers.12.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.898041     qint16         0(0%)
backbone.stages.2.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.891324     qint8          4048551936(0.18%)
head.layers.34.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.886292     qint16         0(0%)
head.layers.0.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.886155     qint8          3489660928(0.16%)
backbone.stages.2.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.883783     qint8          4048551936(0.18%)
head.layers.34.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.883286     qint16         0(0%)
head.anchor_encoder.vel_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.882876     qint8          354418688(0.02%)
head.anchor_encoder.vel_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.882305     qint8          16613376(0.00%)
head.layers.41.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.881738     qint8          872415232(0.04%)
head.layers.34.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.880495     qint16         0(0%)
head.layers.13.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.880399     qint16         0(0%)
head.layers.41.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.880097     qint8          872415232(0.04%)
head.layers.34.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.879938     qint16         0(0%)
head.anchor_encoder.pos_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.879173     qint16         0(0%)
neck.conv_add.0                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.877468     qint8          5398069248(0.25%)
head.layers.34.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.876573     qint16         0(0%)
head.layers.7.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.874452     qint16         0(0%)
head.layers.34.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.873433     qint8          0(0%)
head.anchor_encoder.pos_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.871909     qint16         0(0%)
backbone.downsample_block.2.proj.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.871901     qint8          8097103872(0.37%)
head.layers.5.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.871345     qint16         0(0%)
head.layers.4.pre_norm.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.870211     qint16         0(0%)
head.anchor_encoder.pos_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.870097     qint16         0(0%)
head.layers.34.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.869759     qint16         0(0%)
head.layers.0.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.869704     qint16         0(0%)
head.layers.8.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.869698     qint8          3489660928(0.16%)
backbone.stages.2.block.7.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.869603     qint8          4048551936(0.18%)
head.layers.9.rsqrt                                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.869557     qint16         0(0%)
head.layers.0.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.869351     qint8          1744830464(0.08%)
backbone.stages.3.block.0.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.869019     qint8          4048551936(0.18%)
head.layers.30.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.868788     qint16         0(0%)
backbone.stages.3.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.868771     qint8          0(0%)
head.anchor_encoder.pos_fc.0                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.868241     qint8          33226752(0.00%)
head.anchor_encoder.size_fc.5.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.867798     qint8          0(0%)
head.anchor_encoder.yaw_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.867331     qint16         0(0%)
head.anchor_encoder.size_fc.8.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.867292     qint16         0(0%)
head.layers.34.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.867174     qint8          37486592(0.00%)
head.layers.13.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.865836     qint16         0(0%)
backbone.stages.1.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.865296     qint8          0(0%)
head.layers.17.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.865071     qint16         0(0%)
head.layers.41.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.864853     qint8          872415232(0.04%)
backbone.stages.1.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.863448     qint8          7197425664(0.33%)
backbone.stages.1.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.863432     qint8          0(0%)
backbone.stages.2.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.863029     qint8          4048551936(0.18%)
head.layers.3.feat_sum                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.862209     qint8          0(0%)
head.layers.3.weights_fc                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.861431     qint8          1308622848(0.06%)
head.layers.32.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.860732     qint8          1744830464(0.08%)
head.layers.27.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.859998     qint16         0(0%)
backbone.stages.2.block.2.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.859827     qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.8.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.859089     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.858515     qint16         0(0%)
head.anchor_encoder.yaw_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.858269     qint16         0(0%)
head.layers.41.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.858259     qint16         0(0%)
backbone.stages.1.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.85793      qint8          0(0%)
head.layers.27.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.857529     qint16         0(0%)
head.anchor_encoder.pos_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.857508     qint16         0(0%)
head.layers.27.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.857116     qint16         0(0%)
head.anchor_encoder.vel_fc.2.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.856893     qint16         0(0%)
backbone.stages.3.block.2.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.856765     qint8          4048551936(0.18%)
backbone.stages.2.block.1.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.856623     qint8          8097103872(0.37%)
head.anchor_encoder.size_fc.6                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.856623     qint8          88604672(0.00%)
head.layers.27.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.855872     qint16         0(0%)
head.layers.12.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.855254     qint16         0(0%)
backbone.stages.2.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.854536     qint8          4048551936(0.18%)
head.anchor_encoder.size_fc.8.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.853436     qint8          0(0%)
head.layers.20.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.853053     qint16         0(0%)
head.layers.8.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.853025     qint8          0(0%)
backbone.downsample_block.2.proj.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.852784     qint8          8097103872(0.37%)
head.layers.4.pre_norm.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.85264      qint16         0(0%)
head.layers.34.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.851689     qint16         0(0%)
head.anchor_encoder.pos_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.85162      qint16         0(0%)
head.layers.10.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.851428     qint8          1308622848(0.06%)
head.layers.8.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.851351     qint16         0(0%)
head.layers.8.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.850625     qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.850089     qint16         0(0%)
head.layers.26.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.850013     qint8          0(0%)
head.layers.39.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.849506     qint8          1744830464(0.08%)
head.anchor_encoder.yaw_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.849001     qint16         0(0%)
head.layers.8.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.848728     qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.848597     qint16         0(0%)
head.layers.40.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.848191     qint16         0(0%)
head.layers.6.layers.9.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.847927     qint16         0(0%)
head.layers.4.pre_norm.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.847107     qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.846241     qint8          0(0%)
backbone.stages.2.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.846052     qint8          4048551936(0.18%)
head.layers.8.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.846024     qint8          3489660928(0.16%)
head.layers.3.residual_op                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.845328     qint8          0(0%)
head.anchor_encoder.pos_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.84524      qint8          1417674752(0.06%)
head.layers.27.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.844956     qint8          0(0%)
backbone.stages.3.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.844604     qint8          4048551936(0.18%)
head.layers.8.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.844318     qint8          3489660928(0.16%)
head.layers.13.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.843973     qint16         0(0%)
head.layers.6.layers.9.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.843785     qint16         0(0%)
head.anchor_encoder.pos_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.843774     qint8          1417674752(0.06%)
head.layers.34.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.843198     qint8          872415232(0.04%)
head.layers.36.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.842931     qint8          3489660928(0.16%)
head.layers.41.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.842538     qint8          872415232(0.04%)
head.anchor_encoder.size_fc.5.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.842485     qint16         0(0%)
head.layers.36.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.842424     qint8          3489660928(0.16%)
backbone.stages.1.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.841826     qint8          7197425664(0.33%)
head.layers.8.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.841812     qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.2.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.84143      qint16         0(0%)
head.layers.3.camera_encoder.5.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.841403     qint16         0(0%)
backbone.stages.1.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.841383     qint8          7197425664(0.33%)
backbone.stages.2.block.2.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.84114      qint8          0(0%)
head.anchor_encoder.vel_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.840924     qint16         0(0%)
head.layers.13.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.840302     qint16         0(0%)
backbone.stages.2.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.840222     qint8          4048551936(0.18%)
head.layers.25.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839918     qint16         0(0%)
backbone.stages.2.block.7.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.839729     qint8          4048551936(0.18%)
head.layers.20.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839695     qint16         0(0%)
head.layers.6.layers.4.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839607     qint16         0(0%)
head.anchor_encoder.size_fc.3                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.83951      qint8          88604672(0.00%)
head.layers.15.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.839431     qint8          3489660928(0.16%)
head.layers.22.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839365     qint16         0(0%)
head.anchor_encoder.yaw_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839114     qint8          0(0%)
head.layers.34.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.839107     qint8          872415232(0.04%)
head.layers.25.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.839034     qint8          0(0%)
head.layers.13.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.838621     qint8          0(0%)
head.layers.41.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.838542     qint16         0(0%)
head.anchor_encoder.pos_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.838082     qint16         0(0%)
backbone.stages.2.block.0.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.837799     qint8          8097103872(0.37%)
backbone.stages.3.block.4.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.837398     qint8          4048551936(0.18%)
backbone.stages.3.block.1.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.837368     qint8          4048551936(0.18%)
head.layers.30.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.837296     qint8          0(0%)
head.layers.7.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.837257     qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.836893     qint8          0(0%)
head.layers.41.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.836246     qint16         0(0%)
head.layers.2.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.836212     qint16         0(0%)
head.layers.17.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.836145     qint8          479232(0.00%)
head.anchor_encoder.vel_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.835893     qint8          0(0%)
head.anchor_encoder.vel_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.835509     qint16         0(0%)
head.layers.1.attn.k_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.835473     qint8          3489660928(0.16%)
head.layers.22.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.835458     qint8          3489660928(0.16%)
head.layers.41.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.83534      qint16         0(0%)
head.anchor_encoder.pos_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.835292     qint16         0(0%)
neck.conv_extract.3.0                                weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.835053     qint8          2699034624(0.12%)
head.layers.13.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.834646     qint8          872415232(0.04%)
head.layers.1.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.834558     qint8          3489660928(0.16%)
head.layers.37.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.834523     qint16         0(0%)
backbone.stages.3.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.834093     qint8          4048551936(0.18%)
head.layers.4.layers.1                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.833953     qint8          3489660928(0.16%)
head.layers.15.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.833812     qint8          3489660928(0.16%)
backbone.stages.3.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.833756     qint8          12145655808(0.55%)
head.layers.6.layers.9.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.833636     qint16         0(0%)
head.layers.15.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.833509     qint16         0(0%)
head.layers.6.layers.9.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.833302     qint16         0(0%)
head.layers.15.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.833196     qint16         0(0%)
head.layers.24.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.833174     qint8          872415232(0.04%)
backbone.stages.3.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.833016     qint8          0(0%)
head.layers.41.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.832944     qint16         0(0%)
head.layers.41.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.832895     qint16         0(0%)
head.layers.29.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.832809     qint16         0(0%)
head.layers.41.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.832617     qint8          872415232(0.04%)
head.layers.26.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.832444     qint16         0(0%)
head.layers.27.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.832057     qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.2.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.832041     qint8          0(0%)
head.layers.15.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.832034     qint8          3489660928(0.16%)
head.layers.5.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.831943     qint16         0(0%)
head.anchor_encoder.size_fc.11.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.831895     qint16         0(0%)
backbone.stages.2.block.3.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.831647     qint8          8097103872(0.37%)
backbone.stages.3.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.831598     qint8          12145655808(0.55%)
head.layers.2.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.831488     qint16         0(0%)
head.layers.22.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.831334     qint16         0(0%)
head.layers.33.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.830948     qint16         0(0%)
backbone.stages.3.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.830809     qint8          4048551936(0.18%)
head.layers.20.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.830704     qint8          872415232(0.04%)
backbone.stages.1.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.830672     qint8          506068992(0.02%)
head.layers.41.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.830507     qint16         0(0%)
head.layers.24.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.830418     qint8          0(0%)
head.layers.27.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.830164     qint16         0(0%)
head.layers.38.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.830048     qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.829886     qint16         0(0%)
head.layers.15.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.829835     qint8          3489660928(0.16%)
head.layers.0.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.829809     qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.829718     qint16         0(0%)
head.layers.13.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.829548     qint16         0(0%)
head.layers.17.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.829452     qint16         0(0%)
head.layers.13.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.829446     qint8          37486592(0.00%)
head.layers.34.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.829356     qint8          872415232(0.04%)
backbone.stages.3.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.829148     qint8          12145655808(0.55%)
backbone.stages.2.block.5.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.829034     qint8          8097103872(0.37%)
head.layers.17.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.828809     qint8          1308622848(0.06%)
head.layers.41.quality_layers.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.828726     qint16         0(0%)
head.layers.29.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.828596     qint16         0(0%)
backbone.stages.2.block.1.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.82849      qint8          0(0%)
head.anchor_encoder.vel_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.828483     qint8          354418688(0.02%)
head.anchor_encoder.size_fc.8.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.828471     qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.828444     qint16         0(0%)
backbone.stages.3.block.3.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.828431     qint8          4048551936(0.18%)
head.layers.10.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.828259     qint8          0(0%)
head.layers.34.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.828092     qint8          0(0%)
head.layers.35.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.827945     qint16         0(0%)
head.anchor_encoder.vel_fc.2.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.827851     qint16         0(0%)
head.layers.28.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.827756     qint8          3489660928(0.16%)
head.layers.3.camera_encoder.0                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.827691     qint8          479232(0.00%)
head.layers.24.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.827657     qint8          81788928(0.00%)
head.layers.3.weight_softmax.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.827532     qint16         0(0%)
backbone.stages.2.block.4.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.827504     qint8          8097103872(0.37%)
head.anchor_encoder.vel_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.827217     qint16         0(0%)
head.layers.38.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.826845     qint8          0(0%)
head.anchor_encoder.size_fc.8.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.82679      qint16         0(0%)
head.layers.34.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.826777     qint8          872415232(0.04%)
backbone.stages.2.block.0.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.826649     qint8          0(0%)
head.layers.23.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.826487     qint16         0(0%)
head.layers.41.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.826334     qint8          872415232(0.04%)
head.layers.10.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.826227     qint8          0(0%)
head.layers.19.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.826223     qint16         0(0%)
head.anchor_encoder.vel_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.826117     qint16         0(0%)
head.layers.13.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.825979     qint16         0(0%)
head.anchor_encoder.vel_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.825941     qint8          354418688(0.02%)
head.layers.9.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.825869     qint16         0(0%)
head.anchor_encoder.vel_fc.8.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8258       qint16         0(0%)
head.layers.28.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.825646     qint8          1744830464(0.08%)
head.layers.37.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.825458     qint8          0(0%)
backbone.stages.2.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.825429     qint8          4048551936(0.18%)
head.layers.8.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.825416     qint16         0(0%)
head.layers.38.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.825415     qint16         0(0%)
head.layers.34.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.825298     qint16         0(0%)
backbone.stages.3.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.825234     qint8          12145655808(0.55%)
head.layers.21.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.82494      qint8          0(0%)
head.layers.10.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.824825     qint16         0(0%)
head.layers.10.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.824786     qint16         0(0%)
head.layers.28.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.824755     qint8          0(0%)
head.layers.34.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.824752     qint16         0(0%)
backbone.stages.2.block.5.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.824723     qint8          105431040(0.00%)
backbone.stages.2.block.5.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.824679     qint8          105431040(0.00%)
head.layers.34.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.824659     qint8          872415232(0.04%)
head.layers.17.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.824557     qint16         0(0%)
backbone.stages.0.block.1.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.824483     qint8          0(0%)
head.layers.27.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.824458     qint8          872415232(0.04%)
backbone.stages.2.block.4.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.824451     qint8          8097103872(0.37%)
head.layers.4.pre_norm.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.824325     qint16         0(0%)
head.layers.33.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.824311     qint16         0(0%)
head.anchor_encoder.yaw_fc.3                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.824006     qint8          88604672(0.00%)
head.layers.41.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.823686     qint8          0(0%)
head.layers.17.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.823657     qint8          10223616(0.00%)
backbone.stages.2.block.2.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.823608     qint8          8097103872(0.37%)
head.layers.10.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.823498     qint16         0(0%)
backbone.stages.3.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.823491     qint8          12145655808(0.55%)
head.layers.10.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.823437     qint16         0(0%)
head.anchor_encoder.pos_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.823386     qint16         0(0%)
head.layers.0.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.823345     qint8          0(0%)
head.layers.33.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.823247     qint16         0(0%)
head.layers.29.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.823246     qint8          3489660928(0.16%)
backbone.stages.2.block.7.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.823183     qint8          105431040(0.00%)
head.layers.11.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.823174     qint8          6979321856(0.32%)
head.layers.7.attn.q_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.823139     qint8          3489660928(0.16%)
head.layers.24.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.823094     qint8          1308622848(0.06%)
backbone.stages.3.block.4.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.823075     qint8          12145655808(0.55%)
backbone.stages.3.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.822957     qint8          94887936(0.00%)
head.layers.8.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822945     qint8          3489660928(0.16%)
head.layers.28.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.822747     qint8          1744830464(0.08%)
head.layers.24.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.82266      qint8          0(0%)
head.layers.28.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822564     qint16         0(0%)
head.layers.10.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822553     qint8          0(0%)
head.layers.17.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.822515     qint8          479232(0.00%)
head.layers.19.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822471     qint8          0(0%)
head.layers.27.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822449     qint16         0(0%)
head.layers.10.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.822434     qint8          872415232(0.04%)
head.layers.35.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.822392     qint16         0(0%)
head.anchor_encoder.yaw_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822379     qint16         0(0%)
head.layers.31.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.822372     qint16         0(0%)
head.layers.24.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822372     qint16         0(0%)
backbone.stages.2.block.3.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.82234      qint8          8097103872(0.37%)
head.anchor_encoder.vel_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.822291     qint16         0(0%)
head.layers.6.quality_layers.6                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.822229     qint8          0(0%)
head.layers.24.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.82204      qint8          0(0%)
head.layers.40.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821985     qint16         0(0%)
head.layers.38.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821969     qint8          0(0%)
head.layers.17.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821894     qint16         0(0%)
head.fc_after                                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.821891     qint16         20937965568(0.95%)
head.layers.40.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821774     qint16         0(0%)
head.layers.34.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.821721     qint8          872415232(0.04%)
head.layers.20.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.821629     qint16         0(0%)
head.layers.34.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.821356     qint16         37486592(0.00%)
head.layers.28.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.821317     qint16         0(0%)
head.layers.13.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821294     qint8          0(0%)
head.layers.3.camera_encoder.3                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.821233     qint8          10223616(0.00%)
backbone.stages.2.block.6.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.821126     qint8          4048551936(0.18%)
neck.fpn_conv.0.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.821051     qint8          1036429295616(47.24%)
head.anchor_encoder.yaw_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.821039     qint16         0(0%)
head.layers.17.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.820976     qint16         0(0%)
head.layers.0.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.820863     qint8          3489660928(0.16%)
head.layers.27.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.820833     qint8          0(0%)
backbone.stages.3.block.4.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.820751     qint8          94887936(0.00%)
head.layers.15.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.820731     qint8          0(0%)
head.layers.32.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.820678     qint8          3489660928(0.16%)
backbone.stages.2.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.82067      qint8          4048551936(0.18%)
head.layers.39.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.820516     qint8          0(0%)
head.layers.7.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.820384     qint16         0(0%)
head.layers.3.weight_softmax.exp                     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.820366     qint16         0(0%)
head.layers.17.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.82027      qint16         0(0%)
backbone.stages.0.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.820257     qint8          7197425664(0.33%)
backbone.stages.3.block.0.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.820204     qint8          4048551936(0.18%)
backbone.stages.0.block.1.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.820161     qint8          14394851328(0.66%)
head.instance_bank.instance_feature_quant_stub       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.820156     qint8          0(0%)
head.layers.20.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.820041     qint8          872415232(0.04%)
head.anchor_encoder.size_fc.8.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.820037     qint16         0(0%)
head.layers.41.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.82003      qint16         0(0%)
head.layers.41.cls_layers.5.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819945     qint16         0(0%)
head.layers.31.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8199       qint16         0(0%)
head.layers.24.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.819872     qint16         0(0%)
head.layers.7.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819838     qint8          0(0%)
backbone.stages.2.block.7.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.819809     qint8          105431040(0.00%)
head.layers.3.camera_encoder.5.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.819766     qint16         0(0%)
head.layers.17.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.819763     qint16         0(0%)
head.anchor_encoder.vel_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.819699     qint16         0(0%)
head.layers.17.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81968      qint16         0(0%)
head.layers.0.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.819634     qint16         0(0%)
head.anchor_encoder.size_fc.11.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819449     qint16         0(0%)
head.layers.37.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819438     qint16         0(0%)
head.layers.41.cls_layers.5.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819408     qint16         0(0%)
head.layers.31.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819325     qint16         0(0%)
backbone.stages.0.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.819196     qint8          0(0%)
neck.conv_add.1                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.819185     qint8          14394851328(0.66%)
head.instance_bank.anchor_quant_stub                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.819171     qint16         0(0%)
head.layers.13.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819151     qint16         0(0%)
head.layers.31.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.819078     qint16         0(0%)
head.layers.15.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818967     qint8          0(0%)
head.layers.10.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.818918     qint16         0(0%)
head.layers.17.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.818909     qint16         81788928(0.00%)
head.anchor_encoder.pos_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818902     qint16         0(0%)
head.layers.6.layers.0                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.818837     qint8          872415232(0.04%)
head.layers.6.layers.4.bias_add                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818796     qint8          0(0%)
head.layers.6.layers.2                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.81879      qint8          872415232(0.04%)
head.layers.3.point_quant_stub                       activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.818733     qint16         0(0%)
head.layers.41.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818665     qint16         0(0%)
head.anchor_encoder.size_fc.5.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.8186       qint16         0(0%)
head.layers.17.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.818561     qint8          81788928(0.00%)
head.layers.7.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.818516     qint8          1744830464(0.08%)
head.layers.22.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818462     qint8          0(0%)
head.layers.36.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818437     qint16         0(0%)
head.layers.8.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818419     qint16         0(0%)
backbone.stages.0.block.1.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.818411     qint8          7197425664(0.33%)
head.layers.28.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818325     qint16         0(0%)
head.layers.0.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.818308     qint16         0(0%)
head.layers.27.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.818288     qint8          872415232(0.04%)
head.layers.3.camera_encoder.5.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.818263     qint16         0(0%)
head.layers.34.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.818233     qint8          0(0%)
head.layers.41.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.818086     qint16         0(0%)
head.layers.36.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.818045     qint16         0(0%)
head.layers.29.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817956     qint16         0(0%)
head.layers.24.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.81789      qint16         0(0%)
head.anchor_encoder.vel_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817799     qint16         0(0%)
head.layers.22.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817785     qint16         0(0%)
head.layers.21.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.817705     qint8          3489660928(0.16%)
head.layers.12.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817623     qint16         0(0%)
head.layers.10.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.817584     qint8          81788928(0.00%)
head.layers.0.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817554     qint16         0(0%)
head.layers.6.layers.4.weight_quant                  activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.817533     qint16         0(0%)
head.layers.15.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817514     qint16         0(0%)
neck.fpn_conv.1.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.817462     qint8          259107323904(11.81%)
head.layers.23.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817458     qint16         0(0%)
head.layers.32.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.817398     qint8          6979321856(0.32%)
head.layers.29.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.817339     qint8          3489660928(0.16%)
head.fc_before                                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.817329     qint16         15703474176(0.72%)
head.layers.24.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.817326     qint8          0(0%)
head.anchor_encoder.size_fc.2.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.817285     qint16         0(0%)
head.layers.38.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.817237     qint8          81788928(0.00%)
backbone.stages.0.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.817157     qint8          1012137984(0.05%)
head.layers.21.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.817117     qint16         0(0%)
head.layers.36.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.817097     qint16         0(0%)
head.layers.3.camera_encoder.2.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816996     qint16         0(0%)
backbone.stages.2.block.3.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.816959     qint8          105431040(0.00%)
head.layers.26.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816927     qint16         0(0%)
head.anchor_encoder.size_fc.8.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816917     qint16         0(0%)
head.layers.41.quality_layers.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816872     qint16         0(0%)
head.layers.22.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.816843     qint8          3489660928(0.16%)
head.layers.38.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816803     qint16         0(0%)
head.layers.38.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816738     qint8          0(0%)
head.layers.38.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816665     qint16         0(0%)
head.layers.31.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816577     qint8          0(0%)
head.layers.8.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816516     qint8          0(0%)
head.layers.30.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816462     qint16         0(0%)
head.layers.38.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.816434     qint8          1308622848(0.06%)
head.layers.25.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81636      qint16         0(0%)
head.layers.27.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.816285     qint16         0(0%)
head.layers.10.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.81627      qint8          10223616(0.00%)
head.layers.31.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.816245     qint8          1308622848(0.06%)
head.layers.27.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.816227     qint8          872415232(0.04%)
head.layers.38.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816211     qint8          0(0%)
head.layers.27.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81621      qint16         0(0%)
head.layers.34.layers.4.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816205     qint16         0(0%)
head.layers.41.quality_layers.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.816204     qint8          872415232(0.04%)
head.layers.37.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8162       qint16         0(0%)
head.layers.25.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.816196     qint8          3489660928(0.16%)
head.layers.15.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.816164     qint8          3489660928(0.16%)
head.layers.35.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816161     qint16         0(0%)
head.layers.31.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816134     qint16         0(0%)
head.layers.31.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.816107     qint8          10223616(0.00%)
head.layers.38.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816104     qint8          0(0%)
head.layers.39.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81607      qint16         0(0%)
head.layers.21.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.816029     qint16         0(0%)
head.layers.31.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.816012     qint16         0(0%)
head.layers.14.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815986     qint16         0(0%)
head.layers.18.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815969     qint8          0(0%)
head.layers.37.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.815956     qint16         0(0%)
head.layers.40.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815949     qint16         0(0%)
head.anchor_encoder.vel_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.81594      qint16         0(0%)
head.anchor_encoder.vel_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815934     qint16         0(0%)
head.layers.10.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815916     qint16         0(0%)
head.layers.31.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81591      qint8          0(0%)
head.layers.35.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.81591      qint8          1744830464(0.08%)
head.layers.0.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.815891     qint8          3489660928(0.16%)
head.layers.27.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.815871     qint16         0(0%)
head.layers.24.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815866     qint16         0(0%)
head.layers.38.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815823     qint16         0(0%)
head.layers.37.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815785     qint16         0(0%)
head.layers.35.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815771     qint8          1744830464(0.08%)
head.layers.0.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.815764     qint8          1744830464(0.08%)
head.layers.39.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.815738     qint8          3489660928(0.16%)
head.layers.41.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.815736     qint8          13631488(0.00%)
backbone.stages.2.block.2.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.815732     qint8          105431040(0.00%)
head.anchor_encoder.pos_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.815718     qint16         0(0%)
head.layers.41.cls_layers.2.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815634     qint16         0(0%)
head.layers.17.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815592     qint16         0(0%)
backbone.stages.3.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.815572     qint8          94887936(0.00%)
head.layers.29.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815522     qint8          0(0%)
head.layers.10.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.815517     qint8          10223616(0.00%)
head.layers.3.camera_encoder.2.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815508     qint16         0(0%)
head.layers.36.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815491     qint16         0(0%)
head.layers.38.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.815484     qint16         0(0%)
head.layers.32.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.815473     qint8          1744830464(0.08%)
head.layers.20.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815455     qint8          0(0%)
backbone.stages.0.block.2.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.815436     qint8          14394851328(0.66%)
head.layers.41.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.815421     qint8          6815744(0.00%)
head.layers.3.camera_encoder.2.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815391     qint8          0(0%)
head.layers.13.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815355     qint16         0(0%)
head.layers.17.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815352     qint8          0(0%)
head.layers.29.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815343     qint8          0(0%)
backbone.stage_norm.3                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>                    0.815335     qint8          0(0%)
head.layers.41.quality_layers.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815318     qint16         0(0%)
head.layers.41.quality_layers.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.815283     qint8          872415232(0.04%)
head.layers.3.camera_encoder.0                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.815273     qint8          479232(0.00%)
head.layers.11.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.815265     qint8          1744830464(0.08%)
backbone.stages.3.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.815182     qint8          12145655808(0.55%)
head.layers.13.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.815152     qint8          872415232(0.04%)
head.layers.41.quality_layers.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815142     qint8          0(0%)
head.layers.38.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.815123     qint16         0(0%)
neck.conv_add.1                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.815104     qint8          14394851328(0.66%)
head.layers.38.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815087     qint16         0(0%)
head.layers.4.pre_norm.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.815077     qint16         0(0%)
head.anchor_encoder.size_fc.2.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.815047     qint16         0(0%)
head.layers.31.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.815011     qint16         0(0%)
backbone.stages.3.block.4.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.814983     qint8          4048551936(0.18%)
head.layers.10.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814926     qint16         0(0%)
head.layers.32.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.81491      qint16         0(0%)
head.layers.18.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814907     qint16         0(0%)
head.layers.25.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.814882     qint8          3489660928(0.16%)
backbone.stages.1.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.814792     qint8          7197425664(0.33%)
head.layers.38.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814791     qint16         0(0%)
head.layers.27.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814689     qint16         0(0%)
head.layers.41.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814686     qint16         0(0%)
head.layers.0.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814641     qint16         0(0%)
head.layers.17.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814606     qint8          0(0%)
head.layers.29.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814587     qint8          0(0%)
head.layers.3.reciprocal_op.reciprocal               activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.814539     qint16         0(0%)
head.layers.17.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.814526     qint8          10223616(0.00%)
head.layers.7.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.814493     qint8          3489660928(0.16%)
head.layers.26.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814451     qint16         0(0%)
head.layers.27.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81444      qint16         0(0%)
backbone.stages.0.block.3.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.814373     qint8          14394851328(0.66%)
head.layers.31.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814372     qint16         0(0%)
head.layers.36.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.814304     qint8          3489660928(0.16%)
head.layers.37.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.814278     qint16         0(0%)
head.layers.12.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814268     qint16         0(0%)
head.layers.15.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.814259     qint8          3489660928(0.16%)
head.layers.24.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814116     qint16         0(0%)
head.layers.23.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.814108     qint16         0(0%)
head.layers.31.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814057     qint8          0(0%)
head.layers.41.cls_layers.2.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.814036     qint16         0(0%)
head.layers.25.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.814003     qint8          6979321856(0.32%)
head.layers.20.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813997     qint8          0(0%)
head.layers.38.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813926     qint16         0(0%)
head.layers.41.cls_layers.5.weight_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813895     qint16         0(0%)
head.layers.38.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813873     qint16         0(0%)
head.layers.17.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813872     qint8          0(0%)
head.anchor_encoder.vel_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81377      qint16         0(0%)
head.layers.28.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813751     qint8          0(0%)
head.layers.17.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81375      qint16         0(0%)
head.layers.39.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.813743     qint8          6979321856(0.32%)
head.layers.28.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813694     qint8          1744830464(0.08%)
head.layers.7.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813649     qint8          3489660928(0.16%)
head.layers.8.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81363      qint16         0(0%)
head.layers.38.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813624     qint16         81788928(0.00%)
head.layers.29.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81359      qint8          3489660928(0.16%)
head.layers.26.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813558     qint16         0(0%)
head.layers.33.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813555     qint16         0(0%)
head.layers.1.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813542     qint8          3489660928(0.16%)
head.layers.32.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813514     qint16         0(0%)
head.anchor_encoder.yaw_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813487     qint16         0(0%)
head.layers.18.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813456     qint16         0(0%)
head.layers.38.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.813418     qint16         0(0%)
head.anchor_encoder.vel_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.813396     qint16         0(0%)
head.layers.37.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813334     qint16         0(0%)
head.layers.3.point_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813317     qint16         0(0%)
head.layers.36.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813316     qint16         0(0%)
backbone.stages.2.block.6.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.813316     qint8          8097103872(0.37%)
head.layers.24.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.813291     qint16         0(0%)
head.layers.22.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813275     qint8          3489660928(0.16%)
head.layers.33.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.813261     qint16         0(0%)
head.layers.35.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813255     qint16         0(0%)
head.instance_bank.anchor_cat                        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813224     qint16         0(0%)
head.layers.24.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813195     qint16         0(0%)
head.anchor_encoder.size_fc.9                        activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.813172     qint8          88604672(0.00%)
head.layers.41.quality_layers.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.813143     qint8          872415232(0.04%)
head.layers.36.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813104     qint8          0(0%)
head.layers.14.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.813093     qint8          3489660928(0.16%)
head.layers.20.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.813083     qint8          872415232(0.04%)
head.layers.38.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.813025     qint16         0(0%)
head.layers.32.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812984     qint16         0(0%)
head.layers.29.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812882     qint16         0(0%)
head.layers.20.layers.11.mul                         activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812791     qint16         0(0%)
head.anchor_encoder.yaw_fc.8.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.812775     qint16         0(0%)
head.layers.10.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812765     qint16         0(0%)
head.layers.37.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.812737     qint16         0(0%)
head.layers.17.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812725     qint16         0(0%)
head.layers.3.kps_generator.keypoints_add            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812723     qint16         0(0%)
head.anchor_encoder.pos_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812665     qint16         0(0%)
head.layers.35.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812664     qint8          0(0%)
head.layers.0.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812617     qint16         0(0%)
head.layers.10.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81261      qint16         0(0%)
head.layers.17.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812588     qint8          0(0%)
head.layers.0.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812586     qint16         0(0%)
neck.fpn_conv.0.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.812584     qint8          1036429295616(47.24%)
head.layers.13.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812573     qint16         0(0%)
head.layers.3.weight_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812531     qint8          0(0%)
head.layers.28.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812482     qint16         0(0%)
head.layers.21.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81248      qint16         0(0%)
head.layers.21.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812479     qint8          1744830464(0.08%)
head.layers.41.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812478     qint16         0(0%)
head.layers.29.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.812466     qint8          3489660928(0.16%)
head.layers.36.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812465     qint8          3489660928(0.16%)
backbone.stages.2.block.5.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.812447     qint8          0(0%)
head.layers.10.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812438     qint16         0(0%)
head.layers.17.feat_sum                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812417     qint8          0(0%)
head.layers.35.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81241      qint8          0(0%)
head.layers.14.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.8124       qint8          1744830464(0.08%)
head.layers.24.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.812358     qint16         0(0%)
head.layers.39.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812329     qint16         0(0%)
head.layers.28.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.812312     qint8          1744830464(0.08%)
head.layers.14.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.812239     qint8          3489660928(0.16%)
head.layers.14.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812204     qint8          0(0%)
head.layers.38.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.812201     qint8          1308622848(0.06%)
head.layers.24.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.812198     qint16         0(0%)
head.layers.39.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812182     qint16         0(0%)
head.layers.31.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812164     qint8          0(0%)
neck.fpn_conv.3.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.812132     qint8          16194207744(0.74%)
backbone.stages.0.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.812026     qint8          1012137984(0.05%)
head.layers.24.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.812017     qint8          10223616(0.00%)
head.layers.25.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.812015     qint16         0(0%)
head.layers.14.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.811986     qint8          1744830464(0.08%)
head.layers.14.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811981     qint16         0(0%)
head.layers.17.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.811929     qint8          1308622848(0.06%)
head.layers.31.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811897     qint16         0(0%)
head.layers.16.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811892     qint8          0(0%)
head.layers.1.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81189      qint16         0(0%)
head.layers.13.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811857     qint16         0(0%)
head.layers.14.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811846     qint8          1744830464(0.08%)
head.layers.26.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811742     qint16         0(0%)
neck.fpn_conv.2.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.811735     qint8          64776830976(2.95%)
head.layers.15.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.811715     qint8          3489660928(0.16%)
head.layers.24.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811682     qint8          0(0%)
head.layers.35.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.811664     qint8          3489660928(0.16%)
head.layers.31.camera_encoder.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.811642     qint16         0(0%)
head.layers.27.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811575     qint16         0(0%)
head.layers.28.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811517     qint16         0(0%)
head.layers.8.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.8115       qint8          3489660928(0.16%)
head.layers.41.cls_layers.2.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811491     qint16         0(0%)
head.layers.41.quality_layers.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81145      qint16         0(0%)
head.layers.17.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811446     qint16         0(0%)
head.layers.38.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.811397     qint8          479232(0.00%)
head.layers.32.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.811395     qint16         0(0%)
backbone.stages.2.block.6.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.811388     qint8          105431040(0.00%)
head.layers.8.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811384     qint16         0(0%)
backbone.stages.3.block.3.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811379     qint8          0(0%)
head.layers.3.weight_softmax.sum                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811364     qint16         0(0%)
head.layers.0.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811323     qint8          1744830464(0.08%)
head.layers.3.camera_encoder.2.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811321     qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811315     qint16         0(0%)
head.layers.38.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811302     qint16         0(0%)
head.layers.3.weights_fc                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.81129      qint8          1308622848(0.06%)
head.layers.31.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811282     qint8          0(0%)
head.layers.27.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811251     qint8          0(0%)
head.layers.17.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.811247     qint16         0(0%)
head.layers.41.cls_layers.3                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.81121      qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.9                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.811206     qint8          88604672(0.00%)
backbone.stages.3.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.811192     qint8          12145655808(0.55%)
head.layers.1.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811171     qint16         0(0%)
head.layers.3.camera_encoder.5.var_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811155     qint16         0(0%)
head.layers.27.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.811148     qint16         0(0%)
head.layers.34.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.811142     qint8          872415232(0.04%)
head.layers.35.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811106     qint16         0(0%)
head.layers.31.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811102     qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.811056     qint8          14394851328(0.66%)
head.layers.41.quality_layers.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.811041     qint16         0(0%)
head.layers.29.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811027     qint16         0(0%)
head.layers.13.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.811026     qint8          872415232(0.04%)
head.layers.21.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.811012     qint8          3489660928(0.16%)
head.layers.24.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.811007     qint16         0(0%)
head.layers.24.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81092      qint16         0(0%)
head.layers.18.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.810879     qint8          6979321856(0.32%)
backbone.stages.2.block.1.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.810852     qint8          8097103872(0.37%)
head.layers.24.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810792     qint16         0(0%)
head.layers.24.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.810777     qint8          1308622848(0.06%)
head.layers.41.cls_layers.0                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.810762     qint8          872415232(0.04%)
head.layers.16.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810741     qint16         0(0%)
head.anchor_encoder.vel_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810734     qint16         0(0%)
head.layers.31.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810698     qint16         0(0%)
head.layers.31.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.810684     qint16         0(0%)
backbone.stages.0.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.810666     qint8          14394851328(0.66%)
head.layers.16.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.810656     qint16         0(0%)
head.layers.41.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.81065      qint8          37486592(0.00%)
head.layers.15.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810645     qint8          0(0%)
head.layers.36.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810628     qint8          0(0%)
head.layers.14.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.810579     qint8          1744830464(0.08%)
head.layers.10.weights_fc                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.810565     qint8          1308622848(0.06%)
head.layers.41.quality_layers.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.810547     qint16         0(0%)
head.layers.18.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.810506     qint8          6979321856(0.32%)
head.layers.17.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810496     qint16         0(0%)
head.layers.41.quality_layers.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.810488     qint16         0(0%)
head.layers.39.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.81041      qint16         0(0%)
head.layers.21.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810408     qint8          0(0%)
head.layers.4.short_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.810344     qint8          1744830464(0.08%)
head.layers.28.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810324     qint16         0(0%)
head.layers.24.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.810297     qint16         0(0%)
head.layers.21.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810273     qint8          1744830464(0.08%)
head.layers.29.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.810234     qint8          3489660928(0.16%)
head.layers.38.camera_encoder.3                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.810167     qint8          10223616(0.00%)
head.layers.9.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810144     qint16         0(0%)
head.layers.17.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810137     qint8          0(0%)
head.layers.24.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.810134     qint16         0(0%)
backbone.stages.1.block.1.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.810127     qint8          506068992(0.02%)
head.layers.20.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.81012      qint16         0(0%)
backbone.stages.3.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.810116     qint8          12145655808(0.55%)
head.layers.15.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.810045     qint8          3489660928(0.16%)
backbone.stages.0.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.809969     qint8          1012137984(0.05%)
head.layers.33.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.809954     qint16         0(0%)
head.layers.30.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809945     qint16         0(0%)
head.layers.24.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.809941     qint8          479232(0.00%)
backbone.stages.0.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.809823     qint8          1012137984(0.05%)
head.layers.17.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.809823     qint16         0(0%)
head.layers.17.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.80979      qint16         0(0%)
head.layers.3.camera_encoder.2.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.809787     qint16         0(0%)
head.layers.24.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809741     qint16         0(0%)
head.layers.16.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809727     qint16         0(0%)
head.layers.34.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.809722     qint16         0(0%)
head.layers.24.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8097       qint16         0(0%)
head.layers.39.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8097       qint16         0(0%)
neck.conv_add.0                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.809658     qint8          5398069248(0.25%)
head.layers.15.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.809656     qint8          3489660928(0.16%)
head.layers.23.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809615     qint16         0(0%)
head.layers.35.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.809574     qint8          1744830464(0.08%)
head.layers.17.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809571     qint16         0(0%)
backbone.stages.1.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.809556     qint8          7197425664(0.33%)
head.layers.13.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.80955      qint8          872415232(0.04%)
head.layers.23.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809519     qint16         0(0%)
head.layers.10.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809506     qint8          0(0%)
head.layers.10.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809506     qint16         0(0%)
head.layers.41.cls_layers.5.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809498     qint16         0(0%)
head.layers.22.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.809484     qint8          3489660928(0.16%)
head.layers.38.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809468     qint16         0(0%)
head.layers.31.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809462     qint16         0(0%)
head.layers.1.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809355     qint16         0(0%)
head.layers.31.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.80935      qint8          872415232(0.04%)
head.layers.22.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809314     qint16         0(0%)
head.layers.38.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.809309     qint16         0(0%)
backbone.stages.2.block.4.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.80929      qint8          105431040(0.00%)
head.layers.3.camera_encoder.5.bias_add              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80926      qint8          0(0%)
head.layers.17.point_matmul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809247     qint16         0(0%)
head.layers.7.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809229     qint8          0(0%)
head.layers.13.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.809219     qint8          872415232(0.04%)
head.layers.21.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.809214     qint8          1744830464(0.08%)
head.layers.35.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809213     qint8          0(0%)
head.anchor_encoder.size_fc.2.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809211     qint16         0(0%)
head.layers.10.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809192     qint16         0(0%)
head.anchor_encoder.pos_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.809189     qint16         0(0%)
head.layers.14.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809172     qint16         0(0%)
head.layers.3.feat_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80915      qint8          0(0%)
head.layers.38.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809133     qint16         0(0%)
head.layers.8.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8091       qint16         0(0%)
head.layers.19.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.809032     qint16         0(0%)
backbone.stages.0.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.80903      qint8          1012137984(0.05%)
head.layers.41.cls_layers.2.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.809002     qint16         0(0%)
head.layers.24.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808926     qint16         0(0%)
head.anchor_encoder.size_fc.2.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808886     qint16         0(0%)
head.layers.3.point_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80886      qint16         0(0%)
head.layers.17.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808838     qint8          0(0%)
head.layers.26.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808786     qint16         0(0%)
head.layers.38.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.808779     qint8          872415232(0.04%)
backbone.stages.3.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808716     qint8          0(0%)
head.layers.38.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.808709     qint16         0(0%)
head.layers.29.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808705     qint16         0(0%)
head.layers.39.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.808678     qint8          1744830464(0.08%)
head.layers.22.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808676     qint16         0(0%)
head.layers.41.cls_layers.2.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808557     qint16         0(0%)
head.layers.21.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.808541     qint8          3489660928(0.16%)
head.layers.30.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.808512     qint16         0(0%)
backbone.stages.2.block.0.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.808511     qint8          8097103872(0.37%)
head.layers.21.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808455     qint8          0(0%)
backbone.stages.2.block.6.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.808434     qint8          105431040(0.00%)
head.layers.16.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80843      qint16         0(0%)
head.layers.27.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808419     qint8          0(0%)
head.layers.41.quality_layers.2.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.808399     qint16         0(0%)
head.layers.40.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.808385     qint16         0(0%)
head.anchor_encoder.pos_fc.5.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80838      qint16         0(0%)
head.layers.15.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808358     qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.808338     qint16         0(0%)
head.layers.5.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808282     qint16         0(0%)
backbone.stage_norm.1                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>                    0.808209     qint8          0(0%)
head.layers.14.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.808176     qint8          3489660928(0.16%)
head.layers.35.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808166     qint8          1744830464(0.08%)
head.layers.0.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.808157     qint8          1744830464(0.08%)
head.layers.14.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808107     qint16         0(0%)
head.layers.41.quality_layers.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808071     qint8          0(0%)
head.layers.38.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.808058     qint16         0(0%)
head.layers.35.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.808004     qint8          1744830464(0.08%)
backbone.stages.0.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.807967     qint8          14394851328(0.66%)
head.layers.38.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807939     qint8          0(0%)
head.layers.22.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807927     qint16         0(0%)
head.layers.10.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807922     qint16         0(0%)
head.layers.14.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.807922     qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807893     qint16         0(0%)
backbone.stages.1.block.1.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.807856     qint8          7197425664(0.33%)
head.layers.9.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807823     qint16         0(0%)
head.layers.10.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.807762     qint16         0(0%)
head.layers.2.weight_quant                           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807754     qint16         0(0%)
head.layers.15.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807737     qint16         0(0%)
head.layers.34.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807724     qint16         0(0%)
head.layers.24.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807689     qint16         0(0%)
head.anchor_encoder.size_fc.8.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807604     qint16         0(0%)
backbone.stages.1.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.807593     qint8          506068992(0.02%)
head.layers.19.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80756      qint16         0(0%)
head.layers.10.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807533     qint16         0(0%)
head.layers.17.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80752      qint16         0(0%)
head.layers.30.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80747      qint16         0(0%)
head.layers.5.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807469     qint16         0(0%)
head.layers.38.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807461     qint8          0(0%)
head.layers.10.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.807455     qint16         81788928(0.00%)
head.anchor_encoder.vel_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807425     qint16         0(0%)
head.layers.13.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.807353     qint16         0(0%)
neck.fpn_conv.3.0                                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.807351     qint8          16194207744(0.74%)
head.layers.41.layers.5                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.807347     qint8          872415232(0.04%)
head.layers.5.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807341     qint16         0(0%)
head.layers.24.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807339     qint16         0(0%)
head.anchor_encoder.yaw_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.80733      qint16         0(0%)
head.layers.19.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.8073       qint16         0(0%)
backbone.stages.2.block.3.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.8073       qint8          105431040(0.00%)
head.layers.10.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.807272     qint8          479232(0.00%)
head.layers.17.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807245     qint16         0(0%)
head.layers.41.quality_layers.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807238     qint16         0(0%)
backbone.stages.0.block.0.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.807236     qint8          0(0%)
head.layers.10.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807174     qint8          0(0%)
head.layers.24.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807168     qint16         0(0%)
head.layers.16.var_mean.pre_mean                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807166     qint16         0(0%)
backbone.stages.2.block.1.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.807136     qint8          105431040(0.00%)
head.layers.41.quality_layers.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.807076     qint16         0(0%)
head.layers.37.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807072     qint16         0(0%)
head.layers.24.feat_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.807045     qint8          0(0%)
head.layers.38.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.806981     qint8          10223616(0.00%)
head.layers.41.quality_layers.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.806957     qint8          872415232(0.04%)
neck.conv_add.2                                      weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.806938     qint8          28789702656(1.31%)
head.layers.1.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806924     qint8          0(0%)
backbone.stages.1.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.806907     qint8          506068992(0.02%)
backbone.stage_norm.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>                    0.806868     qint8          0(0%)
head.layers.1.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80686      qint8          0(0%)
head.layers.3.camera_encoder.5.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806744     qint16         0(0%)
head.layers.41.quality_layers.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806728     qint16         0(0%)
head.layers.24.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806685     qint16         0(0%)
head.layers.24.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80668      qint16         0(0%)
head.layers.17.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.806679     qint16         0(0%)
neck.conv_add.2                                      activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.806679     qint8          28789702656(1.31%)
head.layers.24.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.806678     qint16         0(0%)
head.layers.2.input_mean.pre_mean                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806652     qint16         0(0%)
head.layers.35.attn.softmax.sum                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80664      qint16         0(0%)
head.layers.41.quality_layers.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806623     qint16         0(0%)
neck.fpn_conv.1.0                                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.806535     qint8          259107323904(11.81%)
head.layers.10.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8065       qint16         0(0%)
head.layers.27.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.806429     qint16         37486592(0.00%)
head.anchor_encoder.pos_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.806397     qint16         0(0%)
head.layers.24.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.806395     qint16         81788928(0.00%)
backbone.stages.0.block.2.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.806337     qint8          7197425664(0.33%)
head.instance_bank.feature_cat                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806335     qint8          0(0%)
head.layers.28.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.80633      qint8          1744830464(0.08%)
head.layers.20.layers.7                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.806312     qint8          872415232(0.04%)
head.layers.13.quality_layers.6                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.806303     qint8          0(0%)
head.layers.41.cls_layers.5.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806285     qint8          0(0%)
head.layers.13.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.806279     qint8          0(0%)
head.layers.21.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806255     qint16         0(0%)
head.layers.6.cls_layers.6                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.806224     qint8          0(0%)
head.layers.33.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80617      qint16         0(0%)
head.layers.27.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806155     qint16         0(0%)
head.layers.40.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806107     qint16         0(0%)
head.layers.17.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806081     qint16         0(0%)
head.layers.38.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.806056     qint16         0(0%)
head.layers.24.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.806025     qint8          479232(0.00%)
head.anchor_encoder.pos_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.806023     qint16         0(0%)
head.anchor_encoder.yaw_fc.5.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.806        qint16         0(0%)
head.layers.7.key_cat                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805979     qint16         0(0%)
head.layers.4.pre_norm.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80597      qint16         0(0%)
head.layers.28.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805939     qint16         0(0%)
head.layers.41.cls_layers.5.var_mean.pre_mean        activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805906     qint16         0(0%)
head.layers.31.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.805898     qint8          479232(0.00%)
head.layers.30.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.805896     qint16         0(0%)
head.anchor_encoder.size_fc.11.bias_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.805841     qint16         0(0%)
head.layers.24.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.805791     qint8          10223616(0.00%)
head.layers.36.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.805731     qint8          3489660928(0.16%)
head.layers.8.attn.q_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.805628     qint8          3489660928(0.16%)
head.layers.41.cls_layers.3                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.80558      qint8          872415232(0.04%)
backbone.stages.3.block.4.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.805578     qint8          94887936(0.00%)
head.layers.24.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.805563     qint16         0(0%)
head.layers.17.point_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805563     qint16         0(0%)
head.layers.41.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.805558     qint16         0(0%)
head.anchor_encoder.vel_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.8055       qint16         0(0%)
head.layers.15.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805496     qint16         0(0%)
head.anchor_encoder.pos_fc.5.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805487     qint16         0(0%)
head.layers.1.attn.k_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.805424     qint8          3489660928(0.16%)
head.layers.6.layers.4.rsqrt                         activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.805408     qint16         0(0%)
head.layers.36.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.805316     qint8          3489660928(0.16%)
head.layers.8.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805263     qint8          0(0%)
head.layers.31.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805249     qint16         0(0%)
head.layers.41.cls_layers.5.rsqrt                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.805192     qint16         0(0%)
head.layers.41.cls_layers.2.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.805159     qint16         0(0%)
backbone.stages.2.block.1.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.805148     qint8          105431040(0.00%)
head.layers.38.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805036     qint16         0(0%)
head.layers.12.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.805032     qint16         0(0%)
head.layers.41.quality_layers.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80503      qint16         0(0%)
head.layers.7.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.80497      qint16         0(0%)
head.layers.20.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.804893     qint16         0(0%)
head.layers.31.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.804858     qint8          479232(0.00%)
head.layers.33.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804828     qint16         0(0%)
head.layers.10.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804768     qint16         0(0%)
head.layers.6.layers.5                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.804751     qint8          872415232(0.04%)
head.layers.24.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804741     qint8          0(0%)
head.layers.16.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80474      qint16         0(0%)
backbone.stages.1.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.804726     qint8          506068992(0.02%)
head.layers.1.attn.softmax.sub                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804725     qint16         0(0%)
head.layers.23.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804695     qint16         0(0%)
head.layers.10.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804613     qint16         0(0%)
head.layers.26.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804594     qint16         0(0%)
head.layers.16.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804568     qint16         0(0%)
head.anchor_encoder.vel_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804531     qint16         0(0%)
head.layers.21.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80453      qint16         0(0%)
head.layers.20.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.80451      qint8          0(0%)
head.layers.10.camera_encoder.2.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804497     qint16         0(0%)
head.layers.12.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804382     qint16         0(0%)
head.layers.3.camera_encoder.5.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804358     qint16         0(0%)
head.layers.21.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.804358     qint8          1744830464(0.08%)
head.layers.7.attn.softmax.sum                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804344     qint16         0(0%)
head.layers.34.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804317     qint16         0(0%)
head.layers.41.cls_layers.0                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.804316     qint8          872415232(0.04%)
head.layers.18.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.804271     qint16         0(0%)
head.layers.39.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804258     qint16         0(0%)
head.layers.3.camera_encoder.2.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804244     qint16         0(0%)
head.layers.3.point_matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804239     qint16         0(0%)
head.anchor_encoder.vel_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804235     qint16         0(0%)
head.layers.41.cls_layers.5.weight_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.804211     qint16         0(0%)
head.layers.41.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80418      qint16         0(0%)
head.layers.36.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.804163     qint8          3489660928(0.16%)
head.layers.32.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.804105     qint8          3489660928(0.16%)
head.layers.38.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804049     qint16         0(0%)
head.layers.40.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804037     qint16         0(0%)
head.layers.3.camera_encoder.5.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.804016     qint16         0(0%)
backbone.stages.3.block.3.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.803918     qint8          94887936(0.00%)
backbone.stages.0.block.3.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.803917     qint8          7197425664(0.33%)
head.layers.7.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803902     qint8          1744830464(0.08%)
head.layers.29.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.803887     qint8          3489660928(0.16%)
head.layers.17.camera_encoder.5.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803845     qint16         0(0%)
backbone.stages.0.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.803821     qint8          1012137984(0.05%)
head.layers.41.cls_layers.5.bias_quant               activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.803802     qint16         0(0%)
backbone.stages.2.block.4.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.803729     qint8          0(0%)
head.layers.10.camera_encoder.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803686     qint16         0(0%)
head.layers.41.cls_layers.2.bias_add                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803674     qint8          0(0%)
head.layers.31.camera_encoder.5.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803636     qint8          0(0%)
head.layers.41.cls_layers.2.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803633     qint16         0(0%)
head.layers.39.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.803617     qint16         0(0%)
head.layers.6.layers.7                               activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.8036       qint8          872415232(0.04%)
head.layers.5.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803581     qint16         0(0%)
head.anchor_encoder.yaw_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.803504     qint16         0(0%)
head.layers.31.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.803501     qint16         0(0%)
head.layers.14.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803474     qint8          0(0%)
head.layers.31.camera_encoder.2.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.803425     qint16         0(0%)
head.layers.11.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803403     qint16         0(0%)
head.layers.31.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.803377     qint8          872415232(0.04%)
head.layers.31.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803365     qint16         0(0%)
head.layers.35.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.80334      qint8          3489660928(0.16%)
head.layers.22.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803229     qint8          3489660928(0.16%)
head.layers.24.weight_softmax.sum                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803196     qint16         0(0%)
head.layers.11.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.803184     qint8          3489660928(0.16%)
head.anchor_encoder.vel_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.803094     qint16         0(0%)
head.layers.25.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.803089     qint8          6979321856(0.32%)
head.layers.14.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803047     qint8          0(0%)
head.layers.1.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.803017     qint8          3489660928(0.16%)
backbone.stages.2.block.3.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.80301      qint8          0(0%)
head.layers.13.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.80299      qint16         0(0%)
head.layers.21.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.802927     qint8          1744830464(0.08%)
head.anchor_encoder.yaw_fc.5.rsqrt                   activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.802834     qint16         0(0%)
head.layers.30.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802807     qint16         0(0%)
head.layers.10.camera_encoder.0                      activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.802802     qint8          479232(0.00%)
head.layers.41.quality_layers.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802791     qint16         0(0%)
head.layers.28.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.802755     qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.6                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.802729     qint8          88604672(0.00%)
backbone.stages.2.block.4.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802725     qint8          105431040(0.00%)
head.anchor_encoder.pos_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.802694     qint16         0(0%)
head.layers.31.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802672     qint16         0(0%)
head.layers.14.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.802643     qint16         0(0%)
head.layers.25.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.802546     qint16         0(0%)
head.layers.24.camera_encoder.5.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802488     qint16         0(0%)
backbone.stages.3.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802469     qint8          94887936(0.00%)
head.layers.14.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80241      qint8          1744830464(0.08%)
head.layers.17.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802408     qint16         0(0%)
head.layers.31.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.802401     qint16         0(0%)
head.layers.2.mul                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802388     qint16         0(0%)
head.layers.13.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.802382     qint8          872415232(0.04%)
neck.conv_extract.3.0                                activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802371     qint8          2699034624(0.12%)
backbone.stages.3.block.5.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802331     qint8          12145655808(0.55%)
head.layers.6.layers.4.mul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802283     qint16         0(0%)
head.layers.18.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.802254     qint8          3489660928(0.16%)
head.layers.41.cls_layers.2.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802234     qint16         0(0%)
head.layers.27.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.802225     qint16         0(0%)
head.layers.34.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.802206     qint16         0(0%)
head.anchor_encoder.vel_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.802195     qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802145     qint8          94887936(0.00%)
head.layers.31.weight_softmax.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.802093     qint16         0(0%)
head.anchor_encoder.vel_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.802075     qint16         0(0%)
backbone.stages.2.block.5.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.802018     qint8          8097103872(0.37%)
head.layers.31.kps_generator.offset                  weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.801954     qint8          81788928(0.00%)
head.layers.3.point_mul                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801937     qint16         0(0%)
backbone.stage_norm.2                                activation        <class 'horizon_plugin_pytorch.nn.qat.batchnorm.BatchNorm2d'>                    0.801895     qint8          0(0%)
head.layers.21.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801876     qint16         0(0%)
head.layers.36.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801784     qint16         0(0%)
head.layers.29.query_cat                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801724     qint16         0(0%)
head.anchor_encoder.pos_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801567     qint16         0(0%)
backbone.stages.0.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.801511     qint8          1012137984(0.05%)
head.layers.41.cls_layers.2.input_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801395     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.801357     qint16         0(0%)
head.anchor_encoder.vel_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801347     qint16         0(0%)
head.layers.38.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80134      qint16         0(0%)
head.layers.3.camera_encoder.3                       activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.801269     qint8          10223616(0.00%)
head.layers.7.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.801207     qint8          1744830464(0.08%)
head.layers.41.cls_layers.5.out_mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80116      qint16         0(0%)
head.layers.2.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801154     qint16         0(0%)
head.layers.19.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.801151     qint16         0(0%)
head.layers.35.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.801108     qint8          3489660928(0.16%)
head.layers.10.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.801108     qint16         0(0%)
backbone.stages.2.block.2.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.801007     qint8          105431040(0.00%)
backbone.stages.2.block.7.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.800981     qint8          8097103872(0.37%)
head.anchor_encoder.pos_fc.2.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.80096      qint16         0(0%)
head.layers.41.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.800949     qint16         0(0%)
backbone.stages.0.block.3.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.800871     qint8          14394851328(0.66%)
backbone.stages.2.block.7.pwconv1.0                  weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.80086      qint8          8097103872(0.37%)
head.layers.17.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800857     qint16         0(0%)
head.layers.10.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800777     qint16         0(0%)
head.layers.20.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.800752     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800642     qint16         0(0%)
head.layers.22.attn._generated_mul_0                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800453     qint8          0(0%)
backbone.stages.2.block.6.pwconv1.0                  activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.800452     qint8          8097103872(0.37%)
head.anchor_encoder.yaw_fc.5.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.800432     qint16         0(0%)
head.layers.17.point_sum                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800355     qint16         0(0%)
head.layers.18.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800347     qint16         0(0%)
head.layers.20.layers.9.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.80029      qint16         0(0%)
backbone.stages.3.block.5.add                        activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.800218     qint8          4048551936(0.18%)
head.layers.31.camera_encoder.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800157     qint16         0(0%)
head.layers.7.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.800139     qint8          0(0%)
head.layers.3.camera_encoder.2.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.800095     qint16         0(0%)
head.anchor_encoder.size_fc.5.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.800049     qint16         0(0%)
head.layers.38.camera_encoder.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799972     qint16         0(0%)
head.anchor_encoder.pos_fc.8.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.799944     qint16         0(0%)
head.layers.10.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799927     qint16         0(0%)
head.layers.28.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799904     qint8          0(0%)
head.layers.35.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.799898     qint8          1744830464(0.08%)
head.layers.12.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799865     qint8          0(0%)
head.layers.38.camera_encoder.0                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.799805     qint8          479232(0.00%)
head.layers.34.layers.9.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.799755     qint16         0(0%)
head.layers.3.camera_encoder.5.weight_mul            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799746     qint16         0(0%)
head.layers.36.attn.softmax.mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799693     qint8          0(0%)
head.layers.32.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799684     qint16         0(0%)
head.layers.39.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.799484     qint16         0(0%)
head.anchor_encoder.vel_fc.8.weight_quant            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.799475     qint16         0(0%)
head.layers.22.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.799461     qint8          3489660928(0.16%)
head.anchor_encoder.pos_fc.8.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799405     qint16         0(0%)
head.layers.22.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.799388     qint8          3489660928(0.16%)
head.layers.3.camera_encoder.2.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799359     qint16         0(0%)
head.layers.3.camera_encoder.5.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799348     qint16         0(0%)
head.layers.6.layers.9.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799337     qint16         0(0%)
head.layers.41.quality_layers.5.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799316     qint16         0(0%)
head.layers.20.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.799275     qint16         0(0%)
head.layers.3.kps_generator.offset                   activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.799189     qint16         81788928(0.00%)
head.layers.38.weight_softmax.exp                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.799146     qint16         0(0%)
head.layers.18.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.799104     qint8          1744830464(0.08%)
head.layers.28.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.799093     qint8          1744830464(0.08%)
head.layers.14.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.799069     qint8          3489660928(0.16%)
head.layers.10.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.798995     qint16         0(0%)
head.layers.31.camera_encoder.2.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.798929     qint16         0(0%)
head.anchor_encoder.vel_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.798837     qint8          0(0%)
head.layers.32.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.798386     qint8          6979321856(0.32%)
head.layers.36.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.798261     qint16         0(0%)
head.anchor_encoder.pos_fc.5.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.798256     qint16         0(0%)
head.layers.11.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.798154     qint16         0(0%)
head.layers.34.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.798094     qint8          0(0%)
head.layers.35.attn.softmax.sub                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.797983     qint16         0(0%)
head.fc_before                                       weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.797973     qint8          15703474176(0.72%)
head.layers.10.camera_encoder.2.bias_add             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.797862     qint8          0(0%)
head.layers.31.kps_generator.offset                  activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.797717     qint16         81788928(0.00%)
backbone.stages.2.block.6.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.797691     qint8          0(0%)
backbone.stages.3.block.4.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.797568     qint8          12145655808(0.55%)
head.layers.12.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.797567     qint16         0(0%)
head.layers.31.camera_encoder.2.mul                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.797537     qint16         0(0%)
head.layers.36.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.797465     qint8          3489660928(0.16%)
head.layers.0.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.797246     qint8          1744830464(0.08%)
backbone.stages.0.block.0.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.797194     qint8          1012137984(0.05%)
backbone.stages.0.block.2.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.797164     qint8          0(0%)
head.layers.27.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.797117     qint16         0(0%)
head.layers.12.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.797077     qint16         0(0%)
head.layers.10.weight_softmax.reciprocal.reciprocal  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.797073     qint16         0(0%)
head.layers.41.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.797029     qint16         0(0%)
head.layers.38.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796994     qint16         0(0%)
head.layers.36.attn.out_proj                         activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.7969       qint8          3489660928(0.16%)
head.layers.16.input_mean.pre_mean                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796896     qint16         0(0%)
backbone.stages.3.block.4.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.796826     qint8          0(0%)
head.layers.39.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796719     qint16         0(0%)
head.layers.36.attn.matmul                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796621     qint8          3489660928(0.16%)
head.layers.31.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796561     qint8          0(0%)
head.layers.9.bias_quant                             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.796558     qint16         0(0%)
head.layers.17.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.79655      qint16         0(0%)
head.layers.31.camera_encoder.2.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796501     qint16         0(0%)
backbone.stages.3.block.0.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.796431     qint8          94887936(0.00%)
head.anchor_encoder.size_fc.8.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.79637      qint16         0(0%)
head.layers.32.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.796347     qint16         0(0%)
head.layers.10.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.796339     qint16         0(0%)
head.layers.41.quality_layers.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796328     qint16         0(0%)
head.layers.27.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.796255     qint8          37486592(0.00%)
head.anchor_encoder.yaw_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796241     qint16         0(0%)
head.layers.18.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.796176     qint8          3489660928(0.16%)
head.layers.17.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796081     qint8          0(0%)
head.layers.7.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.796001     qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.796        qint16         0(0%)
head.layers.11.pre_norm.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.795953     qint16         0(0%)
head.anchor_encoder.pos_fc.8.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795946     qint16         0(0%)
backbone.stages.3.block.5.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.795914     qint8          94887936(0.00%)
head.layers.28.attn.q_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.795867     qint8          3489660928(0.16%)
head.layers.34.layers.4.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.795844     qint16         0(0%)
head.layers.14.add                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795763     qint16         0(0%)
head.layers.15.key_cat                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795751     qint16         0(0%)
head.layers.4.pre_norm.sub                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795658     qint16         0(0%)
head.layers.9.sub                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795608     qint16         0(0%)
head.layers.10.camera_encoder.2.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.795537     qint16         0(0%)
head.layers.25.pre_norm.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795417     qint16         0(0%)
head.layers.38.camera_encoder.2.out_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795416     qint16         0(0%)
head.layers.31.weights_fc                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.795401     qint8          1308622848(0.06%)
backbone.stages.0.block.2.pwconv1                    activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.795285     qint8          14394851328(0.66%)
head.layers.41.layers.4.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795278     qint8          0(0%)
head.layers.38.point_quant_stub                      activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.795269     qint16         0(0%)
head.layers.30.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.79526      qint16         0(0%)
head.layers.27.cls_layers.6                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.795166     qint8          0(0%)
head.layers.9.var_mean.pre_mean                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.795131     qint16         0(0%)
head.layers.0.attn._generated_mul_0                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.794984     qint8          0(0%)
head.layers.41.quality_layers.5.weight_mul           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.794949     qint16         0(0%)
head.layers.35.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.79492      qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.8.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.794827     qint16         0(0%)
head.layers.10.weight_add                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.794636     qint8          0(0%)
head.layers.24.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.794538     qint16         0(0%)
head.layers.1.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.794537     qint8          3489660928(0.16%)
head.layers.13.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.7945       qint16         0(0%)
head.layers.27.layers.0                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.794373     qint8          872415232(0.04%)
head.layers.1.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.794301     qint8          3489660928(0.16%)
head.layers.41.quality_layers.5.var_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.7941       qint16         0(0%)
head.layers.21.attn.out_proj                         weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.793992     qint8          3489660928(0.16%)
backbone.stages.1.block.0.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.79396      qint8          7197425664(0.33%)
head.layers.18.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.793893     qint16         0(0%)
head.layers.15.attn.softmax.exp                      activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.793766     qint16         0(0%)
head.layers.31.point_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.793674     qint16         0(0%)
head.layers.38.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.793654     qint16         0(0%)
head.layers.24.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.793604     qint16         0(0%)
head.anchor_encoder.size_fc.5.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.793515     qint16         0(0%)
head.anchor_encoder.vel_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.793489     qint16         0(0%)
head.layers.28.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.793453     qint8          3489660928(0.16%)
head.layers.20.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.793399     qint16         37486592(0.00%)
head.layers.13.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.793287     qint16         37486592(0.00%)
head.layers.17.output_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.793159     qint8          872415232(0.04%)
head.layers.10.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.792921     qint8          872415232(0.04%)
head.layers.3.camera_encoder.2.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.792848     qint16         0(0%)
head.layers.22.attn.attn_weights_mean                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.792652     qint8          0(0%)
head.layers.24.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.792505     qint8          872415232(0.04%)
head.layers.13.layers.9.rsqrt                        activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.792499     qint16         0(0%)
backbone.stages.1.block.1.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.792263     qint8          506068992(0.02%)
head.layers.41.layers.9.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.792129     qint16         0(0%)
head.layers.27.layers.7                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.791974     qint8          872415232(0.04%)
backbone.stages.2.block.0.dwconv.0.0                 weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.791954     qint8          105431040(0.00%)
head.layers.17.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.791904     qint8          0(0%)
head.layers.31.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.791669     qint8          0(0%)
head.layers.31.camera_encoder.3                      weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.791577     qint8          10223616(0.00%)
head.anchor_encoder.vel_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.791442     qint16         0(0%)
head.layers.7.query_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.79144      qint16         0(0%)
backbone.stages.3.block.3.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.791399     qint8          94887936(0.00%)
head.anchor_encoder.size_fc.5.var_mean.pre_mean      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.791353     qint16         0(0%)
head.layers.19.mul                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.791201     qint16         0(0%)
head.layers.13.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.791189     qint8          872415232(0.04%)
head.layers.7.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.790809     qint16         0(0%)
head.layers.38.weight_softmax.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.790744     qint8          0(0%)
head.layers.20.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.790637     qint16         0(0%)
head.anchor_encoder.size_fc.11.mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.790582     qint16         0(0%)
head.layers.14.attn.v_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.790425     qint8          1744830464(0.08%)
head.layers.25.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.790304     qint16         0(0%)
head.layers.20.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.79009      qint16         0(0%)
backbone.stages.3.block.2.dwconv.0                   weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.78995      qint8          94887936(0.00%)
backbone.stages.2.block.7.pwconv1.1.lut              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.789868     qint8          0(0%)
head.layers.39.layers.0.0                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.789717     qint8          6979321856(0.32%)
head.anchor_encoder.vel_fc.11.weight_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.78967      qint16         0(0%)
head.layers.27.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.789559     qint16         0(0%)
head.layers.25.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.789534     qint16         0(0%)
head.layers.3.camera_encoder.2.out_mul               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.789486     qint16         0(0%)
head.layers.19.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.789322     qint16         0(0%)
backbone.stages.3.block.5.add                        weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.ConvAdd2d'>                         0.788973     qint8          4048551936(0.18%)
head.layers.38.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.788949     qint8          0(0%)
head.layers.29.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.78881      qint8          3489660928(0.16%)
head.layers.20.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.788728     qint16         0(0%)
head.layers.31.camera_encoder.5.weight_quant         activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.788727     qint16         0(0%)
head.anchor_encoder.size_fc.2.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.788685     qint16         0(0%)
head.anchor_encoder.size_fc.5.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.788637     qint16         0(0%)
head.layers.34.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.788618     qint16         0(0%)
backbone.stages.3.block.5.pwconv1                    weight            <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.788416     qint8          12145655808(0.55%)
head.layers.4.layers.0.0                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.78838      qint8          6979321856(0.32%)
head.layers.21.attn.softmax.reciprocal.reciprocal    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.788104     qint16         0(0%)
head.layers.13.layers.4.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.788014     qint16         0(0%)
head.layers.20.layers.4.mul                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.787953     qint16         0(0%)
head.layers.41.layers.4.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.786955     qint16         0(0%)
head.layers.4.pre_norm.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.786547     qint16         0(0%)
backbone.stages.3.block.5.act.lut                    activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.786381     qint8          0(0%)
head.layers.20.layers.2                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.786333     qint8          872415232(0.04%)
head.layers.41.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.786285     qint8          0(0%)
head.layers.13.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.78624      qint16         0(0%)
head.layers.41.layers.10                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.785874     qint8          37486592(0.00%)
head.layers.7.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.785808     qint8          1744830464(0.08%)
head.layers.41.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.78528      qint16         0(0%)
head.layers.13.layers.4.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.785207     qint16         0(0%)
head.layers.34.layers.4.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.785198     qint16         0(0%)
head.layers.39.layers.1                              activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.784054     qint8          3489660928(0.16%)
head.layers.22.attn.v_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.784011     qint8          3489660928(0.16%)
head.layers.32.pre_norm.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.783919     qint16         0(0%)
head.layers.23.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.783707     qint16         0(0%)
head.layers.13.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.783416     qint8          872415232(0.04%)
head.layers.22.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.783339     qint8          3489660928(0.16%)
head.layers.34.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.783309     qint8          872415232(0.04%)
head.layers.32.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.783301     qint16         0(0%)
head.layers.30.sub                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.7831       qint16         0(0%)
head.anchor_encoder.yaw_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.782993     qint16         0(0%)
head.layers.22.attn.attn_matmul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.782742     qint8          3489660928(0.16%)
head.layers.3.cam_add                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.782711     qint8          0(0%)
head.anchor_encoder.yaw_fc.11.rsqrt                  activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.78259      qint16         0(0%)
head.layers.11.pre_norm.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.782389     qint16         0(0%)
head.layers.34.layers.4.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.782177     qint16         0(0%)
head.layers.33.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.78212      qint16         0(0%)
head.layers.18.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.781735     qint16         0(0%)
head.layers.19.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.78164      qint16         0(0%)
head.layers.26.bias_quant                            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.780999     qint16         0(0%)
head.layers.12.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.78053      qint16         0(0%)
head.layers.11.pre_norm.bias_quant                   activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.780023     qint16         0(0%)
head.layers.32.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.779555     qint8          0(0%)
head.layers.18.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.779164     qint16         0(0%)
head.layers.10.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.778339     qint8          0(0%)
head.anchor_encoder.size_fc.5.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.778329     qint16         0(0%)
head.anchor_encoder.vel_fc.11.input_mean.pre_mean    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.778091     qint16         0(0%)
head.layers.11.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.7778       qint16         0(0%)
head.layers.16.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.777778     qint16         0(0%)
head.layers.34.add1                                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.777636     qint8          0(0%)
head.layers.13.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.777524     qint8          0(0%)
head.layers.6.add1                                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.777244     qint8          0(0%)
head.layers.4.pre_norm.input_mean.pre_mean           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.776902     qint16         0(0%)
head.layers.25.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.776747     qint8          1744830464(0.08%)
head.layers.40.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.776426     qint16         0(0%)
head.layers.1.attn.attn_matmul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.776185     qint8          3489660928(0.16%)
backbone.stages.3.block.2.dwconv.0                   activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.776025     qint8          94887936(0.00%)
head.layers.32.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.775731     qint16         0(0%)
head.layers.33.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.775693     qint8          0(0%)
head.layers.29.attn.q_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.775418     qint8          3489660928(0.16%)
head.layers.27.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.775417     qint8          872415232(0.04%)
head.layers.1.attn.softmax.exp                       activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.775237     qint16         0(0%)
head.layers.1.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.7745       qint16         0(0%)
head.layers.6.layers.4.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.774342     qint16         0(0%)
head.anchor_encoder.vel_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.773638     qint16         0(0%)
head.layers.24.cam_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.773136     qint8          0(0%)
head.anchor_encoder.yaw_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.772858     qint8          0(0%)
head.anchor_encoder.yaw_fc.8.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.772475     qint16         0(0%)
head.layers.27.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.772347     qint16         0(0%)
head.anchor_encoder.size_fc.11.weight_quant          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.77221      qint16         0(0%)
head.layers.6.layers.4.var_mean.pre_mean             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.771764     qint16         0(0%)
head.layers.10.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.771676     qint16         0(0%)
head.layers.13.layers.4.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.771509     qint16         0(0%)
head.layers.10.camera_encoder.5.rsqrt                activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.771495     qint16         0(0%)
head.layers.11.pre_norm.var_mean.pre_mean            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.771415     qint16         0(0%)
head.anchor_encoder.pos_fc.11.mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.770963     qint16         0(0%)
head.layers.8.add                                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.770389     qint16         0(0%)
head.anchor_encoder.pos_fc.8.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.770373     qint8          0(0%)
head.layers.21.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.770174     qint8          1744830464(0.08%)
head.layers.40.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.769493     qint8          0(0%)
head.layers.1.attn.out_proj                          weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.768828     qint8          3489660928(0.16%)
head.layers.8.attn.v_proj                            activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.767704     qint8          3489660928(0.16%)
head.layers.1.attn.softmax.mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.767666     qint8          0(0%)
head.layers.2.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.767097     qint16         0(0%)
head.layers.1.attn.v_proj                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.766877     qint8          3489660928(0.16%)
head.layers.20.layers.11.scale_quant_stub            activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.766724     qint16         0(0%)
head.layers.25.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.766674     qint16         0(0%)
head.anchor_encoder.yaw_fc.11.weight_mul             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.765492     qint16         0(0%)
backbone.stages.2.block.0.dwconv.0.0                 activation        <class 'horizon_plugin_pytorch.nn.qat.conv2d.Conv2d'>                            0.764075     qint8          105431040(0.00%)
head.layers.29.attn.k_proj                           weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.763907     qint8          3489660928(0.16%)
head.anchor_encoder.yaw_fc.8.sub                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.762431     qint16         0(0%)
head.layers.29.attn.k_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.76178      qint8          3489660928(0.16%)
head.anchor_encoder.size_fc.11.rsqrt                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.761493     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.mul                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.761301     qint16         0(0%)
head.layers.0.attn.attn_weights_mean                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.760766     qint8          0(0%)
head.layers.1.attn.softmax.reciprocal.reciprocal     activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.760353     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.bias_quant              activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.759787     qint16         0(0%)
head.layers.2.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.759735     qint8          0(0%)
head.layers.23.out_mul                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.758856     qint16         0(0%)
head.layers.20.layers.9.input_mean.pre_mean          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.757914     qint16         0(0%)
head.anchor_encoder.size_fc.11.sub                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.75728      qint16         0(0%)
head.layers.20.layers.9.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.755482     qint8          0(0%)
head.layers.18.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.754527     qint16         0(0%)
head.layers.18.pre_norm.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.753333     qint16         0(0%)
head.layers.25.short_add                             activation        <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.751906     qint8          1744830464(0.08%)
head.layers.6.layers.4.weight_mul                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.75151      qint16         0(0%)
head.layers.20.layers.5                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.750336     qint8          872415232(0.04%)
head.anchor_encoder.pos_fc.5.input_mean.pre_mean     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.750097     qint16         0(0%)
head.layers.20.layers.10                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.749376     qint8          37486592(0.00%)
head.layers.7.attn.out_proj                          activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.747834     qint8          3489660928(0.16%)
head.layers.18.short_add                             weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearAdd'>                         0.744574     qint8          1744830464(0.08%)
head.anchor_encoder.vel_fc.11.bias_add               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.744413     qint8          0(0%)
head.layers.6.layers.4.out_mul                       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.743457     qint16         0(0%)
head.anchor_encoder.pos_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.74314      qint16         0(0%)
head.layers.31.camera_encoder.5.sub                  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.741996     qint16         0(0%)
head.layers.40.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.741755     qint16         0(0%)
head.layers.20.layers.9.weight_mul                   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.74078      qint16         0(0%)
head.layers.20.layers.2                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.739789     qint8          872415232(0.04%)
head.layers.20.layers.9.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.736918     qint16         0(0%)
head.anchor_encoder.yaw_fc.5.weight_mul              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.736661     qint16         0(0%)
head.layers.6.layers.9.bias_quant                    activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.735253     qint16         0(0%)
head.layers.38.reciprocal_op.reciprocal              activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.730416     qint16         0(0%)
head.layers.17.camera_encoder.2.input_mean.pre_mean  activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.730056     qint16         0(0%)
head.layers.11.layers.0.0                            weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.729394     qint8          6979321856(0.32%)
head.layers.11.pre_norm.weight_quant                 activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.728067     qint16         0(0%)
head.layers.24.kps_generator.keypoints_add           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.728049     qint16         0(0%)
head.layers.23.weight_mul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.727553     qint16         0(0%)
head.layers.11.pre_norm.bias_add                     activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.727193     qint8          0(0%)
head.layers.20.layers.9.out_mul                      activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.727171     qint16         0(0%)
head.anchor_encoder.yaw_fc.8.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.725492     qint16         0(0%)
head.anchor_encoder.yaw_fc.2.var_mean.pre_mean       activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.725134     qint16         0(0%)
head.layers.20.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.724524     qint8          872415232(0.04%)
head.anchor_encoder.yaw_fc.5.out_mul                 activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.723924     qint16         0(0%)
head.layers.24.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.723425     qint8          0(0%)
head.anchor_encoder.yaw_fc.11.bias_quant             activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.722456     qint16         0(0%)
head.layers.11.layers.1                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.721604     qint8          3489660928(0.16%)
head.layers.27.layers.0                              weight            <class 'horizon_plugin_pytorch.nn.qat.linear.LinearReLU'>                        0.718819     qint8          872415232(0.04%)
head.layers.19.weight_quant                          activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.718252     qint16         0(0%)
head.anchor_encoder.vel_fc.11.sub                    activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.71756      qint16         0(0%)
head.layers.26.rsqrt                                 activation        <class 'horizon_plugin_pytorch.nn.qat.segment_lut.SegmentLUT'>                   0.714341     qint16         0(0%)
head.layers.25.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.713346     qint16         0(0%)
head.layers.23.bias_add                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.70933      qint8          0(0%)
head.layers.9.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.702153     qint16         0(0%)
head.layers.17.output_proj                           activation        <class 'horizon_plugin_pytorch.nn.qat.linear.Linear'>                            0.698987     qint8          872415232(0.04%)
head.layers.11.pre_norm.sub                          activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.698504     qint16         0(0%)
head.layers.0.attn.matmul                            activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.697211     qint8          1744830464(0.08%)
head.layers.2.out_mul                                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.697015     qint16         0(0%)
head.anchor_encoder.yaw_fc.5.bias_add                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.695162     qint8          0(0%)
head.layers.10.residual_op                           activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.686369     qint8          0(0%)
head.anchor_encoder.vel_fc.11.out_mul                activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.685789     qint16         0(0%)
head.layers.31.camera_encoder.5.bias_quant           activation        <class 'horizon_plugin_pytorch.nn.qat.stubs.QuantStub'>                          0.672117     qint16         0(0%)
head.layers.9.bias_add                               activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.669303     qint8          0(0%)
head.anchor_encoder.size_fc.11.input_mean.pre_mean   activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.665712     qint16         0(0%)
head.layers.9.weight_mul                             activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.664432     qint16         0(0%)
head.layers.31.feat_cat                              activation        <class 'horizon_plugin_pytorch.nn.qat.functional_modules.FloatFunctional'>       0.460732     qint8          0(0%)