The floating-point models that meet quantized awareness training requirements.
The process of obtaining quantitative parameters using calibration data.
Pseudo-quantized model obtained after Calibration.
Training for quantized awareness.
Pseudo-quantized models obtained after quantized awareness training.
The process of first quantizing and then dequantizing floating-point data which is generally implemented in network models through pseudo-quantized nodes.
Models with pseudo-quantized nodes which are typically obtained by Calibration or QAT.
Convert the floating-point parameters in a pseudo-quantized model to fixed-point parameters through parameter transformations, and convert the floating-point operators to fixed-point operators, the transformed model is called a Quantized model or fixed-point model or quantized model.
Models exported for deployment, typically exported from a QAT model, can be used for accuracy simulation and compilation on boards.
Name of the BPU architecture.
Name of the processor.
| Processor | J6E | J6M |
|---|---|---|
| BPU | Nash-e | Nash-m |
| enum | March.NASH_E | March.NASH_M |