Skip to content

Quantization Aware Training using the LSQ method

Context

Now that Post Training Quantization is implemented, we want to explore Quantization Aware Training.

A straight forward approach for the QAT is to use the LSQ node, that quantizes both the weights and the activation.

Key steps

  • test the LSQ operator (cpu)
  • test the LSQ operator (cuda)
  • implement the quantizers insertion and init routines
  • benchmark the LSQ QAT on the MNIST (cpu)
  • benchmark the LSQ QAT on the CIFAR-10 (cuda)
  • benchmark the LSQ QAT on the ImageNet (cuda)

Results

TODO

Edited by Benjamin Halimi