To produce this sort of an FPGA implementation, we compress the 2D CNN by applying dynamic quantization tactics. As opposed to fine-tuning an already educated network, this move involves retraining the CNN from scratch with constrained bitwidths for that weights and activations. This method known as quantization-informed training (QAT). The https://devendrab980tdn6.wikievia.com/user