When I quantized the model (RTMDet-Ins-m) downloaded from mmdetection with VitisAI 3.5, the computation time was very slow because a transpose layer (computed by the CPU) was generated for all CNN layers.
I think that the transpose layer was generated because the XIR model is NHWC, while the Pytorch model is in NCHW format, but is this inference correct?
Also, I want to convert from NCHW format to NHWC format so that the transpose layer is not generated, how can I do that? Please let me know if there is a valid conversion program.