1. Abstract
We aim to further advance the performance-efficiency boundary of YOLOs from both the post-processing and the model architecture.
- We first tackle the problem of redundant predictions in the post-processing by presenting a consistent dual assignments strategy for NMS-free YOLOs with the dual label assignments { one-to-many head and one-to-one head } and consistent matching metric.
- It allows the model to enjoy rich and harmonious supervision during training while eliminating the need for NMS during inference, leading to competitive performance with high efficiency.
- We introduce the holistic efficiency-accuracy driven model design strategy for YOLOs
- For efficiency, we propose the lightweight classification head, spatial channel decoupled downsampling, and rank-guided block design.
- For accuracy, we explore the large-kernel convolution and present the effective partial self-attention (PSA) module to enhance the model capability.
2. Our Method
1. Consistent dual assignments (one-to-many head and one-to-one head) for NMS-free
- In dual label assignments, the one-to-many branch provides much richer supervisory signals than
one-to-one branch. Intuitively, if we can harmonize the supervision of the one-to-one head with that
of one-to-many head, we can optimize the one-to-one head towards the direction of one-to-many
head’s optimization. As a result, the one-to-one head can provide improved quality of samples during
inference, leading to better performance.
2. Holistic Efficiency-Accuracy Driven Model Design