Lightweitght AI,
[Arxiv] CR-QAT: Curriculum Relational Quantization-Aware Training for Open-Vocabulary Object Detection
Extreme low-bit (e.g., 4-bit) quantization of open-vocabulary object detection (OVOD) models severely degrades fine-grained vision-language alignment and distorts inter-region relational structures. To resolve this, we propose Curriculum Relational Quantization-Aware Training (CR-QAT), which mitigates error accumulation through progressive model partitioning and preserves multi-dimensional structural knowledge via text-centric relational distillation. Experiments on LVIS and COCO zero-shot benchmarks demonstrate that CR-QAT consistently outperforms existing baselines under aggressive low-bit settings, achieving relative AP improvements of up to 40.9%.