🤖 AI Summary
Pedestrian attribute recognition (PAR) faces three key challenges: severe class imbalance in fine-grained attribute prediction, strong inter-attribute correlations, and poor cross-domain generalization. To address these, we propose a modular vision-language framework built upon a frozen multilingual SigLIP encoder. First, both the visual and textual encoders of SigLIP are frozen to preserve rich multilingual semantic priors. Second, a lightweight cross-modal cross-attention mechanism is introduced to precisely align image features with learnable prompt embeddings. Third, a prompt-driven visual feature refinement module is incorporated to enhance discriminability and mitigate distribution shift across domains. Evaluated on PA100K, PETA, and Market-1501, our method achieves state-of-the-art performance on PA100K and yields substantial average accuracy gains on PETA and Market-1501. It effectively alleviates class imbalance and domain discrepancy, demonstrating superior robustness and generalization capability.
📝 Abstract
Pedestrian Attribute Recognition (PAR) involves predicting fine-grained attributes such as clothing color, gender, and accessories from pedestrian imagery, yet is hindered by severe class imbalance, intricate attribute co-dependencies, and domain shifts. We introduce VLM-PAR, a modular vision-language framework built on frozen SigLIP 2 multilingual encoders. By first aligning image and prompt embeddings via refining visual features through a compact cross-attention fusion, VLM-PAR achieves significant accuracy improvement on the highly imbalanced PA100K benchmark, setting a new state-of-the-art performance, while also delivering significant gains in mean accuracy across PETA and Market-1501 benchmarks. These results underscore the efficacy of integrating large-scale vision-language pretraining with targeted cross-modal refinement to overcome imbalance and generalization challenges in PAR.