🤖 AI Summary
Visual representations are often corrupted by noise and task-irrelevant features in latent space, degrading model robustness and generalization. To address this, we propose an unsupervised dynamic feature selection method that operates during forward propagation: it performs instance-wise importance scoring and applies a lightweight gating mechanism to selectively retain discriminative features—entirely without supervision or ground-truth labels. This is the first work to introduce unsupervised dynamic feature selection for latent-space representation optimization, leading to substantially improved representation quality. Extensive experiments on image clustering and generation tasks demonstrate consistent performance gains across multiple benchmark datasets, along with enhanced generalization capability. Crucially, the method incurs only negligible computational overhead, preserving efficiency while significantly boosting representation fidelity.
📝 Abstract
Latent representations are critical for the performance and robustness of machine learning models, as they encode the essential features of data in a compact and informative manner. However, in vision tasks, these representations are often affected by noisy or irrelevant features, which can degrade the model's performance and generalization capabilities. This paper presents a novel approach for enhancing latent representations using unsupervised Dynamic Feature Selection (DFS). For each instance, the proposed method identifies and removes misleading or redundant information in images, ensuring that only the most relevant features contribute to the latent space. By leveraging an unsupervised framework, our approach avoids reliance on labeled data, making it broadly applicable across various domains and datasets. Experiments conducted on image datasets demonstrate that models equipped with unsupervised DFS achieve significant improvements in generalization performance across various tasks, including clustering and image generation, while incurring a minimal increase in the computational cost.