🤖 AI Summary
In open-world learning, existing out-of-distribution (OOD) detection methods lack mechanisms for knowledge updating, while incremental learning typically relies on supervised signals—introducing a critical mismatch with real-world open environments. To bridge this gap, we propose the first unified framework integrating OOD detection, novel-class discovery, and unsupervised incremental learning. Our approach leverages anomaly detection to identify unknown samples, employs contrastive learning coupled with clustering for unsupervised novel-class discovery, and introduces a gradient-constrained, prototype-aligned unsupervised fine-tuning strategy to ensure stable model evolution. Extensive experiments across multiple standard benchmarks demonstrate significant improvements in unknown-class recognition accuracy and long-term learning stability. Notably, our method achieves fully unsupervised open-world autonomous evolution—the first such result in the literature. Code and experimental artifacts are publicly released.
📝 Abstract
Substantial progress has been made in various techniques for open-world recognition. Out-of-distribution (OOD) detection methods can effectively distinguish between known and unknown classes in the data, while incremental learning enables continuous model knowledge updates. However, in open-world scenarios, these approaches still face limitations. Relying solely on OOD detection does not facilitate knowledge updates in the model, and incremental fine-tuning typically requires supervised conditions, which significantly deviate from open-world settings. To address these challenges, this paper proposes OpenHAIV, a novel framework that integrates OOD detection, new class discovery, and incremental continual fine-tuning into a unified pipeline. This framework allows models to autonomously acquire and update knowledge in open-world environments. The proposed framework is available at https://haiv-lab.github.io/openhaiv .