Sharper Error Bounds in Late Fusion Multi-view Clustering Using Eigenvalue Proportion

📅 2024-12-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multi-view clustering suffers from severe noise and redundant partitions, alongside inadequate modeling of high-order cross-view correlations. Method: This paper proposes a theoretical error-bound framework based on the principal eigenvalue ratio and designs a low-pass graph filtering strategy to enhance the robustness and discriminability of late fusion. It integrates local Rademacher complexity analysis, multilinear k-means, multiple kernel learning, and spectral graph theory. Contribution/Results: We establish, for the first time, an $O(1/n)$ convergence rate for the generalization error of multi-kernel k-means—improving upon the prior $O(sqrt{k/n})$ bound—and introduce the principal eigenvalue ratio as an optimizable metric for fusion quality. Evaluated on standard benchmarks, our method achieves an average 3.2% improvement in clustering accuracy and demonstrates significantly enhanced robustness against both feature-level noise and view-level missingness.

Technology Category

Application Category

📝 Abstract
Multi-view clustering (MVC) aims to integrate complementary information from multiple views to enhance clustering performance. Late Fusion Multi-View Clustering (LFMVC) has shown promise by synthesizing diverse clustering results into a unified consensus. However, current LFMVC methods struggle with noisy and redundant partitions and often fail to capture high-order correlations across views. To address these limitations, we present a novel theoretical framework for analyzing the generalization error bounds of multiple kernel $k$-means, leveraging local Rademacher complexity and principal eigenvalue proportions. Our analysis establishes a convergence rate of $mathcal{O}(1/n)$, significantly improving upon the existing rate in the order of $mathcal{O}(sqrt{k/n})$. Building on this insight, we propose a low-pass graph filtering strategy within a multiple linear $k$-means framework to mitigate noise and redundancy, further refining the principal eigenvalue proportion and enhancing clustering accuracy. Experimental results on benchmark datasets confirm that our approach outperforms state-of-the-art methods in clustering performance and robustness. The related codes is available at https://github.com/csliangdu/GMLKM .
Problem

Research questions and friction points this paper is trying to address.

Multi-view Clustering
Redundancy Handling
Deep Correlation Capturing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-view Clustering
K-means Optimization
Noise Reduction
🔎 Similar Papers
No similar papers found.
Liang Du
Liang Du
Associate Professor, Villanova University
electric power systems
H
Henghui Jiang
School of Computer and Information Technology, Shanxi University, Taiyuan, 030006, China
X
Xiaodong Li
School of Computer and Information Technology, Shanxi University, Taiyuan, 030006, China
Y
Yiqing Guo
School of Computer and Information Technology, Shanxi University, Taiyuan, 030006, China
Y
Yan Chen
College of Computer Science and Technology, Taiyuan University of Technology, Taiyuan, 030600, China
F
Feijiang Li
School of Computer and Information Technology, Shanxi University, Taiyuan, 030006, China
P
Peng Zhou
School of Computer Science and Technology, Anhui University, Hefei, 230601, China
Yuhua Qian
Yuhua Qian
山西大学大数据科学与产业研究院
机器学习、数据挖掘、复杂网络