🤖 AI Summary
Graph Convolutional Networks (GCNs) suffer from insufficient utilization of label information and weak label influence on unlabeled nodes. Method: This paper proposes a two-stage framework: (1) a label-driven graph structure self-adaptation mechanism to construct an Enhanced Label-propagation Utility (ELU) graph, improving label diffusion capability; and (2) dual-graph contrastive learning between the original graph and the ELU graph, incorporating cross-graph message passing and consistency constraints to enhance representation robustness. Contribution/Results: Theoretical analysis demonstrates that the proposed method improves the generalization bound of GCNs. Extensive experiments on multiple benchmark datasets show significant performance gains over state-of-the-art GCN variants and graph contrastive learning methods, validating the synergistic improvement in label utilization efficiency and node representation capability.
📝 Abstract
The message-passing mechanism of graph convolutional networks (i.e., GCNs) enables label information to be propagated to a broader range of neighbors, thereby increasing the utilization of labels. However, the label information is not always effectively utilized in the traditional GCN framework. To address this issue, we propose a new two-step framework called ELU-GCN. In the first stage, ELU-GCN conducts graph learning to learn a new graph structure (ie ELU-graph), which enables GCNs to effectively utilize label information. In the second stage, we design a new graph contrastive learning on the GCN framework for representation learning by exploring the consistency and mutually exclusive information between the learned ELU graph and the original graph. Moreover, we theoretically demonstrate that the proposed method can ensure the generalization ability of GCNs. Extensive experiments validate the superiority of the proposed method.