Enhancing Graph Neural Networks: A Mutual Learning Approach

📅 2025-10-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the performance limitations of shallow graph neural networks (GNNs) in multi-task learning—particularly in teacher-free settings where pre-trained teacher models are unavailable—this paper proposes a Teacher-Free Mutual Learning (TFML) framework. TFML enables collaborative training of multiple homogeneous shallow GNNs, incorporating an adaptive logit-weighting module to dynamically modulate inter-model knowledge exchange intensity and an entropy-regularized mechanism to enhance prediction confidence and generalization. Unlike conventional knowledge distillation, TFML eliminates reliance on a fixed teacher model and instead achieves bidirectional knowledge transfer through ensemble-based mutual supervision. Extensive experiments on three node classification and three graph classification benchmark datasets demonstrate that TFML significantly improves both accuracy and robustness of shallow GNNs. Notably, it achieves superior performance in joint multi-task learning scenarios, outperforming standard baselines and teacher-dependent distillation methods.

Technology Category

Application Category

📝 Abstract
Knowledge distillation (KD) techniques have emerged as a powerful tool for transferring expertise from complex teacher models to lightweight student models, particularly beneficial for deploying high-performance models in resource-constrained devices. This approach has been successfully applied to graph neural networks (GNNs), harnessing their expressive capabilities to generate node embeddings that capture structural and feature-related information. In this study, we depart from the conventional KD approach by exploring the potential of collaborative learning among GNNs. In the absence of a pre-trained teacher model, we show that relatively simple and shallow GNN architectures can synergetically learn efficient models capable of performing better during inference, particularly in tackling multiple tasks. We propose a collaborative learning framework where ensembles of student GNNs mutually teach each other throughout the training process. We introduce an adaptive logit weighting unit to facilitate efficient knowledge exchange among models and an entropy enhancement technique to improve mutual learning. These components dynamically empower the models to adapt their learning strategies during training, optimizing their performance for downstream tasks. Extensive experiments conducted on three datasets each for node and graph classification demonstrate the effectiveness of our approach.
Problem

Research questions and friction points this paper is trying to address.

Enhancing graph neural networks through mutual learning collaboration
Developing adaptive knowledge exchange without pre-trained teacher models
Improving multi-task inference performance via synergistic student ensembles
Innovation

Methods, ideas, or system contributions that make the work stand out.

Collaborative learning replaces teacher-student knowledge distillation
Mutual teaching among ensemble GNNs during training process
Adaptive logit weighting and entropy enhancement techniques
🔎 Similar Papers
No similar papers found.
Paul Agbaje
Paul Agbaje
University of Texas at Arlington
Internet of VehiclesMachine LearningGraphCyber securityCyber Physical Systems
A
Akajyoti Mitra
Department of Computer Science and Engineering, University of Texas at Arlington, TX
Afia Anjum
Afia Anjum
Los Alamos National Laboratory, University of Texas at Arlington
IoTCommunication ProtocolsSemantic CommunicationLarge Language ModelsSmart Grid
P
Pranali Khose
Department of Computer Science and Engineering, University of Texas at Arlington, TX
E
Ebelechukwu Nwafor
Department of Computing Sciences, Villanova University, PA
Habeeb Olufowobi
Habeeb Olufowobi
Assistant Professor, University of Texas at Arlington
Cyberphysical SystemsSecurityAI/ML