Interpreting Manifolds and Graph Neural Embeddings from Internet of Things Traffic Flows

📅 2026-02-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing network monitoring approaches in capturing dynamic relationships among IoT devices and the lack of interpretability in graph neural network (GNN) embeddings. To this end, we propose an interpretable analysis framework that integrates manifold learning with feature attribution. Our method is the first to map high-dimensional GNN embeddings onto a low-dimensional latent manifold, enabling both visualization of network state evolution and attribution of critical traffic features. The framework not only facilitates the identification of concept drift but also achieves an F1 score of 0.830 in intrusion detection tasks, effectively uncovering the structural dependencies and behavioral evolution mechanisms inherent in IoT traffic.

Technology Category

Application Category

📝 Abstract
The rapid expansion of Internet of Things (IoT) ecosystems has led to increasingly complex and heterogeneous network topologies. Traditional network monitoring and visualization tools rely on aggregated metrics or static representations, which fail to capture the evolving relationships and structural dependencies between devices. Although Graph Neural Networks (GNNs) offer a powerful way to learn from relational data, their internal representations often remain opaque and difficult to interpret for security-critical operations. Consequently, this work introduces an interpretable pipeline that generates directly visualizable low-dimensional representations by mapping high-dimensional embeddings onto a latent manifold. This projection enables the interpretable monitoring and interoperability of evolving network states, while integrated feature attribution techniques decode the specific characteristics shaping the manifold structure. The framework achieves a classification F1-score of 0.830 for intrusion detection while also highlighting phenomena such as concept drift. Ultimately, the presented approach bridges the gap between high-dimensional GNN embeddings and human-understandable network behavior, offering new insights for network administrators and security analysts.
Problem

Research questions and friction points this paper is trying to address.

Interpretability
Graph Neural Networks
IoT Traffic
Manifold Learning
Network Visualization
Innovation

Methods, ideas, or system contributions that make the work stand out.

interpretable GNN
manifold visualization
IoT traffic analysis
feature attribution
concept drift detection
🔎 Similar Papers
No similar papers found.
E
Enrique Feito-Casares
Universidad Rey Juan Carlos, Madrid, Spain
F
F. Melgarejo-Meseguer
Universidad Rey Juan Carlos, Madrid, Spain
Elena Casiraghi
Elena Casiraghi
Associate Professor, Computer Science Department (DI), Università degli Studi di Milano
Pattern recognition and ClassificationManifold Learning & Curve InferenceDigital Image\Signal
Giorgio Valentini
Giorgio Valentini
Computer Science Dept. - University of Milan
BioinformaticsComputational biologyMachine Learning
J
J. Rojo-álvarez
Universidad Rey Juan Carlos, Madrid, Spain; Centro de Investigación for Data, Complex Networks and Cybersecurity Sciences, Universidad Rey Juan Carlos, Madrid, Spain