🤖 AI Summary
Existing point cloud sequence-based 3D action recognition methods struggle to simultaneously preserve fine-grained limb motion fidelity and global pose structural integrity, leading to critical spatiotemporal cue loss. To address this, we propose D-Hyperpoint representation and KAN-HyperpointNet—a decoupled architecture. D-Hyperpoint is the first unified representation that jointly encodes local instantaneous motion and global static pose. We further pioneer the integration of Kolmogorov–Arnold Networks (KANs) into a spatiotemporal hybrid module to model high-order nonlinear interactions. Our framework comprises four key components: D-Hyperpoint embedding, KANsMixer, KAN-enhanced spatiotemporal decoupling, and a hierarchical nested aggregation mechanism. Extensive experiments demonstrate state-of-the-art performance on MSR Action3D and NTU-RGB+D 60, with significant improvements in fine-grained discrimination of complex actions.
📝 Abstract
Point cloud sequence-based 3D action recognition has achieved impressive performance and efficiency. However, existing point cloud sequence modeling methods cannot adequately balance the precision of limb micro-movements with the integrity of posture macro-structure, leading to the loss of crucial information cues in action inference. To overcome this limitation, we introduce D-Hyperpoint, a novel data type generated through a D-Hyperpoint Embedding module. D-Hyperpoint encapsulates both regional-momentary motion and global-static posture, effectively summarizing the unit human action at each moment. In addition, we present a D-Hyperpoint KANsMixer module, which is recursively applied to nested groupings of D-Hyperpoints to learn the action discrimination information and creatively integrates Kolmogorov-Arnold Networks (KAN) to enhance spatio-temporal interaction within D-Hyperpoints. Finally, we propose KAN-HyperpointNet, a spatio-temporal decoupled network architecture for 3D action recognition. Extensive experiments on two public datasets: MSR Action3D and NTU-RGB+D 60, demonstrate the state-of-the-art performance of our method.