🤖 AI Summary
Capsule Networks (CapsNets) offer affine robustness and superior performance on overlapping objects, but their high computational cost and slow inference stem from the large number of Primary Capsules (PCs).
Method: This work presents the first systematic study of PC pruning across MNIST, Fashion-MNIST, CIFAR-10, and SVHN. We propose an importance-based selective pruning strategy that jointly optimizes structural sparsity and dynamic routing efficiency—requiring no fine-tuning post-pruning.
Contribution/Results: Our approach achieves up to 95.2% reduction in capsule count and 95.36% reduction in dynamic routing operations, yielding up to 9.90× inference speedup with zero accuracy degradation. Empirical analysis uncovers a dataset-complexity–PC-redundancy correlation, establishing a reproducible, cross-dataset pruning paradigm for lightweight CapsNets.
📝 Abstract
Capsule Networks (CapsNets) are a generation of image classifiers with proven advantages over Convolutional Neural Networks (CNNs). Better robustness to affine transformation and overlapping image detection are some of the benefits associated with CapsNets. However, CapsNets cannot be classified as resource-efficient deep learning architecture due to the high number of Primary Capsules (PCs). In addition, CapsNets' training and testing are slow and resource hungry. This paper investigates the possibility of Primary Capsules pruning in CapsNets on MNIST handwritten digits, Fashion-MNIST, CIFAR-10, and SVHN datasets. We show that a pruned version of CapsNet performs up to 9.90 times faster than the conventional architecture by removing 95 percent of Capsules without a loss of accuracy. Also, our pruned architecture saves on more than 95.36 percent of floating-point operations in the dynamic routing stage of the architecture. Moreover, we provide insight into why some datasets benefit significantly from pruning while others fall behind.