π€ AI Summary
Capsule Networks (CapsNets) outperform CNNs in overlapping-digit classification and affine-transformed image recognition but suffer from excessive parameter count, high computational complexity, and poor scalability to multi-class tasks. To address these limitations, we propose Deep Lightweight Capsule Network (DL-CapsNet): it employs a deep stacked capsule architecture to enhance representational capacity; introduces a novel Capsule Aggregation Layer to replace redundant fully connected capsule layers; and integrates dynamic routing optimization with parameter compression techniques. DL-CapsNet reduces model parameters by approximately 62% compared to the baseline CapsNet, significantly accelerating both training and inference. Crucially, it maintains or even surpasses the original CapsNetβs classification accuracy on challenging benchmarks including Multi-MNIST and SVHN. To our knowledge, DL-CapsNet is the first capsule-based architecture to jointly achieve high efficiency and competitive accuracy in high-cardinality classification scenarios.
π Abstract
Capsule Network (CapsNet) is among the promising classifiers and a possible successor of the classifiers built based on Convolutional Neural Network (CNN). CapsNet is more accurate than CNNs in detecting images with overlapping categories and those with applied affine transformations. In this work, we propose a deep variant of CapsNet consisting of several capsule layers. In addition, we design the Capsule Summarization layer to reduce the complexity by reducing the number of parameters. DL-CapsNet, while being highly accurate, employs a small number of parameters and delivers faster training and inference. DL-CapsNet can process complex datasets with a high number of categories.