Improving the Robustness of Capsule Networks to Image Affine Transformations

11/18/2019
by   Jindong Gu, et al.
20

Convolutional neural networks (CNNs) achieve translational invariance using pooling operations, which do not maintain the spatial relationship in the learned representations. Hence, they cannot extrapolate their understanding of the geometric transformation of inputs. Recently, Capsule Networks (CapsNets) have been proposed to tackle this problem. In CapsNets, each entity is represented by a vector and routed to high-level entities by a dynamic routing algorithm. The CapsNets have been shown to be more robust than CNNs to affine transformations of inputs. However, there is still a huge gap between their performance on transformed inputs compared to untransformed versions. In this work, we first revisit the routing procedure by (un)rolling its forward and backward passes. Our investigation reveals that the routing procedure contributes neither to generalization ability nor to the affine robustness of the CapsNets. Furthermore, we explore the limitations of capsule transformations and propose affine CapsNets (Aff-CapsNets) that are more robust to affine transformations. On our benchmark task where models are trained on the MNIST dataset and tested on the AffNIST dataset, our Aff-CapsNets improve the benchmark performance by a large margin (from 79% to 93.21%), without using a routing mechanism. We also demonstrate the superiority of Aff-CapsNets on a real-world Brain Tumor Type classification task.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset