Skip to yearly menu bar Skip to main content


Poster

STAR-Caps: Capsule Networks with Straight-Through Attentive Routing

Karim Ahmed · Lorenzo Torresani

East Exhibition Hall B + C #101

Keywords: [ Algorithms -> Classification; Applications -> Computer Vision; Deep Learning; Deep Learning ] [ Attention Models; Deep Learning ] [ Object Recognition ] [ Applications ]


Abstract:

Capsule networks have been shown to be powerful models for image classification, thanks to their ability to represent and capture viewpoint variations of an object. However, the high computational complexity of capsule networks that stems from the recurrent dynamic routing poses a major drawback making their use for large-scale image classification challenging. In this work, we propose Star-Caps a capsule-based network that exploits a straight-through attentive routing to address the drawbacks of capsule networks. By utilizing attention modules augmented by differentiable binary routers, the proposed mechanism estimates the routing coefficients between capsules without recurrence, as opposed to prior related work. Subsequently, the routers utilize straight-through estimators to make binary decisions to either connect or disconnect the route between capsules, allowing stable and faster performance. The experiments conducted on several image classification datasets, including MNIST, SmallNorb, CIFAR-10, CIFAR-100, and ImageNet show that STAR-Caps outperforms the baseline capsule networks.

Live content is unavailable. Log in and register to view live content