CPC G06F 18/21 (2023.01) [G06F 18/2431 (2023.01); G06N 3/08 (2013.01); G06V 10/40 (2022.01)] | 7 Claims |
1. A processor implemented method, comprising:
employing a deep feature based Capsule Network for a task, via one or more hardware processors, the employed deep feature based Capsule Network is a deep neural network including one or more capsules, each capsule of the one or more capsules is a set of neurons of the deep feature based Capsule Network, the deep feature based Capsule Network comprising:
a feature extraction layer comprising a set of feature extraction units, wherein each feature extraction unit of the set of feature extraction units comprises:
alternate convolutional layers with 512 filters having kernal size 3×3 and 256 filters with kernal size 1×1 respectively, and
a batch normalization (BN) layer, and
a capsule layer that includes:
a primary capsule layer comprising a set of primary Capsules for defining representation of initial deep features in form of vectors that are trained to learn geometric transformations in a routing-by-agreement process, wherein the primary capsule layer comprises a plurality of neurons of the deep feature based Capsule Network, and
a class capsule layer comprising a set of class capsules, wherein the routing-by-agreement process comprises passing information from each of the set of primary capsules to each capsule in the set of class capsules only for features that contributed to prediction of at least one class in the past, and refrains from passing the information from each of the set of primary capsules to each capsule in the set of class capsules for features that did not contribute to prediction of at least one class in the past,
wherein the feature extraction layer is prior to the capsule layer, the feature extraction layer configured to extract the deep features from input data,
wherein employing the deep feature based Capsule Network comprises:
passing, successively through each of the set of feature extraction units, an input data to obtain a set of deep features;
performing, by the primary capsule layer, a convolution of the set of deep features into the set of primary capsules, wherein the set of primary capsules in the primary capsule layer comprises an optimal number of capsules based on a number of features in the set of deep features; and
predicting, at the class capsule layer, a set of classes from the set of deep features, wherein each capsule in the set of class capsules predict a single class and is activated only for those primary block capsules which together agrees for a required class during a training phase of the deep feature based Capsule Network, wherein a number of capsules of the one or more capsules in the set of class capsules is equal to a number of classes associated with the task, and further receiving the set of classes predicted by the class capsule layer via a decoder and reconstructing an image associated with the task based on the set of classes during a training phase of the deep feature based Capsule Network.
|