Skip to main content
. 2022 Dec 12;12(12):1854. doi: 10.3390/biom12121854
Algorithm 1 Attention Routing using Scalar Product
 Input: input parameters capsule unl from layer l
 Output: output the Digicaps: Vj
  1: affine transformation for all unl:
U^(nl,nl+1,dl+1)l=unTl×W(nl,nl+1,dl,dl+1)l
  2: Calculating self-attention weights:
A(nl,nl,nl+1)l=U^(nl,nl+1,dl+1)l×U^(nl,nl+1,dl+1)Tldl
  3: Use softmax to Calculate Weights C:
C(nl,nl+1)l=exp(nlA(nl,nl,nl+1)l)nl+1exp(nlA(nl,nl,nl+1)l)
  4: For all capsule j in l+1:
snl+1=U^(nl,nl+1,dl+1)l×(C(nl,nl+1)l+B(nl,nl+1)l)
  5: Compress the capsule length to between 0 and 1:
Vj(x1xm)squash(snl+1)
  6: return Vj