Skip to main content
. 2022 Oct 20;10(10):e41136. doi: 10.2196/41136

Figure 2.

Figure 2

Scaled dot-product attention function (left). Multi-head attention consists of several scaled dot-product attention (right). Concat: concatenate; K: key; Matmul: matrix multiply; Q: query; V: value.