Skip to main content
. Author manuscript; available in PMC: 2023 Sep 27.
Published in final edited form as: Annu Rev Biomed Data Sci. 2023 Apr 27;6:153–171. doi: 10.1146/annurev-biodatasci-020722-020704

Figure 5.

Figure 5

The neural network architectures for deep learning and deep transfer learning. (a) An example architecture of a deep neural network model, which includes an input layer; several hidden layers (marked as F), including fully connected layers and dropout layers; and one output layer C. More fully connected layers can be added to the deep neural network model. (b) The neural network architecture of a stacked denoising auto-encoder (SAE) for transfer learning. F1 (or F2) is the encoder with two layers, including a fully connected layer and a dropout layer; F1 (or F2) is the decoder; the first and the second rows provide the structure of the first and second auto-encoders, respectively; and C is a regression or classification layer. (c) The neural network architecture of classification and contrastive semantic alignment (CCSA) (105). CCSA minimizes the loss function LCCSA(f)=(1γ)LC(bg)+γ(LSA(g)+LS(g)), where f=bg represents the composition of a function g that maps the input data X to an embedding space Z and a function b used to predict the output label from Z; C is a classification layer; LC(bg) is the classification loss; LSA(g) is the semantic alignment loss; LS(g) is the separation loss; and γ is the weight used to balance the classification loss versus the contrastive semantic alignment loss LSA(g)+LS(g).