Skip to main content
. 2024 May 22;14:11735. doi: 10.1038/s41598-024-61798-6

Figure 1.

Figure 1

The deep learning model processes a 3D T1-weighted image via a single-input, dual-output 3D convolutional neural network (CNN) to produce estimated multi-label masks for brain tissues (background, white matter, gray matter, cerebrospinal fluid) and brain structures (background + 22 brain structures). The CNN is based on the widely used 3D U-net architecture, which operates on 3D patches of the input scan. Each convolutional layer utilizes 3×3×3 kernels, except for the two convolutional layers before the softmax layers, which use 1×1×1 kernels. Weight normalization and leaky ReLU (slope = 0.20) are employed. The output patches have dimensions of 88×88×88 voxels, which are smaller than the input patches’ dimensions (128×128×128 voxels) due to the use of valid convolutions, mitigating off-patch-center bias.