Skip to main content
. 2024 Nov 13;14:27883. doi: 10.1038/s41598-024-77633-x

Fig. 2.

Fig. 2

Our proposed framework consists of three main parts: 1) TANQ-based image translation from ceT1 to hrT2 images, 2) Multi-view pseudo-hrT2 representation via CycleGAN, 3) Construction of a VS/cochlea segmentation model using multi-view pseudo-hrT2 images and self-training with real-hrT2 images. Specifically, TANQ divides the features based on the ceT1 labels in both the encoder and decoder, applying target-aware normalization. Furthermore, it includes an additional decoder called SegDecoder. The Encoder E extracts features from both the real ceT1 images and pseudo-hrT2 images and then calculates the contrastive loss between selected features using a sorted attention matrix.