Skip to main content
BioMed Research International logoLink to BioMed Research International
. 2020 Nov 5;2020:5193707. doi: 10.1155/2020/5193707

Comparison of Supervised and Unsupervised Deep Learning Methods for Medical Image Synthesis between Computed Tomography and Magnetic Resonance Images

Yafen Li 1,2, Wen Li 1,2, Jing Xiong 1, Jun Xia 3, Yaoqin Xie 1,
PMCID: PMC7661122  PMID: 33204701

Abstract

Cross-modality medical image synthesis between magnetic resonance (MR) images and computed tomography (CT) images has attracted increasing attention in many medical imaging area. Many deep learning methods have been used to generate pseudo-MR/CT images from counterpart modality images. In this study, we used U-Net and Cycle-Consistent Adversarial Networks (CycleGAN), which were typical networks of supervised and unsupervised deep learning methods, respectively, to transform MR/CT images to their counterpart modality. Experimental results show that synthetic images predicted by the proposed U-Net method got lower mean absolute error (MAE), higher structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR) in both directions of CT/MR synthesis, especially in synthetic CT image generation. Though synthetic images by the U-Net method has less contrast information than those by the CycleGAN method, the pixel value profile tendency of the synthetic images by the U-Net method is closer to the ground truth images. This work demonstrated that supervised deep learning method outperforms unsupervised deep learning method in accuracy for medical tasks of MR/CT synthesis.

1. Introduction

Cross-modality medical image synthesis between magnetic resonance (MR) images and computed tomography (CT) images could benefit medical procedures in many ways. As a multiparameter imaging modality, magnetic resonance imaging (MRI) provides a wide range of image contrast mechanisms without ionizing radiation exposure, while CT images outperform MR images in acquisition time and resolution of bone structure. CT is also related with electron density which is critical for PET-CT attenuation correction and radiotherapy treatment planning [1]. Generating synthetic CT (sCT) images from MR images makes it possible to do MR-based attenuation correction in PET-MR system [26] and radiation dose calculation in MRI-guided radiotherapy planning [79]. Synthesizing MR images from CT images can enlarge the datasets for MR segmentation task and thus improve the accuracy of segmentation [10].

In recent years, there have been many efforts to work on medical image synthesis between MR and CT images. Among all these methods, deep learning method exhibited superior ability of learning a nonlinear mapping from one image domain to another image domain. It can be classified into two categories: supervised and unsupervised deep learning methods. Supervised deep learning methods required paired images for model training. In the MR/CT synthesis task, MR and CT images have to be well-registered at first and then used as inputs and corresponding labels for the neural network model to learn an end-to-end mapping. Nie et al. [11] used three-dimensional paired MR/CT image patches to train a three-layer fully convolutional network for estimating CT images from MR images. Other researchers [4, 5, 1215] have trained deeper network for MR-based CT image prediction. However, as for medical image dataset, it is not that easy to get paired MR and CT images. It may take a long time span to collect patients who are scanned by both MR and CT scanners. Registration of certain accuracy between MR and CT images are also necessary to make paired MR-CT dataset.

Unsupervised deep learning methods enabled the possibility of using unpaired images for image-to-image translation [1620]. It was first proposed for natural image synthesis and now has been implemented by many researchers for medical image synthesis [10, 2124]. Chartsias et al. [10] demonstrate the application of CycleGAN in synthesizing cardiac MR images from CT images, using MR and CT images of different patients. Nie et al. [21] synthesized MR images from CT images with a deep convolutional adversarial network. Since there are plenty of unpaired medical images, the available datasets could be easily enlarged.

Unlike natural images, accuracy is highly emphasized in medical images. In this paper, we aim to compare the accuracy of supervised and unsupervised learning-based image synthesis methods for pseudo-MR/CT generation tasks. Two typical networks of U-Net [25] and CycleGAN [17] were introduced as representatives of supervised and unsupervised learning methods, respectively. Mean absolute error (MAE), structural similarity index (SSIM), and peak signal-to-noise ratio (PSNR) of the synthetic results were calculated to evaluate their performance quantitatively. More detailed comparisons and discussions about the advantage and disadvantage of these methods are included in Results and Discussion.

2. Materials and Methods

2.1. Neural Network Models

In our experiments of pseudo-MR/CT generation tasks, U-Net and CycleGAN were used as the typical representative network of supervised and unsupervised deep learning methods, respectively.

U-Net has made a great achievement in segmentation tasks [2529]. The advantage of U-Net is that it could use very few images to make a good performance. In this study, we adapted U-Net to an end-to-end image synthesis task.

The basic architecture of U-Net consists of a contracting part to capture features and a symmetric expanding part to enable precise localization. As shown in Figure 1, we added LeakyReLU [30, 31] as activation operation before convolution operation in the contracting part of the network. Activation operation of LeakyReLU was replaced with ReLU [32] in the expanding part. Batch normalization [33] was introduced to U-Net to enable faster and more stable training. In Figure 1, the number of channels is denoted on top of each of the convolution operation, and the size of feature maps is signed in the parentheses.

Figure 1.

Figure 1

Architecture of proposed U-Net for image synthesis.

In the medical image synthesis task, input image and its corresponding label were fed to the proposed U-Net to train and learn an end-to-end nonlinear mapping between them. Figure 1 illustrated the MR-to-CT synthesis using U-Net architecture, which takes MR images as input and CT images as label to train a synthetic CT generating model. On the contrary, when we use CT images as input and MR images as labels, U-Net could be trained as a synthetic MR-predicting model. The loss function used in the proposed U-Net is

LU­Net=Ex, y~P^datafxy1 (1)

CycleGAN [17] which is proposed by Zhu et al. could be seen as an updated version of generative adversarial networks (GAN) [16]. GAN methods can learn a nonlinear mapping from input image domain to target image domain by adversarial training. CycleGAN introduced the idea of cycle consistency to general GAN methods. Cycle consistency adds restriction that the generated pseudoimage in target domain should be able to be transformed back to the original input image.

We used the CycleGAN architecture from Zhu et al. [17] for our medical image synthesis task. It takes unpaired MR and CT images as inputs to learn nonlinear mappings between these two image modalities. As illustrated in Figure 2, the CycleGAN architecture has two cycles, forward cycle and backward cycle. The forward cycle consists of three networks: two generative networks of G and F and one discriminator of DCT. The backward cycle uses the same generative networks of F and G and a counterpart discriminator of DMR.

Figure 2.

Figure 2

CycleGAN architecture for bidirection synthesis of MR and CT images. The forward cycle generated synthetic CT from input MR by G while F translate the synthetic CT back to the MR image domain. DCT discriminate whether the generated images is real or fake CT. The backward cycle generated synthetic MR from input CT by F while G translate the synthetic MR back to the CT image domain. DMR discriminate whether the generated images is real or fake MR. Two cycle-consistency loss was introduced to capture the intuition that the synthetic image should be translated back to the original image modality.

In the forward cycle, network G was used to generate synthetic CT (sCT) from input MR images, while network F generated synthetic MR (sMR) from network G-generated sCT images. Network DCT discriminates whether the generated sCT image is real CT or fake. The backward cycle works just the opposite way. Network F took CT images as input images and generated sMR; then, network G synthesized sCT from the F-generated sMR images. Network DMR was used to distinguish whether the sMR image is real MR or fake.

The adversarial losses of CycleGAN are as follows:

LGAN_G_MRtoCT=ECT~PdataCTlogDCTCT1+EMR~PdataMRlog1DCTGMR1,LGAN_F_CTtoMR=EMR~PdataMRlogDMRMR1+ECT~PdataCTlog1DMRFCT1. (2)

The cycle-consistency loss consists of forward cycle loss Lforward_cyc and the backward cycle loss Lbackward_cyc. It is represented as follows:

Lforward_cyc=EMR~PdataMRFGMRMR1,Lbackward_cyc=ECT~PdataCTGFCTCT1,LCycleconsistency=Lforward_cyc+Lbackward_cyc. (3)

Then, we have the full objective as the below equation:

LCycleGAN=LGAN_G_MRtoCT+LGAN_F_CTtoMR+λLCycleconsistency, (4)

where λ is the weight of the objectives of cycle consistency.

2.2. Cross-Modality MR/CT Image Synthesis and Evaluation

We used PyTorch to implement the proposed U-Net and CycleGAN. Both the networks were trained for bidirectional image synthesis, which includes learning a MR-to-CT model for generating synthetic CT images from MR images and a CT-to-MR model for generating synthetic MR images from CT images.

U-Net and CycleGAN used similar parameters for training nonlinear mapping models between MRI/CT images. Adam optimizer was adopted for both the networks. The batch size was set to 1. Both networks were trained for 200 epochs, with fixed learning rate for the first 100 epochs. The learning rate decreased linearly to 0 for the following 100 epochs.

Whole 2D slices of axial medical images with size of 256∗256 pixels were used as inputs. During the training process, the images would be padded to 286∗286 pixels and then random cropped to 256∗256 for data augmentation. While U-Net should utilize paired MR and CT datasets for training nonlinear mapping, CycleGAN can take use of unpaired MR and CT images as inputs for both the forward and backward cycles in training procedure. As for the CycleGAN method, we randomly shuffled the MR image input sequences and CT image input sequences in the paired datasets to make the input MR and CT slices unpaired. The MRI input sequence in unpaired datasets were not the same as that in paired datasets.

Three metrics were used to quantitatively characterize the accuracy of the prediction of synthetic images compared with the ground truth images. The mean absolute error (MAE) measures the discrepancies by voxels. Structural similarity index (SSIM) [34] quantifies the similarities in a whole image scale. Peak signal-noise-ratio (PSNR) assesses the quality of prediction.

These evaluation metrics are expressed as follows:

MAE=1HWi=1Hj=1WXi,jYi,j,SSIM=2μxμy+c12σxy+c2/μx2+μy2+c1σx2+σy2+c2,c1=K1L2,c2=K2L2,PSNR=10log10L/MSE,MSE=1HWi=1Hj=1WXi,jYi,j2, (5)

where H and W are the height and width of the images, respectively. X is the ground truth images, and Y is the predicted synthetic images. μx and μy are the average values of ground truth images and synthetic images, respectively. σx2 and σy2 are the variance of ground truth images and synthetic images, respectively. σxy represents the covariance of ground truth images and synthetic images. L denotes the dynamic range of the voxel values. c1 and c2 are two variables to stabilize the division with a weak denominator. Here, we take k1 = 0.01 and k2 = 0.03 by default.

2.3. Dataset Preparing

The datasets contain 34 patients. Each patient has both T2-weighted MR images and CT images of the head region. We acquired T2-weighted MR images (TR: 2500 ms, TE: 123 ms, 1∗1∗1 mm3, 256∗256) on a 1.5 T Avanto scanner (Siemens). The CT images (120 kV, 330 mA, exposure time: 500 ms, 0.5∗0.5∗1 mm3, 512∗512) were acquired on SOMATOM Definition Flash (Siemens).

In this experiment, CT images were resampled to a size of 256∗256 (1∗1 mm2) by bicubic interpolation [35] to match the voxel size of MR images. Binary head masks were generated by the Otsu threshold method [36] for MR and CT images to remove unnecessary background information around the head region.

Since the head region is mainly a rigid construction of bone structure, we applied rigid registration to the MR and CT images to make paired MR/CT images for the proposed U-Net. CT images were set as a fixed volume. MR images were set as a moving volume to register with CT images by Elastix toolbox [37]. The paired datasets were randomly shuffled to make an unpaired dataset for CycleGAN.

In our medical image synthesis task, 28 patients with 4063 image pairs were randomly selected for model training. The remaining 6 patients with 846 image pairs were used for evaluation procedure.

3. Results and Discussion

The results of synthetic MR and synthetic CT images generated by U-Net and CycleGAN and their ground truth are showed in Figure 3. The first column is the input images, and the second column is ground truth images. The third column showed the generated synthetic images predicted from input images by the two networks. The difference map between synthetic images and ground truth images was calculated and showed in the fourth column.

Figure 3.

Figure 3

(a–d) From left to right: the 4 columns are input images, ground truth images, synthetic images, and the difference maps. sCT results generated by U-Net (a) and CycleGAN (b), respectively; sMR results by U-Net (c) and CycleGAN (d), respectively.

The first two rows in Figure 3 are sCT images synthesized by U-Net and CycleGAN, respectively. For the task of synthesizing CT images from MR images, the soft tissue area is translated from high contrast to low contrast. It could be seen from the difference map images that the soft tissue area of synthetic CT images by both networks is well-translated with little error. The translation error mainly occurred in the bone area. Their difference map demonstrates that the sCT by CycleGAN synthesized more error than sCT by U-Net in the bone areas.

The third and fourth rows in Figure 3 are sMR images generated by U-Net and CycleGAN, respectively. It could be seen that sMR by CycleGAN seems more realistic for it has more complex contrast information than sMR by U-Net. However, the difference map images illustrated that the CycleGAN method generated much more error than U-Net does. The abundant image contrast information in sMR by CycleGAN may be false and unnecessary.

In synthesizing CT tasks, the difference between synthetic images and ground truth mainly occurs in the bone area. But in synthesizing MR tasks, the error is evenly distributed in the whole head region. It means synthesizing high contrast images of MR from low contrast image domain of CT is tougher than its reverse synthesizing direction.

To compare the image details, 1D profiles of pixel intensity were also plotted. Figure 4 shows the 1D profiles passing through the short red lines and long blue lines as indicated in the corresponding images in the first row. The red line is overlapped with the blue lines. The 1D profile in the second row of Figure 4 demonstrates pixel intensities of the long blue lines. The 1D profiles in the third row are the pixel intensities of the short red lines of 20 pixels, which shows close-ups of part of the long blue lines' 1D profile.

Figure 4.

Figure 4

Comparison of 1D profiles of pixel intensity passing through the short red lines and long blue lines as indicated in the images: (a) 1D profile and its close-up marked by the horizontal lines in ground truth CT, U-Net sCT, and CycleGAN sCT images; (b) 1D profile and its close-up marked by the horizontal lines in ground truth MR, U-Net sMR, and CycleGAN sMR images.

In the profiles, the red curve indicates pixel intensities of ground truth CT or MR. The blue curve represented for U-Net and the green curve for CycleGAN. It could be clearly seen in Figure 4(a) that the blue curve is close to the red curve, while some of the peaks of the green curve deviated from the red curve to an opposite direction. It means that the tendency of 1D profiles in sCT by U-Net was closer to the ground truth CT, while the CycleGAN method tends to generate fake contrast information in sCT images.

The profile in Figure 4(b) shows that the blue curve vibrated less from the red curve. Some peaks of the green curve deviated more from the red curve. It could be seen in the close-up 1D profile that some peaks of the green curve are biased to the opposite from the red curve, while the tendency of the blue cure seems like a smoothened or flattened red curve. It means that the pixel value of sMR by U-Net was closer to the ground truth but may lack contrast details. The pixel value of sMR by CycleGAN exhibits more deviation from the ground truth along the profile whereas the tendency may be false or exaggerated.

The quantitative metrics have been calculated for comparison. Figure 5 shows the MAE of sCT and sMR for each of the 6 patients in the evaluation datasets and the average result. It is obvious that the U-Net method generated lower MAE either in sCT image generation or sMR image generation for all the patients. This also demonstrates the robust performance of the U-Net method in bidirection MR/CT image translation tasks.

Figure 5.

Figure 5

(a) MAE of sMR images for all the 6 patients in test set and their average value. (b) MAE of sCT images generated for all the 6 patients in test set and their average value. Both the blue columns denoted the U-Net method, and the red columns represented the CycleGAN method.

Figures 5(a) and 5(b) show that the deviations of the MAE between the U-Net and CycleGAN method for sMR images of all the 6 patients are not as significant as those for sCT images. In Figure 3, the difference map of sMR indicated that the main predicted errors are evenly distributed in the whole head region, while the main error of sCT mainly occurs mainly in the bone structure. This could be interpreted that generating MR images of high soft tissue contrast from CT images of low soft tissue contrast is much complex than the inverse direction synthesis of generating CT from MR images.

Table 1 shows the overall statistics of three quantitative metrics for sCT by both the U-Net and CycleGAN methods. The SSIM values indicate that the sCT images by both methods have fairly high similarity with the ground truth CT images. The U-Net method outperformed the CycleGAN method with a much lower MAE of 65.36 HU, a higher SSIM of 0.972, and a higher PSNR of 28.84 dB. The average sCT MAE deviation between the two methods is nearly 30 HU.

Table 1.

Quantitative evaluation results between ground truth CT images and sCT images: MAE, SSIM, and PSNR.

Model MAE ± SD (HU) SSIM ± SD PSNR ± SD (dB)
U-Net 65.36 ± 4.08 0.972 ± 0.004 28.84 ± 0.57
CycleGAN 93.95 ± 5.89 0.955 ± 0.007 26.32 ± 0.55

Table 2 shows the overall statistics of three quantitative metrics for sMR images by the U-Net method and CycleGAN method. The U-Net method outperformed the CycleGAN method with a lower MAE of 73.43 HU, a higher SSIM of 0.946, and a higher PSNR of 32.35 dB.

Table 2.

Quantitative evaluation results between ground truth MR images and sMR images: MAE, SSIM, and PSNR.

Model MAE ± SD (HU) SSIM ± SD PSNR ± SD (dB)
U-Net 73.43 ± 9.16 0.946 ± 0.004 32.35 ± 0.78
CycleGAN 88.71 ± 10.04 0.924 ± 0.003 30.79 ± 0.73

The qualitative and quantitative results demonstrate that the proposed U-Net, a typical supervised learning method, outperforms CycleGAN, a representative advanced unsupervised learning method, in synthesis accuracy of medical image translation task. Since medical images highly value accuracy for the purpose of disease diagnosing, clinical treatment, and therapeutic effect evaluation, the supervised learning method is more recommended in medical practice.

Nevertheless, the success of supervised learning cannot do without well-registered image pairs. The performance of the trained model also depends on the registration accuracy of the paired images. Unlike natural images, paired medical images are not that easy to get. It would take a long time span to collect enough patients who need to be scanned for both MR and CT images at the same time. It is well-known that big amount of datasets could greatly improve the performance of the deep learning method. Though it outperforms the unsupervised learning method, the limit of dataset volume may constrain the further improvement of the supervised learning method in medical image synthesis tasks.

From the experiments discussed above, the image synthesis by using unsupervised learning methods still has a long way to go for practical application in clinic due to their relatively low accuracy. But still, the unsupervised learning method could benefit when there is lack of paired medical image datasets. The good news is that there are abundant easy-to-obtain retrospective unpaired MR and CT images for the unsupervised learning method to take advantage of. No registration is needed.

Our experiments show that when the same datasets were taken as inputs, the unsupervised learning method got inferior quality in the synthesis accuracy for medical image translation. But nonetheless, if the dataset is large enough, it could be expected that the performance of the unsupervised learning method would be improved to a certain acceptable extent in clinical practice.

4. Conclusions

Cross-modality medical image synthesis between MR and CT images could benefit a lot from the fast growing of deep learning methods. In this paper, we compared different deep learning-based image synthesis methods for pseudo-MR/CT generation, including the unsupervised learning method of CycleGAN and supervised learning methods of the proposed U-Net. Synthetic images produced by the CycleGAN method contain more but fake contrast information in the whole image scale. Though the proposed U-Net method blurred the generated pseudoimages, its pixel value profile tendency is basically close to the ground truth images. The quantitative results also indicate that the U-Net method outperformed the CycleGAN method, especially in synthesizing CT image task. As accuracy is highly demanded in medical procedures, we recommend the supervised method such as the proposed U-Net in cross-modality medical image synthesis at present clinical practice.

Acknowledgments

This work is supported in part by grants from the National Key Research and Development Program of China (2016YFC0105102), the Leading Talent of Special Support Project in Guangdong (2016TX03R139), the Fundamental Research Program of Shenzhen (JCYJ20170413162458312), the Natural Science Foundation of Guangdong Province (2017B020229002, 2015B020233011, and 2014A030312006), the Beijing Center for Mathematics and Information Interdisciplinary Sciences, and the National Natural Science Foundation of China (61871374).

Data Availability

The datasets of MR and CT images used to support the findings in this study are restricted by the Medical Ethics Committee of Shenzhen Second People's Hospital in order to protect patient privacy.

Conflicts of Interest

The authors declare that there is no conflict of interest.

References

  • 1.Dirix P., Haustermans K., Vandecaveye V. The value of magnetic resonance imaging for radiotherapy planning. Seminars in Radiation Oncology. 2014;24(3):151–159. doi: 10.1016/j.semradonc.2014.02.003. [DOI] [PubMed] [Google Scholar]
  • 2.Burgos N., Cardoso M. J., Thielemans K., et al. Attenuation correction synthesis for hybrid PET-MR scanners: application to brain studies. Ieee Transactions on Medical Imaging. 2014;33(12):2332–2341. doi: 10.1109/TMI.2014.2340135. [DOI] [PubMed] [Google Scholar]
  • 3.Mehranian A., Arabi H., Zaidi H. Vision 20/20: magnetic resonance imaging-guided attenuation correction in PET/MRI: challenges, solutions, and opportunities. Medical Physics. 2016;43(3):1130–1155. doi: 10.1118/1.4941014. [DOI] [PubMed] [Google Scholar]
  • 4.Liu F., Jang H., Kijowski R., Bradshaw T., McMillan A. B. Deep learning MR imaging-based attenuation correction for PET/MR imaging. Radiology. 2018;286(2):676–684. doi: 10.1148/radiol.2017170700. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Leynes A. P., Yang J., Wiesinger F., et al. Zero-echo-time and Dixon deep pseudo-CT (ZeDD CT): direct generation of pseudo-CT images for pelvic PET/MRI attenuation correction using deep convolutional neural networks with multiparametric MRI. Journal of Nuclear Medicine. 2018;59(5):852–858. doi: 10.2967/jnumed.117.198051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Wu Y., Yang W., Lu L., et al. Prediction of CT substitutes from MR images based on local diffeomorphic mapping for brain PET attenuation correction. Journal of Nuclear Medicine. 2016;57(10):1635–1641. doi: 10.2967/jnumed.115.163121. [DOI] [PubMed] [Google Scholar]
  • 7.Schmidt M. A., Payne G. S. Radiotherapy planning using MRI. Physics in Medicine and Biology. 2015;60(22):R323–R361. doi: 10.1088/0031-9155/60/22/R323. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Dowling J. A., Sun J., Pichler P., et al. Automatic substitute computed tomography generation and contouring for magnetic resonance imaging (MRI)-alone external beam radiation therapy from standard MRI sequences. International Journal of Radiation Oncology • Biology • Physics. 2015;93(5):1144–1153. doi: 10.1016/j.ijrobp.2015.08.045. [DOI] [PubMed] [Google Scholar]
  • 9.Dinkla A. M., Wolterink J. M., Maspero M., et al. MR-only brain radiation therapy: dosimetric evaluation of synthetic CTs generated by a dilated convolutional neural network. International Journal of Radiation Oncology • Biology • Physics. 2018;102(4):801–812. doi: 10.1016/j.ijrobp.2018.05.058. [DOI] [PubMed] [Google Scholar]
  • 10.Chartsias A., Joyce T., Dharmakumar R., Tsaftaris S. A. Cham: Springer International Publishing; 2017. Adversarial image synthesis for unpaired multi-modal cardiac data. [Google Scholar]
  • 11.Nie D., Cao X. H., Gao Y. Z., Wang L., Shen D. G. Estimating CT image from MRI data using 3D fully convolutional networks. Deep Learning and Data Labeling for Medical Applications. 2016;10008:170–178. doi: 10.1007/978-3-319-46976-8_18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Han X. MR-based synthetic CT generation using a deep convolutional neural network method. Medical Physics. 2017;44(4):1408–1419. doi: 10.1002/mp.12155. [DOI] [PubMed] [Google Scholar]
  • 13.Liu F., Yadav P., Baschnagel A. M., McMillan A. B. MR-based treatment planning in radiation therapy using a deep learning approach. Journal of Applied Clinical Medical Physics. 2019;20(3):105–114. doi: 10.1002/acm2.12554. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Fu J., Yang Y., Singhrao K., et al. Deep learning approaches using 2D and 3D convolutional neural networks for generating male pelvic synthetic computed tomography from magnetic resonance imaging. Medical Physics. 2019;46(9):3788–3798. doi: 10.1002/mp.13672. [DOI] [PubMed] [Google Scholar]
  • 15.Gong K., Yang J., Kim K., El Fakhri G., Seo Y., Li Q. Attenuation correction for brain PET imaging using deep neural network based on Dixon and ZTE MR images. Phys Med Biol. 2018;63(12):p. 125011. doi: 10.1088/1361-6560/aac763. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Goodfellow I., Pouget-Abadie J., Mirza M., et al. Generative adversarial nets. Advances in Neural Information Processing Systems 27 (Nips 2014) 2014;27 [Google Scholar]
  • 17.Zhu J. Y., Park T., Isola P., Efros A. A. Unpaired image-to-image translation using cycle-consistent adversarial networks. 2017 Ieee International Conference on Computer Vision (Iccv); 2017; Venice, Italy. pp. 2242–2251. [Google Scholar]
  • 18.Yi Z. L., Zhang H., Tan P., Gong M. L. DualGAN: unsupervised dual learning for image-to-image translation. 2017 Ieee International Conference on Computer Vision (Iccv); 2017; Venice, Italy. pp. 2868–2876. [Google Scholar]
  • 19.Isola P., Zhu J. Y., Zhou T. H., Efros A. A. Image-to-image translation with conditional adversarial networks. 30th Ieee Conference on Computer Vision and Pattern Recognition (Cvpr 2017); 2017; Hawaii, America. pp. 5967–5976. [Google Scholar]
  • 20.Lin J. X., Xia Y. C., Qin T., Chen Z. B., Liu T. Y. Conditional image-to-image translation. 2018 Ieee/Cvf Conference on Computer Vision and Pattern Recognition (Cvpr); 2018; Salt Lake City, America. pp. 5524–5532. [Google Scholar]
  • 21.Nie D., Trullo R., Lian J., et al. Medical image synthesis with context-aware generative adversarial networks. Medical Image Computing and Computer-Assisted Intervention. 2017. pp. 417–425. [DOI] [PMC free article] [PubMed]
  • 22.Maspero M., Savenije M. H. F., Dinkla A. M., et al. Dose evaluation of fast synthetic-CT generation using a generative adversarial network for general pelvis MR-only radiotherapy. Phys Med Biol. 2018;63(18):p. 185001. doi: 10.1088/1361-6560/aada6d. [DOI] [PubMed] [Google Scholar]
  • 23.Emami H., Dong M., Nejad-Davarani S. P., Glide-Hurst C. K. Generating synthetic CTs from magnetic resonance images using generative adversarial networks. Medical Physics. 2018;45(8):3627–3636. doi: 10.1002/mp.13047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Costa P., Galdran A., Meyer M. I., et al. End-to-end adversarial retinal image synthesis. Ieee Transactions on Medical Imaging. 2018;37(3):781–791. doi: 10.1109/TMI.2017.2759102. [DOI] [PubMed] [Google Scholar]
  • 25.Ronneberger O., Fischer P., Brox T. U-net: convolutional networks for biomedical image segmentation. International Conference on Medical Image Computing and Computer-Assisted Intervention; 2015; Munich, Germany. pp. 234–241. [Google Scholar]
  • 26.Yang Y., Feng C., Wang R. Automatic segmentation model combining U-Net and level set method for medical images. Expert Systems with Applications. 2020;153:p. 113419. doi: 10.1016/j.eswa.2020.113419. [DOI] [Google Scholar]
  • 27.Chen S., Yang H., Fu J., et al. U-Net plus: deep semantic segmentation for esophagus and esophageal cancer in computed tomography images. Ieee Access. 2019;7:82867–82877. doi: 10.1109/ACCESS.2019.2923760. [DOI] [Google Scholar]
  • 28.Abraham N., Khan N. M. A novel focal Tversky loss function with improved attention U -Net for lesion segmentation. 2019 Ieee 16th International Symposium on Biomedical Imaging (Isbi 2019); 2019; Venice, Italy. pp. 683–687. [Google Scholar]
  • 29.Zeng Y. Y., Chen X. Y., Zhang Y., Bai L. F., Han J. Dense-U-Net: densely connected convolutional network for semantic segmentation with a small number of samples. Tenth International Conference on Graphics and Image Processing (Icgip 2018); 2019; Chengdu, China. [Google Scholar]
  • 30.Xu B., Wang N., Chen T., Li M. Empirical evaluation of rectified activations in convolutional network. ICML deep learning workshop. 2015. http://arxiv.org/abs/1505.00853.
  • 31.Maas A. L., Hannun A. Y., Ng A. Y. Rectifier nonlinearities improve neural network acoustic models. Proceedings of the 30th International Conference on Machine Learning; 2013; Atlanta, Georgia, USA. [Google Scholar]
  • 32.Nair V., Hinton G. E. Rectified linear units improve restricted Boltzmann machines. Proceedings of the 27th international conference on machine learning (ICML-10); 2010; Haifa, Israel. pp. 807–814. [Google Scholar]
  • 33.Ioffe S., Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. International conference on machine learning; 2015; Lille, France. pp. 448–456. [Google Scholar]
  • 34.Wang Z., Bovik A. C., Sheikh H. R., Simoncelli E. P. Image quality assessment: from error visibility to structural similarity. Ieee Transactions on Image Processing. 2004;13(4):600–612. doi: 10.1109/TIP.2003.819861. [DOI] [PubMed] [Google Scholar]
  • 35.Prashanth H., Shashidhara H., Balasubramanya Murthy K. N. Image scaling comparison using universal image quality index. Advances in Computing, Control, & Telecommunication Technologies, 2009. ACT'09. International Conference; December 2009; Trivandrum, Kerala, India. pp. 859–863. [DOI] [Google Scholar]
  • 36.Otsu N. A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man, and Cybernetics. 1979;9(1):62–66. doi: 10.1109/TSMC.1979.4310076. [DOI] [Google Scholar]
  • 37.Klein S., Staring M., Murphy K., Viergever M. A., Pluim J. P. W. Elastix: a toolbox for intensity-based medical image registration. Ieee Transactions on Medical Imaging. 2010;29(1):196–205. doi: 10.1109/TMI.2009.2035616. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets of MR and CT images used to support the findings in this study are restricted by the Medical Ethics Committee of Shenzhen Second People's Hospital in order to protect patient privacy.


Articles from BioMed Research International are provided here courtesy of Wiley

RESOURCES