Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2025 Feb 27;94(2):797–809. doi: 10.1002/mrm.30478

Repeatability‐encouraging self‐supervised learning reconstruction for quantitative MRI

Zihao Chen 1,2,3, Zheyuan Hu 1,2,3, Yibin Xie 3, Debiao Li 2,3, Anthony G Christodoulou 1,2,3,
PMCID: PMC12137793  PMID: 40014485

Abstract

Purpose

The clinical value of quantitative MRI hinges on its measurement repeatability. Deep learning methods to reconstruct undersampled quantitative MRI can accelerate reconstruction but do not aim to promote quantitative repeatability. This study proposes a repeatability‐encouraging self‐supervised learning (SSL) reconstruction method for quantitative MRI.

Methods

The proposed SSL reconstruction network minimized cross‐data‐consistency between two equally sized, mutually exclusive temporal subsets of k‐t‐space data, encouraging repeatability by enabling each subset's reconstruction to predict the other's k‐t‐space data. The method was evaluated on cardiac MR Multitasking T1 mapping data and compared with supervised learning methods trained on full 60‐s inputs (Sup60) and on split 30‐s inputs (Sup30/30). Reconstruction quality was evaluated on full 60‐s inputs, comparing results to iterative wavelet‐regularized references using Bland–Altman limits of agreement (LOAs). Repeatability was evaluated by splitting the 60‐s data into two 30‐s inputs, evaluating T1 differences between reconstructions from the two halves of the scan.

Results

On 60‐s inputs, the proposed method produced comparable‐quality images and T1 maps to the Sup60 method, with T1 values in general agreement with the wavelet reference (LOA Sup60 = ±75 ms, SSL = ±81 ms), whereas the Sup30/30 method generated blurrier results and showed poor T1 agreement (LOA Sup30/30 = ±132 ms). On back‐to‐back 30‐s inputs, the proposed method had the best T1 repeatability (coefficient of variation SSL = 6.3%, Sup60 = 12.0%, Sup30/30 = 6.9%). Of the three deep learning methods, only the SSL method produced sharp and repeatable images.

Conclusion

Without the need for labeled training data, the proposed SSL method demonstrated superior repeatability compared with supervised learning without sacrificing sharpness, and reduced reconstruction time versus iterative methods.

Keywords: deep learning, quantitative MRI, reconstruction, repeatability, self‐supervised learning, subspace method

1. INTRODUCTION

Quantitative MRI provides detailed insights into tissue composition and structure, aiding in the early detection and characterization of pathological changes in diseases such as multiple sclerosis, brain tumors, liver cirrhosis, and cardiac diseases. 1 , 2 , 3 Quantitative MRI obtains quantitative parameter maps such as T1 and T2 maps, providing an objective assessment of tissue status. 4 , 5 , 6 These maps allow for precise measurement of tissue properties, enabling the detection of subtle changes that might be missed with qualitative techniques. In addition, quantitative interpretation facilitates a more standardized assessment across time and between different scanners or institutions, thus proving invaluable for longitudinal studies and multicenter clinical trials. 7 , 8 , 9 Repeatability in quantitative MRI is crucial for ensuring the reliability and accuracy of measurements over time. In clinical and research settings, consistent and repeatable results are essential for tracking disease progression, evaluating treatment efficacy, and conducting longitudinal studies. 10 , 11 Repeatability ensures that variations in imaging results are due to actual changes in the biological tissue rather than inconsistencies in the imaging process.

Both multiparameter quantitative MRI and dynamic quantitative MRI are multidimensional imaging problems, as a series of images is necessary for parameter fitting. 4 , 5 , 12 , 13 , 14 Multidimensional approaches not only capture detailed spatial anatomical information but can also resolve temporal dynamics such as cardiac motion, respiration, and quantify relaxation times like T1 and T2. However, the acquisition of multidimensional MRI, especially in moving organs, often involves high undersampling, which brings additional challenges for reconstruction.

Highly undersampled k‐t space data can be iteratively reconstructed with constrained subspace/globally low‐rank, 15 locally low‐rank, 16 , 17 and/or compressed‐sensing (CS) 18 models, especially alongside acquisition schemes such as randomized and/or non‐Cartesian sampling that are motion robust and produce incoherent artifacts. These constrained models are often used in concert with quantitative imaging frameworks as well, such as MR fingerprinting 4 , 19 , 20 and MR Multitasking. 5 , 6 However, iterative reconstructions require intensive computation and long reconstruction time. Recently, deep learning or combinations of deep learning with subspace or compressed sensing (e.g., deep subspace learning) have shown great advantage over iterative reconstruction in reducing reconstruction time and improving image quality, 21 , 22 , 23 , 24 , 25 , 26 especially physics‐driven methods. 27 , 28 , 29 Self‐supervised learning (SSL) models are especially useful when ground‐truth data are limited or unavailable, 30 , 31 , 32 , 33 , 34 , 35 , 36 , 37 which is often the case for MRI data. A popular form of SSL MR reconstruction 33 , 35 is to split undersampled k‐space into two disjoint parts for generating images and for calculating the loss function, respectively, requiring no labels for supervision. Several previous deep‐learning MR reconstruction works have focused on improving quantitative MRI reconstruction 38 , 39 , 40 for static images.

However, there are few if any previous deep‐learning reconstruction works that specifically focus on improving quantitative MRI's repeatability, which is essential for the reliability of quantitative MRI. We propose that SSL methods offer a natural way to enforce repeatability through a data‐splitting scheme that splits the k‐t space data along temporal dimension(s) and loss functions, which encourages the repeatability of reconstruction results between subsets.

Thus, here we propose a repeatability‐encouraging SSL reconstruction approach based on temporal data splitting for quantitative dynamic MRI. Like the consistency loss between image pairs used in previous works, 41 , 42 we developed a temporal cross‐data‐consistency (CDC) loss function and a related evaluation metric to facilitate the repeatability of quantitative MRI. Our study applies the proposed SSL reconstruction method to T1 CMR Multitasking, 5 , 6 a multidimensional subspace‐based quantitative MR technique, so we specifically formulate our SSL method within subspace constraints to reduce the computational burden of multidimensional MR reconstruction.

2. THEORY

2.1. Subspace‐constrained dynamic MRI reconstruction

Subspace constraints efficiently reconstruct quantitative and dynamic MR image series by leveraging the low‐rank nature of correlated images. In subspace imaging, a series of T images with Nvox voxels is represented as a matrix XNvox×T and decomposed into a spatial factor UNvox×L and a temporal factor ΦL×T according to X=UΦ. The rank of the matrix X is therefore no larger than L, which is typically less than minNvox,T due to spatiotemporal correlation throughout the image series. When the rows of Φ constitute an orthonormal basis, U can be interpreted as coordinates within the L‐dimensional subspace spanned by the rows of Φ. A suitable Φ can often be quickly extracted from a subset of acquired data b through principal component analysis or singular value decomposition (SVD), 15 a dictionary, 19 , 43 or by a combination of both. 5 The most time‐consuming step of reconstruction is typically to estimate U, often by solving:

U^=argminUbA(U)22+λR(U), (1)
A(U)=Ω([FSU]Φ). (2)

Here, the encoding operator A consists of a sensitivity map operator S, the Fourier transform F (in practice applied as a fast Fourier transform [FFT] for Cartesian acquisition or the more time‐consuming nonuniform FFT [NUFFT] 44 when using non‐Cartesian acquisition), the temporal factor Φ, and the k‐t space undersampling operator Ω. Most implementations include a regularization term λR(U), often a sparse regularizer to complementarily leverage‐compressed sensing. Conventionally, this minimization problem is solved by iterative methods such as alternating direction method of multipliers (ADMM 45 ) or the fast iterative soft‐thresholding algorithm (FISTA 46 ).

The degrees of freedom in subspace‐based dynamic imaging are dominated by the size of U, which has NvoxL complex elements. In the context of multidimensional imaging, the degrees of freedom grow approximately linearly with the number of time dimensions rather than exponential growth that would otherwise be dictated by the curse of dimensionality. Thus, multidimensional images can be reconstructed and stored in much less memory when using subspace methods. This has motivated the use of the subspace‐based MR Multitasking method for quantitative imaging in the presence of motion without breath‐holds or electrocardiographic monitoring. However, slow iterative non‐Cartesian reconstruction is still a barrier to clinical adoption.

Supervised deep subspace learning has achieved clinically practical reconstruction times 25 for MR Multitasking and other subspace reconstruction frameworks. However, dynamic imaging often lacks ground‐truth data because of the high undersampling rates, so supervised learning is typically trained on labels generated through regularized iterative reconstruction. This process of generating labels is time‐consuming and sensitive to hyperparameter tuning that affects the strength of regularization in compressed‐sensing reconstruction. Furthermore, iterative reconstruction methods can introduce discrepancies from the true image, and training on these approximated labels may lead the model to learn these inaccuracies, ultimately compromising reconstruction fidelity. Importantly for quantitative imaging, supervised learning does not directly enforce the repeatability of quantitative maps and is therefore not aimed at optimizing clinical utility.

2.2. SSL MRI reconstruction

SSL MRI reconstruction 30 , 31 , 32 , 33 , 34 , 35 , 36 , 37 , 47 , 48 , 49 , 50 offers a compelling alternative to supervised learning, particularly when fully sampled data are unavailable. A common approach to SSL‐MRI reconstruction is to split undersampled k‐space locations Ω into two subsets Θ and Λ, where Θ are the input samples used for image reconstruction, and Λ are used to define the loss function for training. The self‐supervised learning via data undersampling (SSDU) 35 framework formulates this training process as follows:

minθ1Ni=1NLbΛi,AΛifbΘi,AΘi;θ, (3)

where N is the number of training data; b contains acquired k‐space data; A is the MRI encoding operator; f is the neural network with trainable parameters θ; and L is the loss function. In training a network that can predict the samples in Λ from the samples in Θ, SSDU achieves comparable reconstruction performance to supervised deep learning, without the need of acquiring fully sampled k‐space data.

In SSDU training designs, the split of Θ and Λ has usually been based on two‐dimensional random sampling in the k x k y plane, including subsampling within readout lines. This training design may cause a potential distribution mismatch between training and testing, as all the acquired k‐space lines are used in testing. 48 In addition, for dynamic MRI, previous SSL methods have performed undersampled k‐space data splitting inside each time frame 31 , 33 , 36 rather than along the time dimension.

3. METHODS

3.1. The proposed framework

Here we propose a new SSL reconstruction framework that treats both Θ and Λ as potential inputs to the network, introducing CDC loss to ensure that the reconstruction from either subset of data predicts the other, as follows:

minθ1Ni=1NLbΛi,AΛifbΘi,AΘi;θ+LbΘi,AΘifbΛi,AΛi;θ. (4)

Here, Θ and Λ are mutually exclusive, equally populated subsets, with entire frequency‐encoding readouts allocated to either Θ or Λ rather than being split between the two. This allocation would be a more physically realistic representation of prospective undersampling patterns in testing data. Because this allocation is done on a readout‐by‐readout basis, we refer to it as temporal splitting. By ensuring that half of the k‐space data are predictable from the other half, cross‐data consistency allows image reconstruction from either Θ or Λ and promotes repeatable imaging.

3.2. Subspace implementation for evaluation within MR Multitasking

In this work, we use T1 CMR Multitasking 5 , 6 as the quantitative imaging framework to evaluate this new SSL approach. MR Multitasking uses subspace modeling to separate the temporal information and spatial information early in the image‐reconstruction pipeline: Individual images at different time points share the same spatial factor U, regardless of any motion or contrast‐weighting differences between those images, as temporal differences such as motion patterns are fully encapsulated within the temporal factor Φ. Multitasking derives a scan‐specific temporal factor Φ from interleaved self‐navigation data (e.g., a repeated 0° spoke in radial acquisitions) that capture respiratory and cardiac motion information, in concert with a relaxation model in the form of a contrast evolution basis calculated from the SVD of a signal evolution dictionary. 5 , 6 More specifically, the self‐navigation data are sorted into a tensor based on their motion states and contrast weightings, and low‐rank tensor completion constrained by the SVD‐derived relaxation basis estimates unseen motion state/contrast weighting combinations. The high‐order SVD of the completed tensor yields Φ. The spatial factor U is then calculated given this fixed estimate of Φ. However, the SSL reconstruction method proposed here focuses on estimating U in the presence of a fixed Φ but does not require Φ to be specifically calculated as previously. The rest of this paper will further formulate Eq. (4) specifically in the context of non‐Cartesian subspace imaging used in MR Multitasking.

In the training phase, we randomly split 60 s of MR Multitasking raw data into two 30‐s subsets Θ and Λ, as shown in Figure 1. The random split was regenerated for each training step. This random split ensures that training is robust to motion mismatches between the first and second halves of the scan, although contiguous first‐half/second‐half inputs can still be used during inference, as we will later show. To obtain the initial spatial factor U0,Θ and U0,Λ from each 30‐s subset of data, we first backprojected the corresponding half of the k‐space data b onto the matching half of the predetermined temporal factor Φ, then we performed regridding and complex coil combination. This process is specifically described by

U0,Θ=SHFNUHΩ*DbΘΦΘH, (5)
U0,Λ=SHFNUHΩ*DbΛΦΛH, (6)

where D applies density compensation (in this case, a fixed function that inverted the Voronoi density of general radial sampling), and FNUH represents the adjoint NUFFT. These two spatial factors U0,Θ and U0,Λ were fed into a U‐Net 51 (Figure S2) with identical weights for both inputs. The training loss function compares the network outputs Unn,Θ and Unn,Λ to the other halves of k‐space (bΛ and bΘ, respectively) using a computationally efficient in‐subspace loss calculation as described in Section 3.3.

FIGURE 1.

FIGURE 1

The proposed temporal‐split subspace self‐supervised learning reconstruction framework. (A) Training phase. (B) Inference phase. The shapes of bΘ and U0,Θ are annotated in (A), where NRO=320 is the number of k‐space points in a readout line; TΘ4000 is the number of time points in Θ set; Nx=Ny=320; and L=32. All other b and U are in the similar shapes.

In the inference phase, the network can accept inputs from the full 60‐s scan or from either contiguous 30‐s half of the scan. Here, the initial spatial factor U0 from any of these inputs was fed into the U‐Net to generate an output Unn, which was then passed through an in‐subspace data‐consistency (DC) layer 25 to produce the final reconstruction Udc.

3.3. Loss functions

Repeatable imaging implies that Unn,ΘUnn,Λ. In the k‐space domain, this is equivalent to

AΛUnn,ΘAΛUnn,Λ=bΛ, (7)
AΘUnn,ΛAΘUnn,Θ=bΘ. (8)

To promote the equivalences in Eqs. (7) and (8), we designed CDC loss functions based on the L2‐norm of the k‐space differences, as follows:

LossΘΛ=AΛUnn,ΘbΛ22=TrUnn,ΘHAΛ*AΛUnn,Θ2Unn,ΘHAΛ*bΛ+bΛHbΛ, (9)
LossΛΘ=AΘUnn,ΛbΘ22=TrUnn,ΛHAΘ*AΘUnn,Λ2Unn,ΛHAΘ*bΘ+bΘHbΘ, (10)
LossCDC=LossΘΛ+LossΛΘ. (11)

The trace formulations in Eqs. (9) and (10) allow fast and memory‐efficient loss calculation, because each A*(b) resides in, and each A*A(·) operates within, the low‐dimensional subspace domain rather than the time domain. 25 Additionally, each A*(b) and the operators within each A*A(·) can be precomputed, and the static bHb terms can be ignored, further streamlining the calculation. The value of LossCDC is the final loss function for optimization.

3.4. Data set and implementation details

We collected 122 T1 CMR Multitasking 6 data sets from three 3T scanners: MAGNETOM Verio, MAGNETOM Vida, and Biograph mMR (Siemens Healthcare, Erlangen, Germany). This study was approved by the institutional review board of Cedars‐Sinai Medical Center, and written informed consent was obtained from all human participants. Each scan acquired 60 s of k‐t space data using a continuous inversion‐recovery (IR) fast low‐angle shot sequence (flip angle [FA] = 5°, repetition time [TR] = 3.6 ms, echo time [TE] = 1.4 ms, IR spacing = 2.5 s) featuring golden‐angle radial trajectories interleaved with a repeated self‐navigator spoke. The image matrix size was 320 × 320, with a spatial resolution of 1.7 × 1.7 mm in‐plane and a slice thickness of 8 mm. Subject‐specific temporal factors Φ were generated from a combination of dictionaries and self‐navigator spoke analysis. 5 , 6 Each multidimensional temporal factor Φ had 20 cardiac phases, 6 respiratory phases, and 344 T1‐recovery timepoints. Thus, multiplying U by Φ produces a multidimensional array of images with Nvox = 320 × 320 = 102 400 voxels by T = 20 × 6 × 344 = 41 280 temporal frames. The subspace dimension was L = 32, such that U was only 102 400 × 32. Reference spatial factors U were obtained by iteratively solving Eq. (1) using wavelet‐regularized ADMM. Of the 122 data sets, 96 cases were used for training, 12 cases were used for validation, and 14 cases were used for testing. Each data case refers to a unique subject.

We used a U‐Net architecture with an initial input channel count of 64, doubling the number of channels after each 2 × 2 max‐pooling operation. The network includes two max‐pooling layers, resulting in 256 channels at the deepest layer. For image upsampling, we used PyTorch's upsample layers. The detailed layer information is shown in Figure S2. The spatial factor U is a complex matrix with L = 32, so we concatenated the real part and imaginary part of U to feed a 64‐channel input layer, treating the rank dimension as the channel dimension. For the network output, we recomposed 64 output channels back into a complex spatial factor with L = 32. The DC layer in Figure 1B used 10 conjugate‐gradient (CG) iterations to minimize the DC function bA(U)22 using GPU calculation.

The network was implemented and trained with PyTorch on a NVIDIA Quadro RTX 8000 GPU with 48GB memory. The network's total number of trainable parameters is 1 996 224. The network was trained for 200 epochs with batch size set to 1. We use the Adam optimizer and a learning rate of 0.001.

3.5. Evaluation methods

The proposed SSL method was compared with supervised deep subspace learning reconstruction. 25 The supervised method (Figure 2), which has an identical U‐Net structure to the proposed method, was trained to minimize L2 loss between the network output and the conventional wavelet iterative reconstruction of a 60‐s scan. The supervised network was trained twice for separate evaluations: once on 30‐s inputs (Sup 30/30; Figure 2B) to facilitate direct comparison with the SSL method, which is also trained on split 30‐s inputs, and once on full 60‐s inputs (Sup 60; Figure 2A) to afford the supervised network the full data available to it. Both the SSL and supervised methods' outputs were passed through a DC layer with 10 CG iterations during inference.

FIGURE 2.

FIGURE 2

Training schemes of two supervised learning methods. (A) Supervised learning trained with 60‐s scan input. (B) Supervised learning trained with split 30‐s scan inputs. MSE, mean square error.

As is shown in Figure 3, the evaluation consists of two inference scenarios. The first scenario tests the proposed SSL method's performance using the standard 60‐s scan input. This scenario provides the most direct comparison against the Sup60 method, which is trained with the full 60‐s input, and against the 60‐s wavelet iterative reconstruction, the reference standard that was unseen by the SSL network. The second scenario compares the repeatability of different deep‐learning reconstruction methods by splitting the 60‐s data in the test set into two 30‐s halves before inference. This allowed us to compare whether and how SSL training also improved CDC during inference, and more importantly, whether this translated to improved T1 repeatability.

FIGURE 3.

FIGURE 3

Evaluation methods. (A) Evaluation on the reconstruction performance with standard 60‐s scan time. (B) Evaluation of repeatability with split 30‐s data. CG, conjugate gradient; DC, data consistency; SSL, self‐supervised learning.

3.6. Evaluation of standard 60‐s scan

We evaluated the weighted images and quantitative T1 maps of various reconstruction methods on 60‐s MR Multitasking inputs from the test set. We compared four reconstruction methods: (i) the proposed SSL method; (ii) supervised learning trained on 60‐s inputs (Sup60); (iii) supervised learning trained on split 30‐s inputs (Sup30/30); and (iv) iterative wavelet‐regularized reconstruction. We used Bland–Altman plots to compare the myocardial septal T1 values between the deep learning methods against the iterative wavelet‐regularized reference reconstructions. Paired t‐tests were used to detect T1 biases between methods, regarding p < 0.05 as significant.

3.7. Evaluation of repeatability

We further compared the repeatability of each deep‐learning reconstruction method (Methods 1–3 explained previously). We split each 60‐s scan in the test set into two contiguous 30‐s halves (first 30 s and second 30 s). Note that unlike the random splitting used during training, this contiguous data‐splitting scheme is equivalent to two back‐to‐back accelerated acquisitions as in a test–retest experiment.

We used a CDC metric (MetricCDC; Figure 3B) as an initial surrogate for repeatability of network outputs. This metric, closely related to our CDC training loss, evaluated how well the reconstructed spatial factor from one 30‐s half of the scan could predict data from the other 30‐s half of the scan, as follows:

MetricCDC=12MetricCDC,1+MetricCDC,2=12A1(Unn,2b122+A2Unn,1b222. (12)

Here, A1 and b1 correspond to the encoding operator and k‐space data for the first 30‐s scan, and A2 and b2 correspond to those for the second 30‐s scan. We calculated MetricCDC for the proposed SSL method and both supervised learning methods, evaluating the network outputs before DC layers (like the training loss function), focusing on the intrinsic properties of each network. Paired t‐tests were used to compare CDC results between methods, regarding p < 0.05 as significant.

In addition to the CDC metric, we generated the quantitative T1 maps for different methods from the 30‐s inputs. We used coefficients of variation (CoVs) and Bland–Altman plots to compare the T1 value differences between the two 30‐s halves for the SSL method and both supervised learning methods to measure the T1 quantification repeatability of each reconstruction method. The CoVs are calculated using the standard deviations of test–retest differences divided by the mean value of all test–retest samples. We evaluated the results both with and without DC layers to evaluate both the intrinsic performance of the neural networks as well as their interactions with DC layers. Paired t‐tests were used to detect T1 biases between repetitions, regarding p < 0.05 as significant.

4. RESULTS

The reconstruction time for all deep learning methods was 2 min per scan, primarily due to time spent in the CG DC layer. In comparison, the conventional wavelet reconstruction took about 25 min. All the reconstructions were performed on GPU.

4.1. Results of standard 60‐s scan

Figure 4 shows reconstructed images and corresponding T1 maps from supervised learning (Sup60 and Sup30/30), the proposed SSL method, and the iterative wavelet reconstruction. Both the Sup60 and SSL methods produced images and T1 maps of comparable quality to the wavelet reconstruction. The Sup30/30 reconstruction appears noticeably blurrier than the other methods.

FIGURE 4.

FIGURE 4

Reconstruction images and T1 maps from different reconstruction methods for 60‐s scan. (A,B) Two subjects. Red arrows point to the myocardial regions with obvious different T1 values. SSL, Self‐supervised learning.

Figure 5 presents Bland–Altman plots comparing myocardial septal T1 values of each deep‐learning reconstruction methods against values from conventional iterative wavelet reconstruction. The Sup60 and SSL methods both showed substantially better agreement with the wavelet reference than did the Sup30/30 method. The Sup30/30 and SSL methods show small but statistically significant T1 biases versus the wavelet reference (p < 0.01), whereas the Sup60 method does not have a statistically significant bias (p = 0.13).

FIGURE 5.

FIGURE 5

Bland–Altman plots of myocardial septal T1 values from different reconstruction methods for 60‐s scan. (A) Sup60 versus wavelet reconstruction. (B) Sup30/30 versus wavelet reconstruction. (C) Self‐supervised learning (SSL) versus wavelet reconstruction. SD, standard deviation.

4.2. Repeatability evaluation

Figure 6 shows the CDC metrics for each supervised learning method and the proposed SSL method. The SSL method produced significantly lower CDC values than both the Sup60 and Sup30/30 methods (p < 0.001), indicating that the SSL‐reconstructed spatial factors from one 30‐s half of the scan could better predict the data from the other half.

FIGURE 6.

FIGURE 6

Cross‐data‐consistency (CDC) metrics for supervised learning method and the proposed self‐supervised learning (SSL) method.

Figure 7 displays reconstructed images and T1 maps from back‐to‐back 30‐s halves from 1 subject, for all three deep learning methods. The SSL results strike a balance between noise and image sharpness: Less noise is evident in the SSL T1 maps than Sup60's T1 maps, without the prominent blurring seen in the Sup 30/30 results. The Sup60 method produces visibly discrepant T1 values between the two 30‐s halves.

FIGURE 7.

FIGURE 7

The reconstruction images and T1 maps of 30‐s scan from supervised learning methods and the proposed self‐supervised learning (SSL) method. All reconstructions are with data‐consistency layers. (A,B) Two subjects.

In Figure 8, Bland–Altman plots illustrate the repeatability of T1 values between the two 30‐s scans for all deep‐learning reconstruction methods. The SSL method produces the best T1 repeatability, with LOAs approximately half that of Sup60's. The SSL method achieves the best CoV of 6.3%, followed by the Sup30/30 method (CoV: 6.9%) and the Sup60 method (CoV: 12.0%). Additionally, the Sup60 method produces significantly biased T1 values (p = 0.04) between scans, whereas the SSL method (p = 0.13) and Sup30/30 (p = 0.19) method do not.

FIGURE 8.

FIGURE 8

Bland–Altman plots of the myocardial septal T1 values for 30‐s scans from different reconstruction methods with data‐consistency (DC layers. (A) Sup60 reconstruction first 30‐s scan versus second 30‐s scan. (B) Sup30/30 reconstruction first 30‐s scan versus second 30‐s scan. (C) Self‐supervised learning (SSL) reconstruction first 30‐s scan versus second 30‐s scan. SD, standard deviation.

Note that all the reconstructions shown in Figures 7 and 8 are after the DC layers. Figure S1 shows Bland–Altman plots for the reconstructions before the CG‐DC layers, revealing similar trends but with even more pronounced differences. Figure S3 shows images and T1 maps before the DC layers for supervised learning methods and the SSL method. Figure 7 and Figure S3 can be viewed together for visualizing the performance differences with and without the DC layer, which is consistent with the results in Figure S1.

The trends in T1 repeatability across methods were aligned between both Bland–Altman comparisons and with the CDC results in Figure 6: The Sup30/30 method produced more repeatable results than the Sup60 method, whereas the SSL method was more repeatable than both supervised methods.

5. DISCUSSION

In this study, we developed a repeatability‐encouraging SSL reconstruction method for quantitative MRI. The key technical contribution of our approach is the introduction of CDC loss and evaluation metrics as a surrogate for repeatability. Additional contributions include the random splitting data in the temporal dimension rather than within readouts for training, and in‐subspace DC calculations for computationally efficient training in non‐Cartesian dynamic image reconstruction. We evaluated our method in the context of CMR Multitasking T1 mapping, comparing repeatability‐encouraging SSL method to supervised deep learning reconstructions and the conventional iterative wavelet reconstruction.

On full‐duration 60‐s inputs, we compared deep learning reconstructions to wavelet reconstruction, which was previously evaluated against references in phantom and in vivo studies. 5 , 6 The Sup60 and SSL reconstructions produced comparable‐quality images and T1 maps, with T1 values in general agreement with the wavelet reference. Notably, the SSL method achieved its results without being exposed to wavelet reconstruction results during training. SSL did show a small but statistically significant bias, which is commonplace between T1 mapping methods 52 and can be corrected if the bias is systematic and therefore repeatable. 3 The Sup30/30 method generated noticeably blurrier results than the other methods and showed poor T1 agreement versus the reference. The Sup30/30 method was trained on 30‐s inputs, so its relatively poor performance on 60‐s inputs may be due to supervised learning's sensitivity to domain differences in scan time/acceleration factor. The SSL method was also trained on 30‐s inputs but was able to maintain image quality and T1 agreement, suggesting potential resilience to different scan durations/acceleration factors.

Comparing the test–retest repeatability of two half‐duration 30‐s inputs, the proposed SSL framework emerged as the most repeatable method. Sup30/30's repeatability approached that of SSL, likely because its training used the same 60‐s‐derived label for each of the two 30‐s inputs per scan. However, Sup30/30's repeatability came at the cost of image sharpness—a drawback that SSL avoided. The Sup60 network produced sharper images and maps than Sup30/30, but its repeatability was significantly worse, with nearly 2 times the LOAs as well as a statistically significant difference in mean values between the first and second halves of the scan, possibly due to respiratory drift. The inconsistent mean values of Sup60 suggest compromised reliability of T1 values for that method, rather than the more benign and correctable systematic bias between methods seen in the 60‐s experiments. This discrepancy may again partially reflect the domain sensitivity of supervised learning, as the Sup60 method was not trained to handle 30‐s inputs. The superior T1 repeatability of the SSL method over supervised learning was even more pronounced in reconstructions before the DC layer, as shown in Figure S1. This suggests that the SSL method inherently enforces T1 repeatability, whereas supervised methods rely more heavily on DC layers to promote consistency. Overall, the contrasting performance of Sup60 versus Sup30/30—sharpness versus repeatability—highlights trade‐offs in supervised learning that were not evident for SSL.

Of the three deep learning methods, SSL was the only one that provided images that were both sharp and repeatable. It did so in the same inference time as the supervised methods, and without needing and reconstructed labels for training. This alleviates the burden of generating iterative reconstruction labels, which are time‐consuming and not a ground truth.

The success of CDC loss in translating to T1 repeatability, along with the consistent trends observed between the CDC metric (Figure 6) and T1 value repeatability (Figures 8 and S1), suggests that the CDC metric is a viable surrogate for quantitative repeatability. CDC loss is simple to evaluate during the training and validation phases and avoids potentially complicated and time‐consuming T1/T2 fitting during the learning process. A 14% improvement in CDC corresponded to a T1 repeatability improvement of 46% for the SSL method compared with the Sup60 method, which suggests that even a modest enhancement in CDC can lead to substantial gains in quantitative repeatability.

We selected the U‐Net as our network structure, because it is widely used in the field of deep‐learning MR reconstruction. However, more advanced network structures, such as Vision Transformers, 53 , 54 could also be readily used with our framework, potentially improving the overall reconstruction quality. In this study, both SSL and supervised learning used a single‐pass network structure and did not train with the DC layer, to simplify implementation and comparison. Nevertheless, we anticipate that our framework would also be compatible with end‐to‐end training and unrolled models, 21 , 22 which may further boost the reconstruction quality. Ablation studies are a potential direction for our future work, which includes testing the optimal number of unrolled iterations in an end‐to‐end training setting, as well as fine‐tuning our loss functions. In our experiments, NUFFT was used to process non‐Cartesian k‐space data. However, the rapid Toeplitz method 25 , 55 presents a potential alternative to NUFFT, with the advantage of reducing reconstruction time for both iterative reconstruction and the proposed SSL reconstruction. It is important to note that even with the adoption of the Toeplitz method, the proposed approach would remain significantly faster than iterative reconstruction due to its substantially reduced number of CG‐DC iterations.

Although our study demonstrates promising results, there are several limitations to acknowledge. First, the test–retest repeatability was assessed without any time gap, by splitting the 60‐s scan into two halves. This approach represents a relatively controlled condition, easier than assessing reproducibility, which would involve changes in experimental conditions such as timing or location. Extending this concept to longer time gaps could be valuable, although it may be challenging to manage significant misregistration from subjects entering or exiting the scanner or moving to a different scanner. Second, we did not investigate other splitting ratios beyond 50/50. Additionally, our study did not focus on patient populations, which could present different challenges and considerations. Finally, while we evaluated the proposed SSL method within the MR Multitasking framework, it may also be applicable to other dynamic quantitative MRI frameworks. However, further investigation would be needed to confirm its generalizability.

6. CONCLUSIONS

In this study, we developed a cross‐data‐consistent SSL reconstruction method to enhance the repeatability of quantitative MR image reconstruction. Our method introduced a CDC loss function and evaluation metric that were effective surrogates for T1 repeatability when evaluated in the context of cardiac MR Multitasking. This CDC SSL method demonstrated superior repeatability compared with supervised learning, and unlike supervised learning, achieved this without sacrificing image and map sharpness. The combination of improved repeatability, reduced reconstruction time versus iterative methods, and the elimination of the need for labeled training data highlights the potential of SSL to advance quantitative MRI techniques in both research and clinical settings.

Supporting information

Figure S1. Bland–Altman plots of the myocardial septal T1 values for 30‐s scans from different reconstruction methods. (A) All reconstructions are without conjugate‐gradient (CG)–data‐consistency (DC) layers. (B) All reconstructions are with CG‐DC layers. Note that (B) is essentially the same as Figure 8 but with y‐axis limit rescaled.

Figure S2. The U‐Net network structure used in our work. The number above each layer shows its number of channels; “IN” denotes the instance normalization.

Figure S3. Without the data‐consistency (DC) layers, the reconstruction images and T1 maps of 30‐s scan from supervised learning methods and the proposed self‐supervised learning (SSL) method. The subject is the same as the one in Figure 7A.

MRM-94-797-s001.docx (1.3MB, docx)

ACKNOWLEDGMENTS

This work was partially supported by National Institutes of Health (NIH) award R01EB028146.

Chen Z, Hu Z, Xie Y, Li D, Christodoulou AG. Repeatability‐encouraging self‐supervised learning reconstruction for quantitative MRI . Magn Reson Med. 2025;94:797‐809. doi: 10.1002/mrm.30478

Correction added after online publication 22 March 2025. Due to a publisher's error, reference 25 has been updated.

REFERENCES

  • 1. Granziera C, Wuerfel J, Barkhof F, et al. Quantitative magnetic resonance imaging towards clinical application in multiple sclerosis. Brain. 2021;144:1296‐1311. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Moura Cunha G, Navin PJ, Fowler KJ, Venkatesh SK, Ehman RL, Sirlin CB. Quantitative magnetic resonance imaging for chronic liver disease. Br J Radiol. 2021;94:20201377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Sheagren CD, Cao T, Patel JH, et al. Motion‐compensated T1 mapping in cardiovascular magnetic resonance imaging: a technical review. Front Cardiovasc Med. 2023;10:1160183. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Ma D, Gulani V, Seiberlich N, et al. Magnetic resonance fingerprinting. Nature. 2013;495:187‐192. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Christodoulou AG, Shaw JL, Nguyen C, et al. Magnetic resonance multitasking for motion‐resolved quantitative cardiovascular imaging. Nat Biomed Eng. 2018;2:215‐226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Shaw JL, Yang Q, Zhou Z, et al. Free‐breathing, non‐ECG, continuous myocardial T1 mapping with cardiovascular magnetic resonance multitasking. Magn Reson Med. 2019;81:2450‐2463. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Schieda N, Lim CS, Zabihollahy F, et al. Quantitative prostate MRI. J Magn Reson Imaging. 2021;53:1632‐1645. [DOI] [PubMed] [Google Scholar]
  • 8. Seraphim A, Knott KD, Augusto J, Bhuva AN, Manisty C, Moon JC. Quantitative cardiac MRI. J Magn Reson Imaging. 2020;51:693‐711. [DOI] [PubMed] [Google Scholar]
  • 9. Seiler A, Nöth U, Hok P, et al. Multiparametric quantitative MRI in neurological diseases. Front Neurol. 2021;12:640239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Bane O, Hectors SJ, Wagner M, et al. Accuracy, repeatability, and interplatform reproducibility of T1 quantification methods used for DCE‐MRI: results from a multicenter phantom study. Magn Reson Med. 2018;79:2564‐2575. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Sorace AG, Wu C, Barnes SL, et al. Repeatability, reproducibility, and accuracy of quantitative MRI of the breast in the community radiology setting. J Magn Reson Imaging. 2018;48:695‐707. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Feng L, Liu F, Soultanidis G, et al. Magnetization‐prepared GRASP MRI for rapid 3D T1 mapping and fat/water‐separated T1 mapping. Magn Reson Med. 2021;86:97‐114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Feng L, Wen Q, Huang C, Tong A, Liu F, Chandarana H. GRASP‐pro: imProving GRASP DCE‐MRI through self‐calibrating subspace‐modeling and contrast phase automation. Magn Reson Med. 2020;83:94‐108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Ong F, Zhu X, Cheng JY, et al. Extreme MRI: large‐scale volumetric dynamic imaging from continuous non‐gated acquisitions. Magn Reson Med. 2020;84:1763‐1780. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Liang Z‐P. Spatiotemporal imaging with partially separable functions. Proc IEEE Int Symp Biomed Imaging. IEEE; 2007:988‐991. [Google Scholar]
  • 16. Trzasko J, Manduca A, Borisch E. Local versus global low‐rank promotion in dynamic MRI series reconstruction. In: Proceedings of the 19th Annual Meeting of ISMRM. ISMRM; 2011:4371. [Google Scholar]
  • 17. Zhang T, Pauly JM, Levesque IR. Accelerating parameter mapping with a locally low rank constraint. Magn Reson Med. 2015;73:655‐661. doi: 10.1002/mrm.25161 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Lustig M, Donoho D, Pauly JM. Sparse MRI: the application of compressed sensing for rapid MR imaging. Magn Reson Med. 2007;58:1182‐1195. [DOI] [PubMed] [Google Scholar]
  • 19. Zhao B, Setsompop K, Adalsteinsson E, et al. Improved magnetic resonance fingerprinting reconstruction with low‐rank and subspace modeling. Magn Reson Med. 2018;79:933‐942. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Assländer J, Cloos MA, Knoll F, Sodickson DK, Hennig J, Lattanzi R. Low rank alternating direction method of multipliers reconstruction for MR fingerprinting. Magn Reson Med. 2018;79:83‐96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Aggarwal HK, Mani MP, Jacob M. MoDL: model‐based deep learning architecture for inverse problems. IEEE Trans Med Imaging. 2018;38:394‐405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Schlemper J, Caballero J, Hajnal JV, Price AN, Rueckert D. A deep cascade of convolutional neural networks for dynamic MR image reconstruction. IEEE Trans Med Imaging. 2017;37:491‐503. [DOI] [PubMed] [Google Scholar]
  • 23. Wang S, Su Z, Ying L, et al. Accelerating magnetic resonance imaging via deep learning. Proc IEEE Int Symp Biomed Imaging. IEEE; 2016:514‐517. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Yang G, Yu S, Dong H, et al. DAGAN: deep de‐aliasing generative adversarial networks for fast compressed sensing MRI reconstruction. IEEE Trans Med Imaging. 2017;37:1310‐1321. [DOI] [PubMed] [Google Scholar]
  • 25. Chen Z, Chen Y, Xie Y, Li D, Christodoulou AG. Data‐consistent non‐Cartesian deep subspace learning for efficient dynamic MR image reconstruction. Proc IEEE Int Symp Biomed Imaging; 2022. doi: 10.1109/isbi52829.2022.9761497 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Hammernik K, Küstner T, Rueckert D. Machine learning for MRI reconstruction. Advances in Magnetic Resonance Technology and Applications. Elsevier; 2022:281‐323. [Google Scholar]
  • 27. Hammernik K, Küstner T, Yaman B, et al. Physics‐driven deep learning for computational magnetic resonance imaging: combining physics and machine learning for improved medical imaging. IEEE Signal Process Mag. 2023;40:98‐114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Knoll F, Hammernik K, Zhang C, et al. Deep‐learning methods for parallel magnetic resonance imaging reconstruction: a survey of the current approaches, trends, and issues. IEEE Signal Process Mag. 2020;37:128‐140. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Feng L, Ma D, Liu F. Rapid MR relaxometry using deep learning: an overview of current techniques and emerging trends. NMR Biomed. 2022;35:e4416. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Huang P, Zhang C, Zhang X, Li X, Dong L, Ying L. Self‐supervised deep unrolled reconstruction using regularization by denoising. IEEE Trans Med Imaging. 2024;43:1203‐1213. doi: 10.1109/TMI.2023.3332614 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Jun Y, Arefeen Y, Cho J, et al. Zero‐DeepSub: zero‐shot deep subspace reconstruction for rapid multiparametric quantitative MRI using 3D‐QALAS. Magn Reson Med. 2024;91:2459‐2482. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Millard C, Chiew M. A theoretical framework for self‐supervised MR image reconstruction using sub‐sampling via variable density Noisier2Noise. IEEE Trans Comput Imaging. 2023;9:707‐720. doi: 10.1109/TCI.2023.3299212 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Miller Z, Johnson KM. Motion compensated self supervised deep learning for highly accelerated 3D ultrashort echo time pulmonary MRI. Magn Reson Med. 2023;89:2361‐2375. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Wang S, Wu R, Jia S, et al. Knowledge‐driven deep learning for fast MR imaging: undersampled MR image reconstruction from supervised to un‐supervised learning. Magn Reson Med. 2024;92:496‐518. [DOI] [PubMed] [Google Scholar]
  • 35. Yaman B, Hosseini SAH, Moeller S, Ellermann J, Uğurbil K, Akçakaya M. Self‐supervised learning of physics‐guided reconstruction neural networks without fully sampled reference data. Magn Reson Med. 2020;84:3172‐3191. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Zhang M, Xu J, Arefeen Y, Adalsteinsson E. Zero‐shot self‐supervised joint temporal image and sensitivity map reconstruction via linear latent space. Proceedings of Machine Learning Research.JMLR; 2024:1713‐1725. [Google Scholar]
  • 37. Zhou B, Schlemper J, Dey N, et al. Dual‐domain self‐supervised learning for accelerated non‐cartesian MRI reconstruction. Med Image Anal. 2022;81:102538. [DOI] [PubMed] [Google Scholar]
  • 38. Bian W, Jang A, Liu F. Improving quantitative MRI using self‐supervised deep learning with model reinforcement: demonstration for rapid T1 mapping. Magn Reson Med. 2024;92:98‐111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Kofler A, Zimmermann FF, Papafitsoros K. Machine learning for quantitative MR image reconstruction. arXiv. 240219396 2024.
  • 40. Zimmermann FF, Kolbitsch C, Schuenke P, Kofler A. PINQI: an end‐to‐end physics‐informed approach to learned quantitative MRI reconstruction. IEEE Trans Comput Imaging. 2024;10:628‐639. doi: 10.1109/TCI.2024.3388869 [DOI] [Google Scholar]
  • 41. Xu S, Früh M, Hammernik K, Gatidis S, Küstner T. Self‐supervised contrastive learning for MR image reconstruction of cardiac CINE on accelerated cohorts. In: Proceedings of the 32nd Annual Meeting of ISMRM, Toronto, Canada, 2023.
  • 42. Desai AD, Ozturkler BM, Sandino CM, et al. Noise2Recon: enabling SNR‐robust MRI reconstruction with semi‐supervised and self‐supervised learning. Magn Reson Med. 2023;90:2052‐2070. [DOI] [PubMed] [Google Scholar]
  • 43. Huang C, Graff CG, Clarkson EW, Bilgin A, Altbach MI. T2 mapping from highly undersampled data by reconstruction of principal component coefficient maps using compressed sensing. Magn Reson Med. 2012;67:1355‐1366. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Fessler JA. On NUFFT‐based gridding for non‐cartesian MRI. J Magn Reson. 2007;188:191‐195. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Boyd S, Parikh N, Chu E, Peleato B, Eckstein J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found Trends Mach Learn. 2011;3:1‐122. [Google Scholar]
  • 46. Beck A, Teboulle M. A fast iterative shrinkage‐thresholding algorithm for linear inverse problems. SIAM J Imaging Sci. 2009;2:183‐202. [Google Scholar]
  • 47. Acar M, Çukur T, Öksüz İ. Self‐supervised dynamic MRI reconstruction. In: Proceedings of the 4th International Workshop on Machine Learning for Medical Image Reconstruction, Strasbourg, France, 2021.
  • 48. Gan W, Ying C, Boroojeni PE, et al. Self‐supervised deep equilibrium models with theoretical guarantees and applications to MRI reconstruction. IEEE Trans Comput Imaging. 2023;9:796‐807. doi: 10.1109/TCI.2023.3304475 [DOI] [Google Scholar]
  • 49. Hamilton JI. A self‐supervised deep learning reconstruction for shortening the breathhold and acquisition window in cardiac magnetic resonance fingerprinting. Front Cardiovasc Med. 2022;9:928546. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Lozenski L, Anastasio MA, Villa U. A memory‐efficient self‐supervised dynamic image reconstruction method using neural fields. IEEE Trans Comput Imaging. 2022;8:879‐892. [Google Scholar]
  • 51. Ronneberger O, Fischer P, Brox T. U‐Net: Convolutional Networks for Biomedical Image Segmentation. Springer; 2015:234‐241. [Google Scholar]
  • 52. Roujol S, Weingärtner S, Foppa M, et al. Accuracy, precision, and reproducibility of four T1 mapping sequences: a head‐to‐head comparison of MOLLI, ShMOLLI, SASHA, and SAPPHIRE. Radiology. 2014;272:683‐689. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Cao H, Wang Y, Chen J, et al. Swin‐Unet: Unet‐like pure transformer for medical image segmentation. In: Proceedings of the European Conference on Computer Vision, 2022.
  • 54. Dosovitskiy A, Beyer L, Kolesnikov A, et al. An image is worth 16 × 16 words: transformers for image recognition at scale. arXiv. 201011929 2020.
  • 55. Baron CA, Dwork N, Pauly JM, Nishimura DG. Rapid compressed sensing reconstruction of 3D non‐Cartesian MRI. Magn Reson Med. 2018;79:2685‐2692. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Figure S1. Bland–Altman plots of the myocardial septal T1 values for 30‐s scans from different reconstruction methods. (A) All reconstructions are without conjugate‐gradient (CG)–data‐consistency (DC) layers. (B) All reconstructions are with CG‐DC layers. Note that (B) is essentially the same as Figure 8 but with y‐axis limit rescaled.

Figure S2. The U‐Net network structure used in our work. The number above each layer shows its number of channels; “IN” denotes the instance normalization.

Figure S3. Without the data‐consistency (DC) layers, the reconstruction images and T1 maps of 30‐s scan from supervised learning methods and the proposed self‐supervised learning (SSL) method. The subject is the same as the one in Figure 7A.

MRM-94-797-s001.docx (1.3MB, docx)

Articles from Magnetic Resonance in Medicine are provided here courtesy of Wiley

RESOURCES