Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2025 Aug 13.
Published in final edited form as: Proc Int Soc Magn Reson Med Sci Meet Exhib Int Soc Magn Reson Med Sci Meet Exhib. 2025 May;33:2742.

Contrastive Learning for Accelerated MR Fingerprinting

Peizhou Huang 1, Brendan L Eck 2, Mingrui Yang 2, Ruiying Liu 3, Xiaoliang Zhang 1, Xiaojuan Li 2, Leslie Ying 1,3
PMCID: PMC12346756  NIHMSID: NIHMS2098965  PMID: 40810093

INTRODUCTION

Magnetic Resonance Fingerprinting (MRF) is a powerful quantitative MRI technique that allows simultaneous estimation of multiple tissue parameters [1]. However, traditional MRF reconstruction methods rely on dictionary matching, which is computationally intensive and noise-sensitive, limiting its scalability for highly undersampled data [27]. Although recent methods [812] have been proposed to mitigate these issues, practical applications of MRF remain constrained. This study proposes CLIP-MRF, a novel approach leveraging contrastive learning for accelerated, robust MRF reconstruction.

METHODS

Inspired by the Contrastive Language-Image Pre-training (CLIP) [13], we proposed a dual-encoder network that maximizes signal similarity in a learnable space for MRF reconstruction. CLIP is a model for learning associations between images and their corresponding textual descriptions by comparing their embeddings directly. This concept shares common with that of MRF pattern matching. We propose to learn the association between measured signal evolutions and known fingerprint in a dictionary with CLIP. With that, the MRF pattern matching can be replaced by the CLIP network for more accurate quantifications. This is because CLIP compares the features instead of the entire time evolution of measure signals and fingerprints, and thus is more robust than the conventional MRF pattern matching.

The schematic overview of our proposed CLIP-MRF model is shown in Fig. 1(a)(b). The model contains two networks, a signal encoder and a fingerprint encoder respectively. The signal encoder extracts the feature of the measured MRF signal evolutions denoted as, and the fingerprint encoder maps all known fingerprints to a low-dimensional space denoted as. If the features of the two encoders match, then the measured signal represents the corresponding quantitative values of the tissue properties. The blue diagonal denotes the matched pairs whose inner products are maximized, and the inner products of the unmatched pairs are minimized during the training. The right of Fig.1 shows the detailed network architecture.

Figure 1.

Figure 1.

Schematic overview of our proposed CLIP-MRF model. The training (a) and the inference process (b) are shown on the left. The network structures of the signal encoder (c) and fingerprint encoder (d) are shown on the right.

Given N input pairs of measured signal evolution and fingerprint of a certain tissue parameter, the signal encoder and fingerprint encoder will be jointly trained to maximize the cosine similarity of the N true pairs in the batch while minimizing that of the rest. This similarity calculation is almost the same as the one in the DM method.

RESULTS & DISCUSSION

The MRF sequence uses a 3D Cartesian trajectory with linear readout in kx and a variable density circular Cartesian undersampling pattern in ky-kz. The MRF acquisition consists of 500 time frames acquired using FISP with each frame acquired at an acceleration factor (AF) of 15. The acquisition matrix size is 96×96×24. Five digital phantoms based on in vivo data from five volunteers under an IRB-approved protocol were used. The in vivo data consist of T1 and M0 maps obtained from 2D multislice phase-sensitive inversion recovery and T2 maps obtained from 3D MAPSS. Quantitative tissue parameter maps were used as input reference values for the generation of MRF data.

To train the CLIP-MRF model, we put the generated MRF data of a pixel into the signal encoder and the fingerprint that corresponds to the T1 and T2 values of this pixel into the fingerprint encoder. The similarity between the features extracted from the two encoders was maximized. Four out of five datasets were used for training and one for testing. The reconstructed maps were compared with conventional dictionary matching (DM) [1], SVD-compressed dictionary matching (LR-DM) [14], and conventional deep-learning-based method (DL) [8]. The same generated MRF data were used to train the DL network. Fig.2 shows the quantification results for a representative sagittal slice. The proposed CLIP-MRF is superior to DM and DL, and comparable to LR-DM but with a significantly less reconstruction time. To demonstrate that image contents have little effect on training, we also trained CLIP-MRF on a synthetic image whose quantification values were randomly assigned for each pixel and tested on the same digital phantom. The results are shown in Fig.3. And Table 1 shows the relative error of all testing data. The same conclusion can be made as in Fig.2, which suggests no need for training datasets with contents similar to the testing ones. We also acquired in vivo MRF data using the same scanner and the same MRF sequence with an AF of 15, and the results are shown in Fig.4. Since there is no ground truth, LR-DM was used as the reference. The CLIP-MRF is the closest to LR-DM in both T1 and T2 maps.

Figure 2.

Figure 2.

Comparison of the proposed CLIP-MRF with other networks when trained on the digital phantoms. From left to right: (a) ground truth (b) DM, (c) DL, (d) LR-DM, and (e) proposed CLIP-MRF. A vertical colorbar shows the quantitative values. The nMSE and relative error are shown at the bottom right corner.

Figure 3.

Figure 3.

Comparison of the proposed CLIP-MRF with other networks when trained on the synthetic image. From left to right: (a) ground truth (b) DM, (c) DL, (d) LR-DM, and (e) proposed CLIP-MRF. A vertical colorbar shows the quantitative values. The nMSE and relative error are shown at the bottom right corner.

TABLE 1.

The relative errors for T1 and T2 of the reconstruction subject (expressed as mean ± standard deviation, in percentage), defined as the ratio between the Frobenius norm of the difference map and that of the ground truth map.

Method T1 T2
DM 16.43±5.19 21.92±3.28
LR-DM 16.92±8.43 7.58±2.58
DL 12.29±5.24 15.79±5.84
CLIP-MRF 8.40±4.22 11.98±4.45

Figure 4.

Figure 4.

Comparison of the proposed CLIP-MRF with other networks for in vivo MRF data. From left to right: (a) DM, (b) LR-DM, (c) DL, and (d) proposed CLIP-MRF. A vertical colorbar shows the quantitative values. (Top row: T1 maps; Bottom row: T2 maps; units are ms)

CONCLUSION

CLIP-MRF shows improvement over conventional MRF methods. Future work will explore the extension of this framework to other quantitative parameters, such as T2* and B1 inhomogeneity, to further enhance its clinical utility.

Synopsis.

Keywords:

MR Fingerprinting, AI/ML Image Reconstruction, Contrastive Learning

Motivation:

MR Fingerprinting (MRF) enables simultaneous multi-parametric mapping but suffers from computational challenges and noise/artifact sensitivity due to dictionary matching.

Goal(s):

This study aims to develop a novel network called CLIP-MRF to improve pattern matching in MRF. It incorporates contrastive learning to enhance quantification accuracy in accelerated MRF.

Approach:

We propose a dual-encoder contrastive training method to robustly map MRF signals to tissue parameters accurately. The model maximizes similarity between matching signal-parameter pairs and minimizes mismatched ones during training.

Results:

CLIP-MRF demonstrates superior performance over the state-of-the-art MRF methods in T1 and T2 quantification, reducing the reconstruction time and errors.

Impact:

The CLIP-MRF network enables accurate parameter mapping and improves computational efficiency for accelerated MRF. Trained on simulated data only, the network offers robust generalization across signals with different noise/artifacts, paving the way for fast and reliable tissue quantification.

Acknowledgements

This work was supported in part by the NIH/NIAMS R01 AR077452, NIH/NIA K25 AG070321 and NIH under a BRP grant U01 EB023829.

References

  • [1].Ma D et al. , “Magnetic resonance fingerprinting,” Nature, vol. 495, no. 7440, pp. 187–192, Mar. 2013, doi: 10.1038/nature11971. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Davies M, Puy G, Vandergheynst P, and Wiaux Y, “A Compressed Sensing Framework for Magnetic Resonance Fingerprinting,” SIAM Journal on Imaging Sciences, vol. 7, no. 4, pp. 2623–2656, 2014, doi: 10.1137/130947246. [DOI] [Google Scholar]
  • [3].Zhao B, Lam F, Bilgic B, Ye H, and Setsompop K, “Maximum likelihood reconstruction for magnetic resonance fingerprinting,” in 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), Apr. 2015, pp. 905–909. doi: 10.1109/ISBI.2015.7164017. [DOI] [Google Scholar]
  • [4].Wang Z, Li H, Zhang Q, Yuan J, and Wang X, “Magnetic Resonance Fingerprinting with compressed sensing and distance metric learning,” ArXiv, vol. abs/2209.08734, 2016, [Online]. Available: https://api.semanticscholar.org/CorpusID:10230819 [Google Scholar]
  • [5].McGivney DF et al. , “SVD Compression for Magnetic Resonance Fingerprinting in the Time Domain,” IEEE Transactions on Medical Imaging, vol. 33, no. 12, pp. 2311–2322, Dec. 2014, doi: 10.1109/TMI.2014.2337321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Doneva M, Amthor T, Koken P, Sommer K, and Börnert P, “Matrix completion-based reconstruction for undersampled magnetic resonance fingerprinting data,” Magn Reson Imaging, vol. 41, pp. 41–52, Sep. 2017, doi: 10.1016/j.mri.2017.02.007. [DOI] [PubMed] [Google Scholar]
  • [7].Hu Y, Li P, Chen H, Zou L, and Wang H, “High-Quality MR Fingerprinting Reconstruction Using Structured Low-Rank Matrix Completion and Subspace Projection,” IEEE Transactions on Medical Imaging, vol. 41, no. 5, pp. 1150–1164, May 2022, doi: 10.1109/TMI.2021.3133329. [DOI] [PubMed] [Google Scholar]
  • [8].Cohen O, Zhu B, and Rosen MS, “MR fingerprinting Deep RecOnstruction NEtwork (DRONE),” Magnetic Resonance in Medicine, vol. 80, no. 3, pp. 885–894, 2018, doi: 10.1002/mrm.27198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Fang Z et al. , “Deep Learning for Fast and Spatially Constrained Tissue Quantification From Highly Accelerated Data in Magnetic Resonance Fingerprinting,” IEEE Transactions on Medical Imaging, vol. 38, no. 10, pp. 2364–2374, Oct. 2019, doi: 10.1109/TMI.2019.2899328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Hamilton JI and Seiberlich N, “Machine Learning for Rapid Magnetic Resonance Fingerprinting Tissue Property Quantification,” Proceedings of the IEEE, vol. 108, no. 1, pp. 69–85, Jan. 2020, doi: 10.1109/JPROC.2019.2936998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [11].Soyak R et al. , “Channel Attention Networks for Robust MR Fingerprint Matching,” IEEE Transactions on Biomedical Engineering, vol. 69, no. 4, pp. 1398–1405, Apr. 2022, doi: 10.1109/TBME.2021.3116877. [DOI] [PubMed] [Google Scholar]
  • [12].Singh M, Jiang S, Li Y, van Zijl P, Zhou J, and Heo H-Y, “Bloch simulator–driven deep recurrent neural network for magnetization transfer contrast MR fingerprinting and CEST imaging,” Magnetic Resonance in Medicine, vol. 90, no. 4, pp. 1518–1536, 2023, doi: 10.1002/mrm.29748. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Radford A et al. , “Learning Transferable Visual Models From Natural Language Supervision,” in Proceedings of the 38th International Conference on Machine Learning, Meila M and Zhang T, Eds., in Proceedings of Machine Learning Research, vol. 139. PMLR, Jul. 2021, pp. 8748–8763. [Online]. Available: https://proceedings.mlr.press/v139/radford21a.html [Google Scholar]
  • [14].Yang M et al. , “Low rank approximation methods for MR fingerprinting with large scale dictionaries,” Magn Reson Med, vol. 79, no. 4, pp. 2392–2400, Apr. 2018, doi: 10.1002/mrm.26867. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Dosovitskiy A et al. , “An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale,” in International Conference on Learning Representations, 2021. [Online]. Available: https://openreview.net/forum?id=YicbFdNTTy [Google Scholar]

RESOURCES