Abstract
Multiparametric magnetic resonance imaging (mpMRI) data are emergingly used in the clinic e.g. for the diagnosis of prostate cancer. In contrast to conventional MR imaging data, multiparametric data typically include functional measurements such as diffusion and perfusion imaging sequences. Conventionally, these measurements are visualized with a one-dimensional color scale, allowing only for one-dimensional information to be encoded. Yet, human perception places visual information in a three-dimensional color space. In theory, each dimension of this space can be utilized to encode visual information. We addressed this issue and developed a new method for tri-variate color-coded visualization of mpMRI data sets. We showed the usefulness of our method in a preclinical and in a clinical setting: In imaging data of a rat model of acute kidney injury, the method yielded characteristic visual patterns. In a clinical data set of N = 13 prostate cancer mpMRI data, we assessed diagnostic performance in a blinded study with N = 5 observers. Compared to conventional radiological evaluation, color-coded visualization was comparable in terms of positive and negative predictive values. Thus, we showed that human observers can successfully make use of the novel method. This method can be broadly applied to visualize different types of multivariate MRI data.
Multiparametric magnetic resonance imaging (mpMRI) is used to measure several physical parameters for each spatial location in a patient. In the clinic, this technology has important applications: for example, it is used to diagnose prostate cancer according to the PI-RADS criteria1. Usually, these physical parameters are visualized as gray scale images or one-dimensional color coded images and human observers visually evaluate these images successively or side-by-side2. Because this procedure yields many independent images, examining multi-channel datasets can be very laborious. For example, to read a typical MRI data set of the prostate according to the current PI-RADS criteria approximately takes 20 minutes3.
Grayscale visualization is the most common way of displaying radiological measurements in clinical routine. This way of visualizing data disregards color. However, humans have evolved in a colorful environment and color is an extremely important carrier of information in many aspects of life4. Not surprisingly, color is a highly efficient carrier of information and humans can quickly spot visual targets based on color5,6. Color processing in the human brain is a complex process that involves many different parts of the nervous system7. This machinery enables humans to distinguish millions of colors8. In contrast, humans can only distinguish approximately 720 different gray shades on typical medical displays2.
The aim of data visualization is to make objective information accessible to the human mind9. Grayscale intensity coding is one possible carrier for this information – but the above-mentioned studies show that color is potentially a much better carrier of information. In the context of imaging, it follows that human observers can potentially extract more information from color-coded images than from grayscale images. This increase of transmittable visual information constitutes an information gain. This idea is the basis for a new visualization method for mpMRI data that we present in this paper.
MpMRI data sets consist of several spatially aligned channels. By using a color-coded representation, this information can theoretically be merged into one image. A commonly used approach to combine several image channels in one image is to overlay a functional measurement on top of a structural CT or MRI image10. However, this approach cannot be used to visualize more than two channels at once. A simple way of visualizing more than two channels is to use the RGB color space that is used in computer displays: to encode tri-variate imaging data, each parameter map is assigned to one of the basic display colors red, green, blue. However, this method leads to distortions of the data and is prone to introduce artifacts, caused by the pronounced nonlinearity of the human perception of colors in the RGB color space11,12.
For computed tomography (CT) imaging, several studies have assessed the use of color transfer functions and their impact on observer performance13,14. For MRI data, published approaches include 3D visualization13, glyph-based methods15 and color coded visualization of temporal differences in dynamic contrast-enhanced (DCE) MRI breast imaging data16. Yet, to our knowledge, there is no color-coded visualization technique available that can be readily applied to three-channel, spatially aligned mpMRI data sets.
In this study, we present a possible solution to these limitations of previous visualization methods. We propose a new method for simultaneous color-based visualization of up to three spatially aligned image channels. First, we describe the development of a tri-variate color map that is adapted to the human visual system. Second, we explore how this color map can be used to process routine MRI data sets. Third, we assess diagnostic performance of blinded observers in a typical diagnostic setting in the clinic.
Material and Methods
Reproducibility and ethics statement
All source codes used for this study are available under the MIT license (http://opensource.org/licenses/MIT) and can be accessed via the following DOI: 10.5281/zenodo.208185. All experimental measurements can be accessed as a raw data via this DOI: 10.5281/zenodo.205081. Anonymized DICOM datasets of all samples shown in the figures can be accessed via this DOI: 10.5281/zenodo.205089. All experiments were carried out in accordance with the Declaration of Helsinki and were approved by the institutional ethics board (2013-824R-MA). The institutional ethics board waived the need for informed consent for this retrospective analysis of fully anonymized samples. Animal data were retrieved from a previously published study17. All animal procedures were performed according to the Guide of the Care and Use of Laboratory Animals published by the National Academy of Sciences and were approved by the local authorities (Regional council Karlsruhe, G40/10).
Data sets
As a preclinical example, we used a data set of perfusion MRI data of experimental kidney injury in rats as described before17. Briefly, measurements were performed on a 3 T scanner (Tim trio, Siemens Healthcare Sector, Erlangen, Germany) using an eight channel receive-only volumetric rat array (RAPID Biomedical GmbH, Rimpar, Germany) for signal detection. DCE-MRI was performed using a 3D time-resolved angiography with stochastic trajectories (TWIST) sequence (16) with the following parameters: TR/TE/FA = 3.4 ms/1.4 ms/20°, matrix = 192 × 84, FOV = 114 × 50 mm2, a GRAPPA factor of 2 and 28 slices. The nominal temporal resolution was 0.9 s per volume. Images were continuously acquired for 6 minutes resulting in 400 volumes. After the 15th volume, 0.05 ml of contrast agent (Dotarem, Guerbet, France) was manually administered in the femoral vein, followed by a 1.0 ml saline flush. Of the acquired data the parametric maps of the Plasma flow, Plasma Volume and Tubular flow were calculated using an in house certified OsiriX plugin (UMMperfusion 1.5.2)18,19.
As a clinically relevant example, we used mpMRI measurements of prostate cancer patients as described by Weidner et al.3. Briefly, the prostate multi-parametric data sets consist of a T2-weighted image, apparent diffusion coefficient (ADC) map, a diffusion weighted image with b = 800, and prostate blood flow (PBF) map. Imaging was performed at a 1.5 T scanner (Magnetom Avanto, Siemens Healthcare, Erlangen, Germany) with parameters are given in Table 1. ADC maps were reconstructed inline at the scanner while PBF maps of the prostate were calculated using the above mentioned perfusion plugin. Clinico-pathological characteristics of the sample collective are given in Table 2.
Table 1. MRI parameters for prostate imaging.
T2 axial | DWI | DCE | |
---|---|---|---|
Sequence | TSE | EPI DWI | 2D FLASH |
TR (ms) | 6360 | 1300 | 193 |
TE (ms) | 108 | 67 | 1.05 |
FOV (mm2) | 160 × 160 | 200 × 200 | 300 × 225 |
Voxel size (mm3) | 0.7 × 0.6 × 3 | 2.2 × 2.2 × 4 | 2.1 × 1.6 × 6 |
Matrix | 256 × 230 | 90 × 90 | 192 × 144 |
PAT factor | 2 | — | — |
Number of slices | 22 | 10 | 4–6 |
B-value | — | 0,50,150,300,600,800 | — |
No. of measurements | 1 | 1 | 50 |
Temporal resolution (sec) | — | — | 4 |
DWI = diffusion weighted imaging, DCE = dynamic contrast enhanced, FLASH = fast low angle shot, TE = echo time, TR = repetition time, TSE = turbo spin echo, PAT = parallel imaging factor.
Table 2. Clinico-pathological characteristics of the prostate sample collective.
Pseudonym | Age | Gleason | TNM, L, V, R |
---|---|---|---|
112 | 69 | 4 + 4 = 8 | pT3b pNx pMx L1 V1 R0 |
121 | 63 | 4 + 5 = 9 | pT3b pN1 pMx L1 V1 R1 |
152 | 59 | 3 + 4 = 7 | pT2c pN0 pMx L0 V0 R0 |
159 | 72 | 4 + 3 = 7 | pT3b pNx pMx L0 V0 R0 |
186 | 52 | 3 + 3 = 6 | pT2c pNx pMx L0 V0 R0 |
187 | 59 | 3 + 4 = 7 | pT2c pNx pMx L0 V0 R1 |
225 | 56 | 4 + 3 = 7 | pT2c pN0 pMx L0 V0 R0 |
226 | 59 | 3 + 4 = 7 | pT2c pN0 pMx V0 L0 R0 |
227 | 67 | 3 + 3 = 6 | pT2c pNx pMx L0 V0 R0 |
234 | 69 | 3 + 4 = 7 | pT2b pN0 pMx L1 V0 R0 |
237 | 68 | 3 + 3 = 6 | pT2c pN0 pMx L0 V0 R0 |
238 | 73 | 3 + 4 = 7 | pT3a pNx pMx L0 V0 R0 |
260 | 58 | 3 + 4 = 7 | pT3a pN0 pMx L0 V0 R0 |
Computational implementation
All computational methods in this study were implemented in MATLAB (MATLAB R2015b, Mathworks, Natick, MA, USA) if not otherwise noted. They were run on a standard computer workstation (2.2 GHz Intel Core i7, 16 GB RAM). All visual experiments were performed on a Macbook Pro Retina Display (Apple, Cupertino, CA) using the Adobe RGB 1998 color profile. Colored MRI image were saved as DICOM files and were viewed with OsiriX Viewer (Pixmeo Sarl, Bernex, Switzerland).
Tri-variate color-based reconstruction
Color maps were constructed in a linear, perceptually uniform manner in CIELAB color space similarly to our previously published method20. By using the CIELAB color space, we ensured that Euclidean distances of signal intensities linearly corresponded to color differences in terms of human perception21. We constructed a three-dimensional color map with eight anchor points: 1st “no signal in any channel” (black), 2nd “maximum signal in all channels” (white), 3rd–5th “no signal in two channels, maximum signal in one channel” and 6th–8th “maximum signal in two channels, no signal in the third channel”. All other elements of the color map were linearly interpolated from these anchor points. The following constraints were applied: 1) Black (L = 0, A = 0, B = 0) and white (L = 100, A = 0, B = 0) should be the most extreme points on either sides (1st and 2nd anchor point). 2) The other anchor points (3rd to 8th) should be part of the L = 50 plane. 3) The volume of the polyhedron should be large in order to include many different colors. These constraints guaranteed optimal characteristics of the three-dimensional color map. It can be seen that the resulting color map is symmetric in CIELAB space (Fig. 1A–C). We used the following anchor points (colors given as RGB hexadecimal codes). 1st black (#000000), 2nd white (#FFFFFF), 3rd red (#F40000), 4th green (#009100), 5th blue (#1173FE), 6th magenta (#EB009C), 7th cyan (#008B8E), 8th dark orange (#A27200). This color map is shown in Fig. 1A–C, while Fig. 1D–I show characteristics of two alternative color maps that do not satisfy all of the above-mentioned criteria. However, in special circumstances, these alternative color maps might have other benefits. For example, the color map shown in Fig. 1G–I could be better suited for observers with deuteranomaly because it does not contain green hues.
The color map in Fig. 1A–C was used to encode tri-variate data. Initially, the assignment of image channels to anchor points of this color map was arbitrary – in other words, the color map could be arbitrarily rotated. However, it is known that reddish visual targets are easier to spots than targets of different colors. This holds true even for desaturated targets, showing that reddish hues can be efficiently used to encode visual cues that are of interest22. Therefore, we rotated the color map in such a way that in the prostate cancer data set, cancerous tissue regions mapped to reddish areas of the three-dimensional color map.
Data preprocessing and postprocessing
Before combining the images to a color-coded visualization, each channel was preprocessed as follows: (1) images were resampled to be spatially aligned to the T2 image (by using the native OsiriX resampling functionality), (2) all pixels with an intensity below the 1st quantile or above the 99th quantile were considered as outliers and were removed, (3) the channels were normalized to a range of 0 to 1, and (4) contrast within each channel was maximized by stretching the histogram so that 1% of the data was maximally saturated. After creating the merged colored image, the result was overlaid to the axial T2 image to allow better co-localization of anatomical structures. This procedure was applied to the prostate data set. For the rat kidney data set, all pixels below 2% or above 98% were removed. The rat data set was acquired with a clinical 3 Tesla system and dedicated coils and presented more noise and thus more outliers than the prostate data set, which was acquired with a clinical 1.5 Tesla MRI. Also, for the rat data, the following modifications were made to this procedure: no T2-overlay was created and the data were already spatially aligned in this application so that no resampling was necessary.
Measuring perceptual distance
After applying the new visualization procedure to prostate MRI data sets, we investigated whether tumorous areas took on different colors than non-tumorous areas. Based on the histopathological annotation, we manually delineated the largest contiguous tumor area and the largest contiguous non-tumor area in a representative axial image. We calculated the mean color of the tumor area in CIELAB space. Then, for each image pixel, we calculated the Euclidean distance to the mean tumor color in CIELAB space. Because distances in the CIELAB space are proportional to perceived color differences, this distance represented the perceptual difference (or perceptual contrast) for each pixel to the mean tumor color.
Observer study
In a previous study by Weidner et al.3, conventional expert reading of MR images (before total prostatectomy) was compared to systematic histopathological evaluation of these samples (after total prostatectomy). Primary end point was the detection of cancer in each segment of the prostate. In each prostate, 28 segments were defined as follows: 4 transversal planes were considered and each plane was divided in 8 sectors, except for the highest plane, which was only divided in 6 sectors (Fig. 2). In the present study, we used the same partitioning when presenting the color-coded visualizations to human observers. One expert observer (attending physician in radiology) and four radiological trainees (three residents within radiological training programs, one specialized medical intern) participated. Each observer evaluated each segment of the prostate and decided whether cancer was present or not. For each segment, this decision was compared to the histopathological gold standard and positive predictive value (PPV) and negative predictive value (NPV) were calculated. Before the actual experiment, each observer received a short training (approx. 10–15 minutes) and was explained the theory behind the visualization approach. Also, he or she was shown one example of a prostate mpMRI data set before and after the visualization procedure (this example was not included within the analyzed set).
Results
Application to perfusion MRI in a preclinical animal model of kidney injury
We applied the new visualization technique to a preclinical and a clinical data set. First, we chose a preclinical model of acute kidney injury in the rat and used three perfusion parameters: Tubular flow, plasma flow and plasma volume (Fig. 3). We found that in control animals, kidneys typically showed a very regular stratified pattern (Fig. 3A) that was disrupted after kidney injury (Fig. 3B). These changes were well visualized by a tri-variate color map while in conventional gray-scale images, this pattern was not as easily visualized because each parameter was shown in a separate image (Fig. 3D–I).
Color-coded images can be used to visualize prostate cancer in MRI data
Second, we applied the method to a clinically relevant case and used mpMRI measurements of perfusion and diffusion in patients with suspected prostate cancer (Fig. 4C). In these data sets, tumors are generally hyperintense in the b800 image, hypointense in the ADC map and hyperintense in the PBF map. Correspondingly, in the multivariate visualization, prostate tumors took on reddish hues while surrounding prostate tissue took on blue-greenish hues.
This color map was applied to mpMRI datasets of patients with suspected prostate cancer. In Fig. 5A, the result can be seen for a control patient with no evidence of prostate cancer. In Fig. 4B, a patient with histological evidence of prostate cancer is shown. The tumor lights up as a red patch whereas normal prostate tissue takes on blue-greenish hues or white. Comparable examples are shown in Fig. 6.
Perceptual contrast between prostate tumors and non-tumorous prostate areas
When comparing the histopathological annotation to the color-coded prostate data sets, it was evident that reddish tumorous areas could be discerned from the blue-greenish background. To quantify this subjective observation, we measured perceptual contrast between tumor and non-tumorous areas. As a reference point, we used the mean color of the tumor in CIELAB space. Then, in each of our N = 14 samples (including the training sample), we measured the perceptual difference from the reference point to all pixels in tumor and non-tumor areas. We found that the median perceptual distance from tumorous areas to the mean tumor color was lower than for non-tumorous areas. This was the case in all 14 samples, as can be seen in Fig. 7. These experiments show that it is theoretically possible to visually discern tumor vs. non-tumor based on the color. Next, we went on to validate this finding in a real-world setting.
Observer performance is comparable to conventional data reading
To quantify the diagnostic accuracy that can be achieved with the proposed method, we performed an observer study with N = 5 human observers at different stages of their training. We compared their performance (measured as PPV and NPV) against an expert reading of the conventional visualization and against a systematic histopathological workup of N = 13 cases. We found that PPV and NPV were comparable in all groups: For the conventional readings, we re-evaluated raw data from a previous study3 and found a PPV of 0.77 and NPV of 0.65 for the presence of prostate cancer in a given prostate segment. Using the new color-coded visualization, an expert (N = 1) achieved a PPV of 0.75 and an NPV of 0.65. The average diagnostic accuracy of radiological trainees (N = 4) was reflected in a PPV of 0.76 and an NPV of 0.61. Table 3 summarizes sensitivity, specificity, PPV and NPV for all observers and Table 4 contains the raw contingency data. In summary, all these values are comparable indicating that the novel color-coded visualization enables both expert and trainee readers to detect prostate cancer with sufficient accuracy.
Table 3. Sensitivity (Sens.), Specificity (Spec.), Positive/Negative Predictive Value (PPV/NPV) for each observer in our prostate cancer detection study.
Sens. | Spec. | PPV | NPV | |
---|---|---|---|---|
Radiologist expert gray | 0.36 | 0.92 | 0.77 | 0.65 |
Radiologist 1 color | 0.19 | 0.92 | 0.64 | 0.60 |
Radiologist 2 color | 0.26 | 0.90 | 0.66 | 0.61 |
Radiologist 3 color | 0.15 | 0.95 | 0.69 | 0.59 |
Radiologist 4 color | 0.17 | 0.96 | 0.75 | 0.60 |
Radiologists color mean | 0.21 | 0.95 | 0.76 | 0.61 |
Radiologist expert color | 0.36 | 0.91 | 0.75 | 0.65 |
Table 4. Contingency table for the presence of prostate cancer in each segment of the prostate, given for each observer in our study.
Pathologist pos. | Pathologist neg. | |
---|---|---|
Radiologist expert gray pos. | 64 | 19 |
Radiologist expert gray neg. | 116 | 217 |
Radiologist 1 color pos. | 35 | 20 |
Radiologist 1 color neg. | 145 | 216 |
Radiologist 2 color pos. | 47 | 24 |
Radiologist 2 color neg. | 133 | 212 |
Radiologist 4 color pos. | 27 | 12 |
Radiologist 4 color neg. | 153 | 224 |
Radiologist 5 color pos. | 30 | 10 |
Radiologist 5 color neg. | 150 | 226 |
Radio. mean color pos. | 37 | 12 |
Radio. mean color neg. | 143 | 224 |
Radio. expert color pos. | 64 | 21 |
Radio. expert color neg. | 116 | 215 |
Discussion
Benefits of the proposed method
Here, we have presented a new method for the color-coded visualization of mpMRI data sets. Conventionally, only a one-dimensional color space is used to map signal intensities to pixel intensities in MRI images. For the first time, our study extends this to a three-dimensional color space. In an environment of ever-growing amounts of data, there is a real need of new data analysis techniques. Our new approach addresses this need, could now be used to explore other advanced visualization techniques and could stimulate research in this field.
Applied in a preclinical and in a clinical setting, our method can be used to visualize three independent channels from a mpMRI data set as one image. Our computational experiments and the observer study demonstrate that this type of visualization indeed enables human observers to discern clinically relevant pathologies.
Normally, a multiparametric data set is visualized as several independent grayscale images that are shown successively or side-by-side – or one channel is overlaid on the other by using an arbitrary one-dimensional color map. Using our method, the whole data set can be read by viewing one combined color-coded image that is based on a perceptually optimized color map. Conceptually, this is a major improvement because the visual system of human observers can be used more efficiently. Thus, more information can be made available for clinical or basic research-related insight based on imaging data.
We envision that our proposed method could be used in clinical routine as a plugin for radiology image viewers. As soon as radiologists gain experience with this new type of visualization, they might use it in a wide range of applications. One possible benefit of the method would be to get a quick initial impression of an imaging data set without having to look at different MRI sequences in parallel. Also, a possible benefit would be to provide intuitive visualizations for non-radiologists or patients. Also in a scientific setting, our method could be used to enable an easy-to-understand visualization of multiparametric measurements.
Training effects
Pattern recognition by human observers requires experience23,24. Especially, the detection of prostate cancer in MR images significantly improves with training25. Consequently, this new way of visualization can probably only be exploited completely after a training period. In the user study that we present, observers received only an extremely short training session – in fact, they were only shown one example of a prostate that contained cancerous and non-cancerous regions. Yet, these observers were already able to detect cancer tissue with a classification performance that was comparable to the gold standard. Possibly, this performance would further improve after a longer training period. The novel visualization method we describe could be used in a larger, prospective study, that would also assess performance as observers gain experience with color coded images.
Limitations
Color-based visualization is more technically challenging than grayscale visualization. For example, the calibration of color displays is more laborious as the calibration of grayscale displays – and not as widely used2. However, the extra effort for calibration of color displays in medical imaging might be justified by other benefits of color-coding information.
Another limitation of color-based methods is related to the observer: Color vision anomalies are quite common and might compromise diagnostic reliability of these methods. Especially, 8% of the male population and 0.4% of the female population have detectable color vision abnormalities, most commonly deuteranomaly26,27,28. For these observers, the color maps we present are probably not optimally suited.
Visualization versus automatic classification of multiparametric data sets
Human observers cannot use an arbitrary number of information channels for their decision-making process. Consequently, in practice, only a small number of different MR measurements is used to make a diagnosis. Computer-based classification algorithms are a way to overcome this limitation, because they can combine an arbitrary number of measurements to a single metric. Automatic tissue classification methods have been established for histological imaging29,30, CT imaging31, and multivariate MR imaging, e.g. for the detection of cancer in breast32 and prostate cancer33. Especially, such a classification method has been successfully applied for automatic prostate cancer grading from mpMRI data34. Automatic methods for the detection of clinically relevant pathologies are also summarized as “computer-aided diagnosis”35. These approaches have been successfully combined with human decision making, thereby complementing, instead of substituting, human observers36. More recently, fully automatic approaches (termed radiomics) have been used to extract even more clinically useful information from imaging datasets37,38. These approaches use raw data and do not require any human interaction.
Still, even today, visual examination by a trained human observer remains the gold standard in diagnostic imaging. Although machine learning algorithms can incorporate a huge number of parameters in their computations, humans still outperform automatic algorithms in many visual pattern recognition tasks in medicine. There are several possible reasons for this – and one of the main reasons is that humans excel in combining heterogeneous data from different sources. A clinical radiologist considers much more information than just the raw image data: he or she considers patient information (age, sex, lifestyle, family history), previous diseases, previous surgery, previous imaging studies, laboratory tests such as prostate-specific antigen (PSA) and other factors.
Given that humans are very good at deriving decisions from data, it is a relevant problem how to make this data accessible to them. In an medical imaging context, making data accessible relates to optimally displaying imaging data, especially high-dimensional MR information so that human observers can make a well-informed clinical decision. Potentially, the method we present here is a powerful tool to achieve this.
Additional Information
How to cite this article: Kather, J. N. et al. Color-coded visualization of magnetic resonance imaging multiparametric maps. Sci. Rep. 7, 41107; doi: 10.1038/srep41107 (2017).
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Acknowledgments
We acknowledge the financial support of the Deutsche Forschungsgemeinschaft and Ruprecht-Karls-Universität Heidelberg within the funding programme Open Access Publishing.
Footnotes
Some of the authors are members of the institutes “Computer Assisted Clinical Medicine” or “Institute of Clinical Radiology and Nuclear Medicine”. These institutes cooperate with Siemens (Munich, Germany) on various projects. However, this cooperation did not influence the present study in any way. None of the authors declare any other competing financial interest.
Author Contributions J.N.K. and F.G.Z. conceived the experiments and planned the study. C.A.W., A.W., U.A., M.W. and L.R.S. contributed data sets and gave important advice for the study design. J.N.K., Y.B. and F.G.Z. analyzed the data. J.N.K. and F.G.Z. wrote the manuscript. All authors revised the manuscript and gave intellectual input to the conclusions being drawn.
References
- Weinreb J. C. et al. PI-RADS Prostate Imaging - Reporting and Data System: 2015, Version 2. Eur Urol 69, 16–40, doi: 10.1016/j.eururo.2015.08.052 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kimpe T. & Tuytschaever T. Increasing the number of gray shades in medical display systems–how much is enough? J Digit Imaging 20, 422–432, doi: 10.1007/s10278-006-1052-3 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Weidner A. M. et al. Value of multiparametric prostate MRI of the peripheral zone. Z Med Phys 21, 198–205, doi: 10.1016/j.zemedi.2010.12.004 (2011). [DOI] [PubMed] [Google Scholar]
- Elliot A. J. & Maier M. A. Color psychology: effects of perceiving color on psychological functioning in humans. Annu Rev Psychol 65, 95–120, doi: 10.1146/annurev-psych-010213-115035 (2014). [DOI] [PubMed] [Google Scholar]
- Bauer B., Jolicoeur P. & Cowan W. B. Visual search for colour targets that are or are not linearly separable from distractors. Vision Res 36, 1439–1466, doi: 10.1016/0042-6989(95)00207-3 (1996). [DOI] [PubMed] [Google Scholar]
- D’Zmura M. Color in visual search. Vision Res 31, 951–966 (1991). [DOI] [PubMed] [Google Scholar]
- Conway B. R. et al. Advances in color science: from retina to behavior. J Neurosci 30, 14955–14963, doi: 10.1523/JNEUROSCI.4348-10.2010 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Judd D. B. & Wyszecki G. Color in Business, Science and Industry. 3rd edn, 388 (Wiley-Interscience, 1975). [Google Scholar]
- Ware C. Information, Visualization, Perception for Design. 2nd ed. edn, 2–5 (Morgan Kaufmann, Elsevier, 2004). [Google Scholar]
- Xu S. et al. Real-time MRI-TRUS fusion for guidance of targeted prostate biopsies. Comput Aided Surg 13, 255–264, doi: 10.3109/10929080802364645 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moffat B. A. et al. Functional diffusion map: a noninvasive MRI biomarker for early stratification of clinical brain tumor response. Proc Natl Acad Sci USA 102, 5524–5529, doi: 10.1073/pnas.0501532102 (2005). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brun A., Park H.-J., Knutsson H. & Westin C.-F. In Computer Aided Systems Theory - EUROCAST 2003: 9th International Workshop on Computer Aided Systems Theory Las Palmas de Gran Canaria, Spain, February 24–28, 2003 Revised Selected Papers (eds Roberto Moreno-Díaz & Franz Pichler) 518–529 (Springer Berlin Heidelberg, 2003).
- Saba L. et al. Carotid artery dissection on non-contrast CT: does color improve the diagnostic confidence? Eur J Radiol 83, 2288–2293, doi: 10.1016/j.ejrad.2014.09.001 (2014). [DOI] [PubMed] [Google Scholar]
- Kumar V. & Udayashankara V. In Proceedings of CCSO. 10–15 (2013).
- Ropinski T., Oeltze S. & Preim B. Survey of glyph-based visualization techniques for spatial multivariate medical data. Comput Graph 35, 392–401 (2011). [Google Scholar]
- Twellmann T., Saalbach A., Gerstung O., Leach M. O. & Nattkemper T. W. Image fusion for dynamic contrast enhanced magnetic resonance imaging. Biomed Eng Online 3, 35, doi: 10.1186/1475-925X-3-35 (2004). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zollner F. G., Zimmer F., Klotz S., Hoeger S. & Schad L. R. Functional imaging of acute kidney injury at 3 Tesla: investigating multiple parameters using DCE-MRI and a two-compartment filtration model. Z Med Phys 25, 58–65, doi: 10.1016/j.zemedi.2014.01.002 (2015). [DOI] [PubMed] [Google Scholar]
- Zollner F. G. et al. An open source software for analysis of dynamic contrast enhanced magnetic resonance images: UMMPerfusion revisited. BMC Med Imaging 16, 7, doi: 10.1186/s12880-016-0109-0 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zollner F. G. et al. UMMPerfusion: an open source software tool towards quantitative MRI perfusion analysis in clinical routine. J Digit Imaging 26, 344–352, doi: 10.1007/s10278-012-9510-6 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kather J. N. et al. New Colors for Histology: Optimized Bivariate Color Maps Increase Perceptual Contrast in Histological Images. PLoS One 10, e0145572, doi: 10.1371/journal.pone.0145572 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jain A. K. Fundamentals of Digital Image Processing. 68–73 (Prentice Hall, 1989). [Google Scholar]
- Lindsey D. T. et al. Color channels, not color appearance or color categories, guide visual search for desaturated color targets. Psychol Sci 21, 1208–1214, doi: 10.1177/0956797610379861 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pugliese F. et al. Learning curve for coronary CT angiography: what constitutes sufficient training? Radiology 251, 359–368, doi: 10.1148/radiol.2512080384 (2009). [DOI] [PubMed] [Google Scholar]
- Rosenkrantz A. B. et al. Comparison of interreader reproducibility of the prostate imaging reporting and data system and likert scales for evaluation of multiparametric prostate MRI. Am J Roentgenol 201, W612–618, doi: 10.2214/AJR.12.10173 (2013). [DOI] [PubMed] [Google Scholar]
- Akin O. et al. Interactive dedicated training curriculum improves accuracy in the interpretation of MR imaging of prostate cancer. Eur Radiol 20, 995–1002, doi: 10.1007/s00330-009-1625-x (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Machado G. M., Oliveira M. M. & Fernandes L. A. A physiologically-based model for simulation of color vision deficiency. IEEE Trans Vis Comput Graph 15, 1291–1298, doi: 10.1109/TVCG.2009.113 (2009). [DOI] [PubMed] [Google Scholar]
- Cole B. L. Assessment of inherited colour vision defects in clinical practice. Clin Exp Optom 90, 157–175, doi: 10.1111/j.1444-0938.2007.00135.x (2007). [DOI] [PubMed] [Google Scholar]
- Birch J. Diagnosis of defective colour vision. (Oxford University Press, 1993). [Google Scholar]
- Nawaz S. & Yuan Y. Computational pathology: Exploring the spatial dimension of tumor ecology. Cancer Lett, doi: 10.1016/j.canlet.2015.11.018 (2015). [DOI] [PubMed] [Google Scholar]
- Kather J. N. et al. Multi-class texture analysis in colorectal cancer histology. Sci Rep 6, 27988, doi: 10.1038/srep27988 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chong D. Y. et al. Robustness-Driven Feature Selection in Classification of Fibrotic Interstitial Lung Disease Patterns in Computed Tomography Using 3D Texture Features. IEEE Trans Med Imaging 35, 144–157, doi: 10.1109/TMI.2015.2459064 (2016). [DOI] [PubMed] [Google Scholar]
- Wang Y., Morrell G., Heibrun M. E., Payne A. & Parker D. L. 3D multi-parametric breast MRI segmentation using hierarchical support vector machine with coil sensitivity correction. Acad Radiol 20, 137–147, doi: 10.1016/j.acra.2012.08.016 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Litjens G. J. et al. Computer-extracted Features Can Distinguish Noncancerous Confounding Disease from Prostatic Adenocarcinoma at Multiparametric MR Imaging. Radiology 278, 135–145, doi: 10.1148/radiol.2015142856 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fehr D. et al. Automatic classification of prostate cancer Gleason scores from multiparametric magnetic resonance images. Proc Natl Acad Sci USA 112, E6265–6273, doi: 10.1073/pnas.1505935112 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Doi K. Computer-aided diagnosis in medical imaging: historical review, current status and future potential. Comput Med Imaging Graph 31, 198–211, doi: 10.1016/j.compmedimag.2007.02.002 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hambrock T., Vos P. C., Hulsbergen-van de Kaa C. A., Barentsz J. O. & Huisman H. J. Prostate cancer: computer-aided diagnosis with multiparametric 3-T MR imaging–effect on observer performance. Radiology 266, 521–530, doi: 10.1148/radiol.12111634 (2013). [DOI] [PubMed] [Google Scholar]
- Aerts H. J. W. L. et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nature Communications 5, 1–8, doi: 10.1038/ncomms5006 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Parmar C., Grossmann P., Bussink J., Lambin P. & Aerts H. J. W. L. Machine Learning methods for Quantitative Radiomic Biomarkers. Sci Rep 5, 1–11, doi: 10.1038/srep13087 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]