Skip to main content
International Journal of Biomedical Imaging logoLink to International Journal of Biomedical Imaging
. 2026 Mar 6;2026:4456355. doi: 10.1155/ijbi/4456355

Deep Learning Framework for Automated MRI Planimetry in Multiple Sclerosis

Stephanie Mangesius 1,2, Daniela Schiefeneder 3, Matthias Schwab 1, Markus Tiefenthaler 1, Florian Deisenhammer 4, Markus Haltmeier 3,, Elke R Gizewski 1,2
Editor: Irfan Ahmad
PMCID: PMC12964314  PMID: 41799369

Abstract

Brain volume changes and infratentorial involvement are key predictors of disability in multiple sclerosis (MS) and can be assessed using magnetic resonance imaging (MRI) planimetry. Although MRI planimetry is less susceptible to methodological and patient‐related confounders than volumetry, it currently depends on manual measurements by unblinded experts, an approach that is time‐consuming and vulnerable to bias. In this study, we present a fully automated deep learning framework for deriving brainstem planimetric measurements from MRI. The pipeline integrates an automated midsagittal plane (MSP) detection algorithm with a convolutional neural network trained to perform the segmentations required for planimetry. The automated method shows strong agreement with manual measurements and remains robust across scanners and acquisition protocols. These findings suggest that the proposed framework enables reliable, reproducible, and scalable MRI planimetry, supporting objective assessment of disease progression and treatment response in patients with MS.

Keywords: deep learning, midsagittal plane detection, MRI planimetry, multiple sclerosis

1. Introduction

MS is a chronic, neuroinflammatory, and neurodegenerative disorder of the central nervous system [1, 2]. Early neuronal and axonal loss leads to measurable brain volume changes [35], and progressive atrophy has been proposed as an MRI‐based predictor of irreversible disability [6, 7]. Infratentorial involvement, in particular, is strongly associated with poorer clinical outcomes [8]. While MRI is a powerful tool for monitoring these structural changes, conventional imaging often provides limited or subjective information directly relevant to disease progression. This highlights the need for quantitative and reproducible imaging biomarkers to support early diagnosis, accurate prognosis, and informed clinical decision‐making.

MRI volumetry can improve diagnostic accuracy but is often limited by scanner variability, availability, and complex processing requirements. In contrast, MRI planimetry provides robust and readily obtainable metrics that can be evaluated using standard imaging software. Several studies have demonstrated the diagnostic utility of brainstem planimetric measures, including midbrain and pontine areas and diameters, as well as the magnetic resonance parkinsonism index (MRPI), for differentiating patients with neurodegenerative disorders [9, 10].

Planimetry is typically performed on the midsagittal plane (MSP), a single well‐defined slice of the MRI volume. Although manual determination and measurement are straightforward for radiologists, the process is time‐consuming, rater‐dependent, and difficult to standardize across sites [11]. Fully automated planimetric workflows have not yet been established. These limitations underscore the need for automated, reproducible methods for both MSP extraction and planimetric measurement, motivating the development of a deep learning–based frameworks for automated planimetry.

The primary aim of this study is to develop and evaluate a deep learning framework for automated and standardized MRI planimetry. The framework focuses on brainstem structures that are particularly susceptible to atrophy and may serve as predictors of disease progression or treatment response. The proposed approach comprises the following:

  • 1.

    Automated extraction of the MSP from 3D MRI volumes.

  • 2.

    Convolutional neural network (CNN)–based segmentation of the midbrain and pons.

  • 3.

    Planimetric measurement of areas and diameters from the segmented MSP.

  • 4.

    Use of a weighted entropy loss function to address class imbalance during network training.

Evaluation was performed on a cohort of MS patients across multiple subtypes, with automated results compared against manual planimetric measurements by medical experts, which is considered a gold standard reference.

While the novelty of this work does not lie in the development of a new CNN architecture, the innovation resides in the first application of CNNs to automated brainstem MRI planimetry, combined with task‐specific loss adjustments. The framework demonstrates robust, reproducible, and scalable performance, and the pipeline can accommodate more advanced network architectures in future work without modification to its overall design.

Table 1 summarizes abbreviations used in this work.

Table 1.

Abbreviations used.

Abbreviation Definition
MRI Magnetic resonance imaging
MS Multiple sclerosis
MSP Midsagittal plane
EDSS Expanded Disability Status Scale
CNN Convolutional neural network
ROI Region of interest
SNR Signal‐to‐noise ratio
DICE Sørensen–Dice coefficient
AD Area difference
RAAD Relative absolute area difference
ADD Absolute diameter difference
MPRAGE Magnetization‐prepared rapid gradient echo
HAUS Hausdorff distance
RADD Relative absolute diameter difference
SD Standard deviation

2. Materials and Methods

2.1. Background

MRI planimetry has been applied to neurodegenerative and neuroinflammatory disorders to quantify midbrain and pontine structures, supporting early diagnosis and disease monitoring. Previous methods rely on manual or semiautomatic segmentation, which is labor‐intensive and subject to interrater variability [1219]. While automated MSP detection and brainstem segmentation methods have been proposed [20, 21], these approaches often suffer from limited generalizability, small datasets, or a lack of integration with downstream clinical applications.

Deep learning–based segmentation techniques have demonstrated high accuracy in neuroimaging tasks [2226]; however, fully automated MRI planimetry remains unexplored. This study addresses this gap by presenting an end‐to‐end framework for MSP detection, segmentation, and planimetric measurement, enabling reproducible assessment across a heterogeneous patient cohort. To the best of our knowledge, no other automated MRI planimetry pipeline currently exists, and we benchmark our method against manual planimetric measurements, considered the gold standard.

2.2. Dataset

High‐resolution MRI scans were acquired using four 1.5 T and two 3 T Siemens scanners with a standard multichannel head coil. T1‐weighted 3D magnetization‐prepared rapid gradient echo (MPRAGE) sequences were obtained for all participants at the Neuroimaging Core Facility. Previous studies indicate that MRI planimetry produces consistent results across 1.5 and 3 T scanners [27]. Clinical assessments included demographic data, disease onset, and the Expanded Disability Status Scale (EDSS); see Table 2.

Table 2.

Demographics and MRI acquisition details for all cohorts.

Training Validation Test Healthy
Patients (MRI) 56 (136) 9 (18) 22 (54) 19 (19)
No. of MRIs per patient 2.43 ± 1.56 2 ± 1.49 2.45 ± 1.92 1
Sex (m/f) 22/34 1/8 9/13 5/14
Age at MRI 46.4 ± 9.71 44.94 ± 6.36 43.02 ± 8.54 37.32 ± 13.15
Disease duration 16.34 ± 9.28 15.06 ± 8.55 15.35 ± 9.7 0
EDSS score 4.9 ± 1.73 4.19 ± 2.15 4.76 ± 1.86 0
Contrast agent 29 (21.32%) 1 (5.56%) 13 (24.07%) 0
Scanner 1.5/3 T 92/44 9/9 45/9

The dataset comprises 227 MRI volumes from 106 participants: 87 patients with MS (208 MRI scans) and 19 healthy controls (19 MRI scans). MS patients were randomly assigned to training (136 MRI scans), validation (18 MRI scans), and test (54 MRI scans) sets. Healthy controls were used exclusively for an additional test set (19 MRI scans). The split was performed at the patient level, as scan‐level splitting can introduce bias and data leakage if data from the same patient appear in both training and test sets, potentially inflating performance metrics.

All procedures were approved by the local Ethics Committee of the Medical University of Innsbruck. Written informed consent was waived for MS patients due to the retrospective study design (Ethics 1267/2019). Prospective healthy controls provided written informed consent (AN 5100 325/4.19 384/5.3 4213a).

2.3. MRI Planimetry

Planimetric measurements are parameters derived from MRI scans restricted to the MSP. Several of these measures have been proposed to quantify atrophy in specific brainstem structures and serve as disease‐specific markers. Previous studies have demonstrated their high diagnostic utility [9, 28, 29]. In this study, we focus on four specific parameters: midbrain diameter, midbrain area, pontine area, and pontine diameter.

The workflow for obtaining brainstem‐derived measures involves determining the MSP, identifying the midbrain and pons regions, and computing the area and diameter of these regions. While previous studies performed the first two steps manually, the aim of the present study is to replace them with fully automated algorithms.

2.4. Automated MSP Computation

The MSP was automatically extracted from each 3D MRI scan using a self‐developed algorithm that combines the 2D Radon transform with symmetry metrics [3033]. The MSP, which separates the brain into left and right hemispheres, is defined by its roll angle ((z, x)‐plane), yaw angle ((y, x)‐plane), and its perpendicular distance from the origin.

The algorithm proceeds as follows (compare with Figure 1). First, an axial slice F in the upper brain is selected, and the center of mass of the thresholded binary image is computed. We then apply the Radon transform, which computes the average of $ F $ along lines characterized by normal direction θ and signed perpendicular distance s. The values θ (yaw angle) and s (perpendicular distance) are identified as the point in the Radon transform with minimal intensity value, corresponding to the intersection of the MSP with the selected axial plane. All axial slices are subsequently rotated by −θ to align them with the coronal plane.

Figure 1.

Workflow of MSP detection. (a) Axial slice with center of mass, (b) Radon transform with the minimal value highlighted, and (c) axial slice showing the line corresponding to this minimal value. (d) Coronal slice with the computed point, (e) intermediate alignment, and (f) final MSP.

graphic file with name IJBI-2026-4456355-g003.jpg

(a)

graphic file with name IJBI-2026-4456355-g004.jpg

(b)

graphic file with name IJBI-2026-4456355-g005.jpg

(c)

graphic file with name IJBI-2026-4456355-g006.jpg

(d)

graphic file with name IJBI-2026-4456355-g007.jpg

(e)

graphic file with name IJBI-2026-4456355-g008.jpg

(f)

Next, the central coronal slice is rotated over candidate roll angles ϕ ∈ (−10, 10), and the angle that maximizes vertical symmetry is selected as the roll angle. All coronal slices are rotated accordingly. The MSP is then defined as the sagittal slice intersecting the computed line in each coronal slice. Median values of yaw, roll, and perpendicular distance across slices are used to improve robustness.

Finally, all MSP images are resampled to 176 × 256 pixels, and a 128 × 128 ROI consistently containing the midbrain and pons is extracted.

2.5. Manual Planimetric Measurements

Manual measurements served as the reference standard for training and validating the automated method. The procedure, following established protocols [9], was independently applied in this study and is described below. Segmentation and quantitative extraction were performed in MATLAB.

Midbrain and pontine areas were derived by first defining two reference lines: one extending from the inferior edge of the quadrigeminal plate (tectum) to the superior pontine notch and a second parallel line passing through the inferior pontine notch, thereby separating the pons from the medulla oblongata (see Figure 2). Using these anatomical boundaries, polygons were manually drawn to delineate the midbrain and pons, and the areas were then computed.

Figure 2.

Figure 2

Manual planimetric measurements of midbrain (green) and pons (blue).

Polygons were used for area measurements, and ellipses for diameter measurements. For diameter measurements, ellipses were fitted to the segmented regions, excluding the pontine tegmentum in the pons and the collicular plate in the midbrain. Maximum midsagittal anteroposterior diameters were then derived from the minor axes of the fitted ellipses.

2.6. Deep Learning Model

Let F denote the 128 × 128 MSP ROI image. For automated segmentation, we train a CNN U θ to output a probability mask: P≔U θ (F). Binary segmentation masks are obtained by thresholding P.

For the CNN architecture U θ , we use a standard U‐Net [34]. The CNN was implemented in Python using the Keras deep learning library. The U‐Net employs convolutions with filter sizes of 3 × 3 and 1 × 1, as well as max‐pooling and upsampling operations with a factor of 2. In total, the network contains 31,031,685 trainable parameters. An overview of the architecture is shown in Figure 3.

Figure 3.

Figure 3

Used CNN architecture.

2.7. Loss Function

Training the network requires an objective function that measures how well the predicted segmentation mask matches the ground‐truth mask. Because the target structures (midbrain and pons) occupy only a small portion of the image, the data are strongly imbalanced. To address this, we use a combination of Dice loss and weighted binary cross‐entropy, which stabilizes optimization and ensures that small anatomical structures are not ignored (compare [35, 36]). The final loss is defined as the sum of Dice loss and weighted cross‐entropy loss:

LG,GLDiceG,G+LWCEG,G,

where

LDiceG,G12iGiGiiGi+iGiLLWCEG,GiIk=01,wi·Gik·logkGiwlog1+k=01,I·1IkIk+I·1CkCk.

To aid clarity, all variables used above are explicitly defined:

  • 1.

    I : set of all pixel indices in the image

  • 2.

    G[i] : ground‐truth label of pixel i (1 = structure, 0 = background)

  • 3.

    Gi: predicted probability for pixel i

  • 4.

    I 1: set of all pixels belonging to the target structure

  • 5.

    C 1: set of contour pixels of the structure

  • 6.

    |A| : number of pixels in a set |A|

  • 7.

    1A (i): indicator function (1 if iA, 0 otherwise)

  • 8.

    w[i]: class‐ and contour‐aware weighting factor for pixel i

Dice loss penalizes spatial mismatches between prediction and ground truth, ensuring high overlap and preserving structural shape while remaining robust to class imbalance. Weighted cross‐entropy emphasizes pixel‐wise accuracy for small target structures by assigning higher weights to minority‐class and boundary pixels, preventing trivial background‐only predictions. Together, these terms form a robust loss function tailored to the small, clinically relevant regions of the MSP.

2.8. Model Training

For each segmentation task (area or diameter; midbrain or pons), a separate model was trained. All models were based on data from 56 patients with multiple sclerosis (61% female, mean age 46.4 years), comprising a total of 136 MRI scans. The training dataset included 136 pairs of MRI scans and their corresponding binary segmentation masks. Networks were trained for 30 epochs with a batch size of 4. After each epoch, model performance was evaluated on the validation set using the Keras accuracy metric, and the weights achieving the best validation performance were retained as the final model.

To improve robustness, data augmentation was applied using the ImageDataGenerator module in Keras, including random horizontal and vertical shifts of −10 to 10 pixels, rotations of −15° to 15°, and zooms of 80%–120%.

2.9. Automated Planimetry

Automated planimetry was applied to the four trained neural networks, each targeting a specific task. For automated area measurements, the number of pixels in the binary segmentation masks produced by the networks was counted and multiplied by the pixel area. For diameter measurements, an ellipse was algorithmically fitted to the corresponding regions. First, the boundary points of the binary segmentation mask were selected, which do not form perfect ellipses. An ellipse was then calculated to best fit this set of boundary points in a least‐squares sense. The diameter was defined as the length of the minor axis of this fitted ellipse.

2.10. Evaluation Metrics

To quantitatively assess agreement between automated and manual segmentations, several complementary metrics were used to capture spatial overlap, area and diameter accuracy, and boundary distance:

  • 1.

    Sørensen–Dice coefficient (DICE)

  • 2.

    Area difference (AD)

  • 3.

    Relative absolute area difference (RAAD)

  • 4.

    Absolute diameter difference (ADD)

  • 5.

    Relative absolute diameter difference (RADD)

  • 6.

    Hausdorff distance (HAUS)

DICE measures spatial overlap between masks (1 = perfect agreement). AD and RAAD quantify absolute and relative differences in segmented area, while ADD and RADD evaluate differences in anatomically relevant diameters. HAUS assesses boundary accuracy by measuring the maximum contour deviation. Together, these metrics provide a comprehensive evaluation of geometric accuracy and segmentation reliability, which is essential for reproducible planimetric measurements.

Formally, these quantities are defined as

DICE2ABA+BADAB×pixelsizeRAAD2ABA+B×100ADDdiamAdiamBRADD2diamAdiamBdiamA+diamBHAUSmaxmaxaAda,B,maxbBdb,A,

where A denotes the manual segmentation mask and B the automated segmentation mask, diam(·) is the diameter, and d(x, Y) denotes the Euclidean distance between a point x and a set Y.

2.11. Evaluation

To evaluate the methods, an independently selected third dataset of 54 images from 22 patients was used. Automated planimetric measurements were compared with manual measurements, which served as the gold standard. Both manual and automated segmentations were performed on the same MSP determined by the algorithm described in Subsection 2.4. All evaluation metrics from Subsection 2.10 were calculated, and each segmentation mask was visually inspected and approved by an independent expert. Correlation analysis was conducted, and robustness was further assessed using a second independent dataset. Data splitting was performed at the patient level to prevent data leakage between training, test, and evaluation sets.

3. Results

3.1. Area Measurements

We first evaluated area measurements of the midbrain and pons on the test set. Figure 4 illustrates the two best and two worst segmentation results for the trained U‐Net based on DICE scores for the pons and midbrain, respectively. The images in the last column of the lower panel show a tumor in the hypothalamus, representing a challenging out‐of‐distribution case. DICE scores for pons segmentations ranged from 0.984 (best) to 0.901 (worst), while midbrain segmentations ranged from 0.984 to 0.892. Even the segmentations with the lowest DICE scores still closely resemble the ground truth.

Figure 4.

Comparison between manual (lower row) and automatic (upper row) segmentations for the (a) best and (b) worst cases based on the pons (two left columns) and midbrain (two right columns) segmentation. Segmentations of the midbrain are marked in green, and segmentations of the pons are marked in blue. These examples show that automated segmentations closely follow expert contours even in difficult cases.

graphic file with name IJBI-2026-4456355-g011.jpg

(a)

graphic file with name IJBI-2026-4456355-g012.jpg

(b)

A quantitative performance analysis is presented in Table 3, where manual segmentations are used as the ground truth. For all accuracy measures, the mean, standard deviation (SD), and 95% confidence interval were calculated on the test data.

Table 3.

Performance analysis of automated segmentation compared to manual segmentation used as ground truth on the test data. Statistics are provided for DICE, RAAD, and HAUS averaged over the MS test dataset including 54 images.

Ellipse metrics (m e a n ± S D)
DICE RAAD HAUS
Midbrain 0.941 ± 0.02 4.52 ± 3.64 1.24 ± 0.43
Pons 0.966 ± 0.011 3.02 ± 2.06 1.59 ± 0.64

3.2. Diameter Measurements

Next, we compare automated and manual ellipses and corresponding diameter measurements. As explained in Subsection 2.9, the calculation of the diameter is done in two consecutive steps consisting of CNN segmentation and ellipse fitting using the minor axis as the diameter. Each ellipse found by the algorithm was visually inspected by an expert. In general, the quality of the automatically found ellipses was viewed as high, meaning that no substantial misclassification was found. Figure 5 shows examples of the automatically found ellipses compared to manually fitted ellipses. Mean values and SD for D I C E, A D D, and R A D D are shown in Table 4.

Figure 5.

Comparison between manual (lower row) and automatic (upper row) segmentations for the (a) best and (b) worst cases based on the pons (two left columns) and midbrain (two right columns) segmentation. Segmentations of the midbrain are marked in green, and segmentations of the pons are marked in blue.

graphic file with name IJBI-2026-4456355-g013.jpg

(a)

graphic file with name IJBI-2026-4456355-g014.jpg

(b)

Table 4.

Performance analysis for diameter measurements. Statistics are for the Sørensen–Dice coefficient (DICE), the absolute diameter difference (ADD), and the relative absolute diameter difference (RADD) taken over an evaluation dataset including 54 images.

Ellipse metrics (m e a n ± S D)
DICE ADD RADD
Midbrain 0.908 ± 0.046 0.400 ± 0.198 8.31 ± 4.86
Pons 0.925 ± 0.033 0.496 ± 0.368 5.90 ± 4.41

3.3. Regression and Correlation Analysis

We evaluated the agreement between manual and automated segmentations on the test data. Automated and manual area measurements for the midbrain and pons were highly correlated (R 2 = 0.878 and 0.886, respectively; p < 10−25), as shown in Figure 6a,b. Bland–Altman plots indicate a slight tendency for automated oversegmentation, with mean ADs of 4.21 mm2 (midbrain) and 9.33 mm2 (pons) and limits of agreement (mean ±1.96σ) corresponding to σ = 7.08 and 17.99 mm2, respectively.

Figure 6.

Regression analysis for (a) midbrain area and (b) pons area on the test data. The horizontal axis is for manual, and the vertical axis is for automated area values. Bland–Altman plots for (c) midbrain and (d) pons. The horizontal axis is the average, and the vertical axis is the difference between automated and manual area values. Strong correlations and tight Bland–Altman limits indicate minimal bias and support the use of automated planimetry for longitudinal monitoring.

graphic file with name IJBI-2026-4456355-g015.jpg

(a)

graphic file with name IJBI-2026-4456355-g016.jpg

(b)

graphic file with name IJBI-2026-4456355-g017.jpg

(c)

graphic file with name IJBI-2026-4456355-g018.jpg

(d)

Ellipse‐derived diameters were similarly well correlated with manual measurements (R 2 = 0.780 for midbrain, R 2 = 0.794 for pons; p < 10−18), confirming that the automated approach accurately captures both area and diameter metrics (see Figure 7).

Figure 7.

Regression analysis for automatically and manually found diameters of (a) midbrain and (b) pons.

graphic file with name IJBI-2026-4456355-g019.jpg

(a)

graphic file with name IJBI-2026-4456355-g020.jpg

(b)

3.4. Robustness Against MRI Settings

The test cohort contains 13 MRI volumes where a contrast agent was used. Further, different pixel spacings and slice thicknesses were used during the MRI measurements. The big subgroups consist of images sized 128 × 256 with a slice thickness of 1.2 mm and images sized 176 × 256 with a slice thickness of 1 mm. The influence of these parameters is shown in Table 5. Student′s t‐test revealed that differences between results with and without contrast agent were not significant (p ≥ 0.5).

Table 5.

Analysis comparing the use of contrast agent (CA), where Y stands for contrast agent and N for no contrast agent, and the influence of the image size N z × 256 where N z = 128 or N z = 176. The symbol “#” denotes the number of test samples.

Area midbrain (m e a n ± S D) Area pons (m e a n ± S D)
CA, N z , # DICE RAAD HAUS DICE RAAD HAUS
N, 128, 28 0.938 ± 0.019 5.20 ± 3.74 1.28 ± 0.470 0.967 ± 0.009 2.60 ± 1.83 1.48 ± 0.424
N, 176, 10 0.954 ± 0.017 2.23 ± 1.87 1.18 ± 0.335 0.973 ± 0.006 2.89 ± 1.52 1.57 ± 0.528
Y, 128, 11 0.933 ± 0.025 5.57 ± 4.30 1.30 ± 0.491 0.960 ± 0.013 4.07 ± 2.89 1.89 ± 0.99

On the other hand, when 176 × 256 MRI scans were compared to the 128 × 256 without a contrast agent, we received a p value of 0.02 for the DICE in the midbrain. The comparison to smaller sized images with contrast agent led to similar results p = 0.03. Further, with p values of 0.003 (without contrast agent) and 0.03 (with contrast agent), respectively, we received significant differences for the midbrain RAAD. The pons DICE led to p values of 0.05 (without contrast agent) and 0.01 (with contrast agent), respectively. On the other hand, there were no significant differences for the pons concerning the RAAD.

3.5. Robustness Against Noise

To further test the robustness of our framework, we evaluated its performance on MRI scans corrupted with noise. As it is well known that noise in MRI magnitude images follows a Rician distribution [37], experiments were performed for both Rician and Gaussian noise. To this end, noise was added at varying signal‐to‐noise ratios (SNRs) to simulate real‐world degradation. The segmentation performance was measured using Dice coefficients, and failure cases were identified as empty segmentation masks. Results show that the network maintains high accuracy at moderate noise levels, but performance begins to decline as the SNR drops below 15. Notably, we observed a sharp threshold effect: The segmentation either works well or fails completely beyond a certain noise level. However, this threshold varies from image to image, suggesting that each scan has an individual noise tolerance limit (see Figure 8).

Figure 8.

(a) Mean Dice coefficients of segmented images corrupted with Gaussian and Rician noise for different SNRs and (b) rate of zero outputs from the network for the two noise models at different SNRs.

graphic file with name IJBI-2026-4456355-g021.jpg

(a)

graphic file with name IJBI-2026-4456355-g022.jpg

(b)

3.6. Test on Healthy Control Group

To test the robustness of our algorithm with respect to the MRI class, we also examined the performance on a test set of 19 healthy controls. The results are shown in Table 6. One notices that error metrics are comparable to the error metrics on the MS patients, even though neither of the networks has been trained on healthy patients. This further supports the robustness of our approach.

Table 6.

Means and SD of the different error measures on a sample of 19 healthy patients.

Areas (m e a n ± S D) Ellipses (m e a n ± S D)
DICE RAAD HAUS DICE RAAD HAUS
Midbrain 0.929 ± 0.018 6.06 ± 4.78 1.12 ± 0.30 0.920 ± 0.030 0.206 ± 0.151 4.03 ± 2.95
Pons 0.961 ± 0.019 2.89 ± 3.07 1.51 ± 0.70 0.934 ± 0.025 0.525 ± 0.345 6.64 ± 4.33

3.7. Comparison to Manual Rating

Table 7 presents a comparison of the automated network segmentations with two independent raters. The results from two raters, one with professional radiological skills and the other a student, were compared. They also had varying levels of experience with MR planimetric measurement performance. Additionally, the two raters are compared to each other. One can observe that the CNN is able to maintain correlations between the two expert raters. This comparison demonstrates that 2D measurements remain unaffected by user experience.

Table 7.

Means and SD comparing the different measurements (two independent raters and the proposed algorithm). The best values are marked in bold.

CNN vs. R1 CNN vs. R2 R1 vs. R2
Area DICE (mean ± SD)
MB 0.902 ± 0.028 0.898 ± 0.033 0.900 ± 0.033
Pons 0.962 ± 0.010 0.954 ± 0.011 0.953 ± 0.012
Area AD (mean ± SD)
MB 7.95 ± 6.60 12.23 ± 8.19 11.4 ± 8.03
Pons 19.7 ± 14.0 31.6 ± 19.9 44.3 ± 15.2
Ellipse DICE (mean ± SD)
MB 0.918 ± 0.045 0.924 ± 0.053 0.920 ± 0.035
Pons 0.928 ± 0.029 0.888 ± 0.042 0.928 ± 0.038
Ellipse ADD (mean ± SD)
MB 0.185 ± 0.292 0.226 ± 0.364 0.025 ± 0.212
Pons 0.240 ± 0.401 0.751 ± 0.413 0.598 ± 0.370

3.8. Out of Distribution Cases

We analyzed patients whose midsagittal structures deviated substantially from those of all cases in the training and evaluation datasets. The first case involves a mass of unknown origin in the mammillary bodies (Figure 9a, marked with an arrow). Despite this anomaly, the U‐Net successfully segments the midbrain, likely benefiting from its learned shape priors and subtle intensity differences. The second case, which also appears only in the test dataset, involves an elongated basilar artery pressing against and slightly deforming the pons (Figure 9b, marked with an arrow). In this instance, the network misclassifies part of the artery as pontine tissue, likely due to minimal intensity contrast and the largely preserved geometric structure.

Figure 9.

Two out‐of‐distribution cases are shown: (a) a tumor in the mammillary bodies (marked by an arrow) and (b) a basilar artery displaced against the pons (marked by an arrow). Each example includes the MRI scan (left), the manual segmentation (middle), and the network prediction (right).

graphic file with name IJBI-2026-4456355-g001.jpg

(a)

graphic file with name IJBI-2026-4456355-g002.jpg

(b)

We emphasize that the training data did not contain any cases in which the basilar artery was displaced against the pons. We therefore conjecture that expanding the training dataset to include such anatomical variations would help overcome this limitation. Another complementary strategy may involve targeted data augmentation tailored to this specific challenging scenario.

4. Discussion

MRI planimetry is an established tool for assessing brain atrophy in MS and is closely linked to disease progression and treatment response [38]. Brainstem atrophy, in particular, has emerged as a significant contributor to disability progression and neurological decline and is a reliable predictor of future physical and cognitive impairment [39, 40]. In this work, we present a deep learning framework that automates MRI planimetry by integrating an MSP detection algorithm with a neural network for segmentation and measurement, aiming to mitigate the limitations of manual assessment and scanner‐related variability [4144]. Because planimetric segmentation is performed on a single MSP slice, the pipeline is fast and reduces analysis time by automating plane selection. Processing all four regions of interest (ellipses and polygons for pons and midbrain) required 0.45 s per patient on a standard PC with 8 GB RAM, compared with 106.7 s for manual segmentation [9].

The method was developed using a large database spanning all clinical MS phenotypes (clinically isolated syndrome, primary progressive MS, secondary progressive MS, and relapsing–remitting MS) and evaluated on an independent clinical cohort acquired on different MRI systems as part of routine care. These results support the feasibility and generalizability of automated infratentorial planimetry in real‐world settings. U‐Net‐based segmentations showed high agreement with manual labels, with mean DICEs of 0.966 (pons) and 0.941 (midbrain) on the test set [45]. Boundaries between brainstem substructures (pons, midbrain, and medulla) were the most challenging, yet automated contours typically deviated by only one‐to‐two‐pixel slices from manual delineations. This segmentation accuracy translated into strong concordance of planimetric areas: The relative absolute difference (RAD) was 3.02% for pons and 4.52% for midbrain, and Dice coefficients exceeded 0.89 for both structures across all test cases.

Ellipse‐based measurements likewise showed high spatial overlap between automated and manual fits (mean Dice: 0.925 for pons; 0.908 for midbrain). Diameter estimates were accurate (mean RAD: 5.9% for pons; 8.31% for midbrain) and strongly correlated with manual measurements. Consistent with prior observations, automated diameters were slightly smaller than manual ones in most cases (85% of pons; 92% of midbrain), but this small negative bias did not materially affect the correlations. This likely reflects mild underestimation in the ellipse‐fitting step, which can yield slightly shorter diameters even when the underlying automatic segmentations delineate larger areas. Alternative fitting strategies—such as minimizing the sum of squared geometric distances to boundary points rather than using algebraic fits—could mitigate this bias in future versions. With respect to clinical thresholds, this bias did not lead to threshold‐based reclassifications in our sample; however, measurements very close to a cutoff may warrant manual confirmation or a small calibration offset. Overall, we consider the higher consistency of the automatic measurements more important than the minor systematic underestimation, which can be accommodated in clinical decision‐making.

Robustness analyses indicated that performance remained stable down to an SNR of 15, well within the typical clinical range of 20–50. Generalization to healthy subjects—despite training exclusively on MS data—yielded Dice coefficients comparable to those in the MS test set. Contrast administration had no measurable effect on segmentation, whereas slice thickness (axial resolution) was a key determinant of accuracy, as expected.

Automated planimetry reduces observer dependence and improves reproducibility compared with manual assessments, which are sensitive to rater interpretation, scanner type, and acquisition parameters [42, 46]. This stability is particularly important for longitudinal monitoring, where small differences can change the clinical interpretation of progression and inform treatment decisions. More reliable tracking of brainstem atrophy could support earlier identification of patients at risk of disability worsening (e.g., in relation to EDSS) and guide escalation when disease‐modifying therapy (DMT) fails to slow neurodegeneration [38, 47]. Standardized, automated measurements may also reduce intersite variability in multicenter trials, strengthening study power and interpretability [41].

Limitations and future work include the following: (1) While segmentation accuracy was high, the method does not enforce parallel boundaries between pons and midbrain; this could be addressed via alternative loss functions, postprocessing constraints, or geometric regularizers. (2) The ellipse‐fitting routine showed a slight negative diameter bias relative to manual measurements, in contrast to the small positive bias in segmented areas. Approaches minimizing geometric distance metrics (e.g., [48]) may mitigate this in future versions. (3) Although manual planimetry correlates with clinical outcomes, analogous analyses for automated measures remain to be established; future studies should test associations with disability and treatment response using longitudinal cohorts.

We observed incidental findings in our cohort, including three cases of basilar artery dolichoectasia mildly indenting the brainstem and one mammillary body tumor. The tumor did not affect segmentation; dolichoectasia led to localized misclassification. Given its prevalence in older, largely male, hypertensive populations and the younger, predominantly female demographic of MS, this represents a limited and incidental constraint on performance. The mammillary body lesion (likely a hamartoma) is extremely rare and does not alter the overall conclusions [49]. Beyond MS, automated planimetry may extend to disorders with brainstem involvement, such as PSP and MSA, where brainstem atrophy is diagnostically informative [28, 29, 50]. Future work should integrate additional biomarkers (e.g., spinal cord atrophy and functional MRI) and correlate automated metrics with standardized disability scales (e.g., EDSS) to refine clinical utility [51].

For clinical adoption, integration into radiology workflows is essential. The system can export DICOM Segmentation and Structured Reports for automatic ingestion into PACS/reporting platforms (including via DICOMweb), enabling overlays and quantitative metrics at the radiologist′s workstation. Deployment models include on‐premises installation, vendor‐neutral microservices within PACS/RIS (via DICOM SR and HL7), or cloud‐based inference for multisite scalability. Prior implementation studies show that structured report transfer of AI‐derived measurements improves efficiency, reduces transcription errors, and enables sub‐5‐min turnaround times at scale [5256].

In summary, the proposed framework delivers fast, robust, and reproducible brainstem planimetry across scanners and acquisition settings. By reducing observer dependence and standardizing measurements, it has the potential to improve longitudinal monitoring, support treatment decisions, and accelerate the translational use of quantitative MRI biomarkers in MS care.

5. Summary and Outlook

In this paper, we present a U‐Net‐based segmentation approach to calculate the areas and diameters of the midbrain and pons on the MSP. Comparing the U‐Net‐based to manual segmentations yielded high Sørensen–Dice coefficients, indicating high accuracy in image segmentation. The automatically and manually measured areas of the pons and midbrain were highly correlated. Similarly, the spatial overlap between manually and automatically drawn ellipses was very high, leading to very accurate predictions of the diameters of the midbrain and pons. Our study demonstrates the feasibility of automated MRI planimetric segmentations of infratentorial structures in a clinical setting using acquired MRI scans from a large cohort of MS patients. The algorithms were successfully tested on an independent cohort of patients recruited from an MS clinic and scanned using different MRI systems as part of their routine diagnostic workup. Evaluating the algorithm on a larger cohort of MS patients and healthy controls is an important direction for future research. Our results show that automatic MRI planimetry in patients with early to moderately advanced MS, recruited in a clinical setting, works stably, regardless of variability in scanners and acquisition parameters. The implementation of the deep learning framework into clinical practice may play an important role in reliably predicting disease progression and response to treatment in patients with MS.

We employ a supervised deep learning strategy based on CNNs, which remain a well‐established and experimentally powerful technique for medical image segmentation. At the time this study was conducted, CNN‐based methods were considered state‐of‐the‐art for this task. Our architecture is based on a standard CNN design (a U‐Net variant). We combine Dice loss and weighted binary cross‐entropy to address class imbalance, a common challenge in medical segmentation tasks. To the best of our knowledge, this paper is the first to present an automated approach to MRI planimetry. Consequently, a direct comparison with previous work is not possible. Future work will focus on integrating more recent segmentation architectures into our framework, potentially replacing the current U‐Net backbone. Promising alternatives include nnU‐Net [49], the transformer‐based TransBTS [50], fine‐tuned versions of the Segment Anything Model (SAM) for medical data [51], and foundation models developed specifically for medical imaging tasks [52]. Table 8 presents an overview of the U‐Net and these recent models, summarizing their main features, advantages, and limitations.

Table 8.

Overview of possible segmentation architectures besides the proposed U‐Net.

Architecture Main feature Pros Cons
U‐Net CNN‐based encoder–decoder with skip connections Simple, efficient, well‐established in medical imaging May underperform on complex shapes or low‐contrast data
nnU‐Net [49] Self‐configuring U‐Net‐based pipeline No manual tuning needed; top performer in many benchmarks Requires high computational resources
TransBTS [50] Transformer‐augmented 3D segmentation Captures long‐range dependencies; better global context More complex training; slower inference
SAM (fine‐tuned) [51] Pretrained on vast general datasets; adaptable Zero‐shot/few‐shot capability; strong generalization Requires fine‐tuning for medical data; not domain‐specific
MedSAM [52] Foundation model for medical segmentation Optimized for medical images; transferable across tasks Still under validation; limited fine‐grained control

Future work may incorporate advanced loss functions that integrate geometric priors, such as boundary‐aware or surface‐based losses, to reduce segmentation drift at structure interfaces. Adding explicit anatomical constraints (e.g., enforcing smoothness, convexity, or topology preservation) could further enhance robustness. In addition, exploring transformer‐enhanced architectures such as TransBTS, Swin‐Unet, or foundation model–based approaches like MedSAM may improve generalization, particularly for out‐of‐distribution cases. These architectures offer stronger global feature modeling and may reduce reliance on high spatial resolution. To ensure clinical scalability, future studies should also validate the pipeline on multicenter datasets with diverse scanner types, acquisition protocols, and field strengths, allowing assessment of how well the method transfers to real‐world imaging environments.

Author Contributions

S.M.: funding acquisition, conceptualization, supervision, methodology, investigation, and writing—original draft; D.S.: funding acquisition, conceptualization, methodology, supervision, and writing—review and editing; M.S.: methodology, software development, experimental validation, formal analysis, statistical analysis, and writing—original draft; M.T.: methodology, software development, experimental validation, formal analysis, statistical analysis, and writing—original draft; F.D.: data curation and writing—review; M.H.: conceptualization, methodology, supervision, resources, and writing—review and editing (major restructuring); E.R.G.: conceptualization, methodology, supervision, resources, and writing—review and editing.

Funding

This work was funded by Medizinische Universität Innsbruck (10.13039/501100009980) and Universität Innsbruck (10.13039/501100012163). Open Access funding provided by Universitat Innsbruck/KEMÖ

Conflicts of Interest

The authors declare no conflicts of interest.

Acknowledgments

The authors acknowledge support in the context of the Jubiläumsfond der Universität Innsbruck and the Medizinischen Universität Innsbruck zur Förderung wissenschaftlicher Kooperationsprojekte.

Mangesius, Stephanie , Schiefeneder, Daniela , Schwab, Matthias , Tiefenthaler, Markus , Deisenhammer, Florian , Haltmeier, Markus , Gizewski, Elke R. , Deep Learning Framework for Automated MRI Planimetry in Multiple Sclerosis, International Journal of Biomedical Imaging, 2026, 4456355, 15 pages, 2026. 10.1155/ijbi/4456355

Academic Editor: Irfan Ahmad

Contributor Information

Markus Haltmeier, Email: markus.haltmeier@uibk.ac.at.

Irfan Ahmad, Email: iahmad@wiley.com.

Data Availability Statement

Research data are not shared.

References

  • 1. Thompson A. J., Baranzini S. E., Geurts J., Hemmer B., and Ciccarelli O., Multiple Sclerosis, Lancet. (2018) 391, no. 10130, 1622–1636, 10.1016/S0140-6736(18)30481-1, 2-s2.0-85044296592. [DOI] [PubMed] [Google Scholar]
  • 2. Lassmann H., The Changing Concepts in the Neuropathology of Acquired Demyelinating Central Nervous System Disorders, Current Opinion in Neurology. (2019) 32, no. 3, 313–319, 10.1097/WCO.0000000000000685, 2-s2.0-85065346296. [DOI] [PubMed] [Google Scholar]
  • 3. Brex P. A., Jenkins R., Fox N. C., Crum W. R., O′Riordan J. I., Plant G. T., and Miller D. H., Detection of Ventricular Enlargement in Patients at the Earliest Clinical Stage of MS, Neurology. (2000) 54, no. 8, 1689–1691, 10.1212/WNL.54.8.1689, 2-s2.0-0034009929, 10762518. [DOI] [PubMed] [Google Scholar]
  • 4. Dalton C. M., Brex P. A., Jenkins R., Fox N. C., Miszkiel K. A., Crum W. R., O′Riordan J. I., Plant G. T., Thompson A. J., and Miller D. H., Progressive Ventricular Enlargement in Patients With Clinically Isolated Syndromes is Associated With the Early Development of Multiple Sclerosis, Journal of Neurology, Neurosurgery & Psychiatry. (2002) 73, no. 2, 141–147, 10.1136/jnnp.73.2.141, 2-s2.0-0036323343, 12122170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Matthews P. M., Brain Imaging of Multiple Sclerosis: The Next 10 Years, Neuroimaging Clinics of North America. (2009) 19, no. 1, 101–112, 10.1016/j.nic.2008.08.003, 2-s2.0-57049185774, 19064203. [DOI] [PubMed] [Google Scholar]
  • 6. Popescu V., Klaver R., Voorn P., Galis-de Graaf Y., Knol D. L., Twisk J. W. R., Versteeg A., Schenk G. J., van der Valk P., Barkhof F., de Vries H. E., Vrenken H., and Geurts J. J., What Drives MRI-Measured Cortical Atrophy in Multiple Sclerosis?, Multiple Sclerosis Journal. (2015) 21, no. 10, 1280–1290, 10.1177/1352458514562440, 2-s2.0-84926449661, 25583833. [DOI] [PubMed] [Google Scholar]
  • 7. Trapp B. D., Peterson J., Ransohoff R. M., Rudick R., Mörk S., and Bö L., Axonal Transection in the Lesions of Multiple Sclerosis, New England Journal of Medicine. (1998) 338, no. 5, 278–285, 10.1056/NEJM199801293380502, 2-s2.0-0032576752, 9445407. [DOI] [PubMed] [Google Scholar]
  • 8. Chung K. K., Altmann D., Barkhof F., Miszkiel K., Brex P. A., O′Riordan J., Ebner M., Prados F., Cardoso M. J., Vercauteren T., Ourselin S., Thompson A., Ciccarelli O., and Chard D. T., A 30-Year Clinical and Magnetic Resonance Imaging Observational Study of Multiple Sclerosis and Clinically Isolated Syndromes, Annals of Neurology. (2020) 87, no. 1, 63–74, 10.1002/ana.25637, 31693200. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Mangesius S., Hussl A., Krismer F., Mahlknecht P., Reiter E., Tagwercher S., Djamshidian A., Schocke M., Esterhammer R., Wenning G., Müller C., Scherfler C., Gizewski E. R., Poewe W., and Seppi K., MR Planimetry in Neurodegenerative Parkinsonism Yields High Diagnostic Accuracy for PSP, Parkinsonism & Related Disorders. (2018) 46, 47–55, 10.1016/j.parkreldis.2017.10.020, 2-s2.0-85034440890, 29126761. [DOI] [PubMed] [Google Scholar]
  • 10. Quattrone A., Morelli M., Bianco M. G., Buonocore J., Sarica A., Caligiuri M. E., Aracri F., Calomino C., de Maria M., Vaccaro M. G., Gramigna V., Augimeri A., Vescio B., and Quattrone A., Magnetic Resonance Planimetry in the Differential Diagnosis Between Parkinson′s Disease and Progressive Supranuclear Palsy, Brain Sciences. (2022) 12, no. 7, 10.3390/brainsci12070949. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Mangesius S., Haider L., Lenhart L., Steiger R., Prados Carrasco F., Scherfler C., and Gizewski E. R., Qualitative and Quantitative Comparison of Hippocampal Volumetric Software Applications: Do All Roads Lead to Rome?, Biomedicine. (2022) 10, no. 2, 10.3390/biomedicines10020432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Ariz M., Abad R. C., Castellanos G., Martínez M., Muñoz-Barrutia A., Fernández-Seara M. A., Pastor P., Pastor M. A., and Ortiz-de-Solórzano C., Dynamic Atlas-Based Segmentation and Quantification of Neuromelanin-Rich Brainstem Structures in Parkinson Disease, IEEE Transactions on Medical Imaging. (2019) 38, no. 3, 813–823, 10.1109/TMI.2018.2872852, 2-s2.0-85054417677, 30281440. [DOI] [PubMed] [Google Scholar]
  • 13. Castellanos G., Fernández-Seara M. A., Lorenzo-Betancor O., Ortega-Cubero S., Puigvert M., Uranga J., Vidorreta M., Irigoyen J., Lorenzo E., Muñoz-Barrutia A., Ortiz-de-Solorzano C., Pastor P., and Pastor M. A., Automated Neuromelanin Imaging as a Diagnostic Biomarker for Parkinson′s Disease, Movement Disorders. (2015) 30, no. 7, 945–952, 10.1002/mds.26201, 2-s2.0-84931955307, 25772492. [DOI] [PubMed] [Google Scholar]
  • 14. Haegelen C., Coupé P., Fonov V., Guizard N., Jannin P., Morandi X., and Collins D. L., Automated Segmentation of Basal Ganglia and Deep Brain Structures in MRI of Parkinson′s Disease, International Journal of Computer Assisted Radiology and Surgery. (2013) 8, no. 1, 99–110, 10.1007/s11548-012-0675-8, 2-s2.0-84871940380, 22426551. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Li B., Jiang C., Li L., Zhang J., and Meng D., Automated Segmentation and Reconstruction of the Subthalamic Nucleus in Parkinson′s Disease Patients, Interface. (2016) 19, no. 1, 13–19, 10.1111/ner.12350, 2-s2.0-84956780870. [DOI] [PubMed] [Google Scholar]
  • 16. Papinutto N., Bakshi R., Bischof A., Calabresi P. A., Caverzasi E., Constable R. T., Datta E., Kirkish G., Nair G., Oh J., Pelletier D., Pham D. L., Reich D. S., Rooney W., Roy S., Schwartz D., Shinohara R. T., Sicotte N. L., Stern W. A., Tagge I., Tauhid S., Tummala S., Henry R. G., and North American Imaging in Multiple Sclerosis Cooperative(NAIMS), Gradient Nonlinearity Effects on Upper Cervical Spinal Cord Area Measurement From 3D T1-Weighted Brain MRI Acquisitions, Magnetic Resonance in Medicine. (2018) 79, no. 3, 1595–1601, 10.1002/mrm.26776, 2-s2.0-85020726579, 28617996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Salvatore C., Cerasa A., Castiglioni I., Gallivanone F., Augimeri A., Lopez M., Arabia G., Morelli M., Gilardi M. C., and Quattrone A., Machine Learning on Brain MRI Data for Differential Diagnosis of Parkinson′s Disease and Progressive Supranuclear Palsy, Journal of Neuroscience Methods. (2014) 222, 230–237, 10.1016/j.jneumeth.2013.11.016, 2-s2.0-84890764457, 24286700. [DOI] [PubMed] [Google Scholar]
  • 18. Visser E., Keuken M. C., Douaud G., Gaura V., Bachoud-Levi A.-C., Remy P., Forstmann B. U., and Jenkinson M., Automatic Segmentation of the Striatum and Globus Pallidus Using MIST: Multimodal Image Segmentation Tool, Neuro Image. (2016) 125, 479–497, 10.1016/j.neuroimage.2015.10.013, 2-s2.0-84946228897, 26477650. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Zhan G., Wang D., Cabezas M., Bai L., Kain K., and Wang C., Learning From Pseudo-Labels: Deep Networks Improve Consistency in Longitudinal Brain Volume Estimation, Frontiers in Neuroscience. (2023) 17, 1196087, 10.3389/fnins.2023.1196087. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Sander L., Pezold S., Andermatt S., Amann M., Meier D., Wendebourg M. J., Sinnecker T., Radue E. W., Naegelin Y., Granziera C., Kappos L., Wuerfel J., Cattin P., Schlaeger R., and for the Alzheimer′s Disease Neuroimaging Initiative, Accurate, Rapid, and Reliable Fully Automated MRI Brainstem Segmentation for Application in Multiple Sclerosis and Neurodegenerative Diseases, Human Brain Mapping. (2019) 40, no. 14, 4091–4104, 10.1002/hbm.24687, 2-s2.0-85067489292. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Bocchetta M., Iglesias J. E., Chelban V., Jabbari E., Lamb R., Russell L. L., Greaves C. V., Neason M., Cash D. M., Thomas D. L., Warren J. D., Woodside J., Houlden H., Morris H. R., and Rohrer J. D., Automated Brainstem Segmentation Detects Differential Involvement in Atypical Parkinsonian Syndromes, Journal of Movement Disorders. (2020) 13, no. 1, 39–46, 10.14802/jmd.19030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Vieira S., Pinaya W. H. L., and Mechelli A., Using Deep Learning to Investigate the Neuroimaging Correlates of Psychiatric and Neurological Disorders: Methods and Applications, Neuroscience & Biobehavioral Reviews. (2017) 74, pt A, 58–75, 10.1016/j.neubiorev.2017.01.002, 2-s2.0-85009921542, 28087243. [DOI] [PubMed] [Google Scholar]
  • 23. Yan W., Qu G., Hu W., Abrol A., Cai B., Qiao C., Plis S. M., Wang Y. P., Sui J., and Calhoun V. D., Deep Learning in Neuroimaging: Promises and Challenges, IEEE Signal Processing Magazine. (2022) 39, no. 2, 87–98, 10.1109/MSP.2021.3128348. [DOI] [Google Scholar]
  • 24. Belwal P. and Singh S., Deep Learning Techniques to Detect and Analysis of Multiple Sclerosis Through MRI: A Systematic Literature Review, Computers in Biology and Medicine. (2025) 185, 109530, 10.1016/j.compbiomed.2024.109530. [DOI] [PubMed] [Google Scholar]
  • 25. Belwal P. and Singh S., Multiple Sclerosis Lesion Segmentation Using MRI: U-Net Approach, Procedia Computer Science. (2025) 260, 513–522, 10.1016/j.procs.2025.03.229. [DOI] [Google Scholar]
  • 26. Belwal P. and Singh S., U-Net Approach to MRI-Based Segmentation of Multiple Sclerosis Lesions, Procedia Computer Science. (2025) 259, 1316–1325, 10.1016/j.procs.2025.04.086. [DOI] [Google Scholar]
  • 27. Mangesius S., Krismer F., Gizewski E. R., Müller C., Hussl A., Schocke M., Scherfler C., Poewe W., and Seppi K., 1.5 Versus 3 Tesla Magnetic Resonance Planimetry in Neurodegenerative Parkinsonism, Movement Disorders: Official Journal of the Movement Disorder Society. (2016) 31, no. 12, 1925–1927, 10.1002/mds.26823, 2-s2.0-84999751882, 27779783. [DOI] [PubMed] [Google Scholar]
  • 28. Heim B., Mangesius S., Krismer F., Wenning G. K., Hussl A., Scherfler C., Gizewski E. R., Schocke M., Esterhammer R., Quattrone A., Poewe W., and Seppi K., Diagnostic Accuracy of MR Planimetry in Clinically Unclassifiable Parkinsonism, Parkinsonism & Related Disorders. (2021) 82, 87–91, 10.1016/j.parkreldis.2020.11.019, 33271461. [DOI] [PubMed] [Google Scholar]
  • 29. Mangesius S., Mariotto S., Ferrari S., Pereverzyev S Jr, Lerchner H., Haider L., Gizewski E. R., Wenning G., Seppi K., Reindl M., and Poewe W., Novel Decision Algorithm to Discriminate Parkinsonism With Combined Blood and Imaging Biomarkers, Parkinsonism & Related Disorders. (2020) 77, 57–63, 10.1016/j.parkreldis.2020.05.033, 32622301. [DOI] [PubMed] [Google Scholar]
  • 30. Liu Y., Collins R. T., and Rothfus W. E., Robust Midsagittal Plane Extraction From Normal and Pathological 3-D Neuroradiology Images, IEEE Transactions on Medical Imaging. (2001) 20, no. 3, 175–192, 10.1109/42.918469, 2-s2.0-0035263302, 11341708. [DOI] [PubMed] [Google Scholar]
  • 31. Rehman H. Z. U. and Lee S., An Efficient Automatic Midsagittal Plane Extraction in Brain MRI, Applied Sciences. (2018) 8, no. 11, 10.3390/app8112203, 2-s2.0-85056229484. [DOI] [Google Scholar]
  • 32. Tan W., Kang Y., Dong Z., Chen C., Yin X., Su Y., Zhang Y., Zhang L., and Xu L., An Approach to Extraction Midsagittal Plane of Skull From Brain CT Images for Oral and Maxillofacial Surgery, IEEE Access. (2019) 7, 118203–118217, 10.1109/ACCESS.2019.2920862. [DOI] [Google Scholar]
  • 33. Volkau I., Prakash K. N. B., Ananthasubramaniam A., Aziz A., and Nowinski W. L., Extraction of the Midsagittal Plane From Morphological Neuroimages Using the Kullback–Leibler′s Measure, Medical Image Analysis. (2006) 10, no. 6, 863–874, 10.1016/j.media.2006.07.005, 2-s2.0-33751314547, 16997609. [DOI] [PubMed] [Google Scholar]
  • 34. Ronneberger O., Fischer P., and Brox T., U-Net: Convolutional Networks for Biomedical Image Segmentation, Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015, Springer, 234–241, 10.1007/978-3-319-24574-4_28, 2-s2.0-84951834022. [DOI] [Google Scholar]
  • 35. Milletari F., Navab N., and Ahmadi S. A., V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation, Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), 2016, IEEE, 565–571, 10.1109/3DV.2016.79, 2-s2.0-85011298810. [DOI] [Google Scholar]
  • 36. Khened G. K. M. and Varghese A., Fully Convolutional Multi-Scale Residual DenseNets for Cardiac Segmentation and Automated Cardiac Diagnosis Using Ensemble of Classifiers, Medical Image Analysis. (2019) 51, 21–45, 10.1016/j.media.2018.10.004, 2-s2.0-85055642314, 30390512. [DOI] [PubMed] [Google Scholar]
  • 37. Gudbjartsson H. and Patz S., The Rician Distribution of Noisy MRI Data, Magnetic Resonance in Medicine. (1995) 34, no. 6, 910–914, 10.1002/mrm.1910340618, 2-s2.0-0028839871, 8598820. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Eshaghi A., Prados F., Brownlee W. J., Altmann D. R., Tur C., Cardoso M. J., de Angelis F., van de Pavert S., Cawley N., de Stefano N., Stromillo M. L., Battaglini M., Ruggieri S., Gasperini C., Filippi M., Rocca M. A., Rovira A., Sastre-Garriga J., Vrenken H., Leurs C. E., Killestein J., Pirpamer L., Enzinger C., Ourselin S., Wheeler-Kingshott C. A. M. G., Chard D., Thompson A. J., Alexander D. C., Barkhof F., Ciccarelli O., and MAGNIMS study group, Deep Gray Matter Volume Loss Drives Disability Worsening in Multiple Sclerosis, Annals of Neurology. (2018) 83, no. 2, 210–222, 10.1002/ana.25145, 2-s2.0-85042264468, 29331092. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Lee C. Y., Mak H. K. F., Chiu P. W., Chang H. C., Barkhof F., and Chan K. H., Differential Brainstem Atrophy Patterns in Multiple Sclerosis and Neuromyelitis Optica Spectrum Disorders, Journal of Magnetic Resonance Imaging. (2018) 47, no. 6, 1601–1609, 10.1002/jmri.25866, 2-s2.0-85047786895. [DOI] [PubMed] [Google Scholar]
  • 40. Andravizou A., Dardiotis E., Artemiadis A., Sokratous M., Siokas V., Tsouris Z., Aloizou A. M., Nikolaidis I., Bakirtzis C., Tsivgoulis G., Deretzi G., Grigoriadis N., Bogdanos D. P., and Hadjigeorgiou G. M., Brain Atrophy in Multiple Sclerosis: Mechanisms, Clinical Relevance and Treatment Options, Auto Immun Highlights. (2019) 10, no. 1, 10.1186/s13317-019-0117-5, 32257063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Sastre-Garriga J., Pareto D., Battaglini M., Rocca M. A., Ciccarelli O., Enzinger C., Wuerfel J., Sormani M. P., Barkhof F., Yousry T. A., de Stefano N., Tintoré M., Filippi M., Gasperini C., Kappos L., Río J., Frederiksen J., Palace J., Vrenken H., Montalban X., Rovira À., and MAGNIMS study group, MAGNIMS Consensus Recommendations on the Use of Brain and Spinal Cord Atrophy Measures in Clinical Practice, Nature Reviews Neurology. (2020) 16, no. 3, 171–182, 10.1038/s41582-020-0314-x, 32094485. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Biberacher V., Schmidt P., Keshavan A., Boucard C. C., Righart R., Sämann P., Preibisch C., Fröbel D., Aly L., Hemmer B., Zimmer C., Henry R. G., and Mühlau M., Intra- and Interscanner Variability of Magnetic Resonance Imaging-Based Volumetry in Multiple Sclerosis, NeuroImage. (2016) 142, 188–197, 10.1016/j.neuroimage.2016.07.035, 2-s2.0-84994335146. [DOI] [PubMed] [Google Scholar]
  • 43. De Stefano N., Silva D. G., and Barnett M. H., Effect of Fingolimod on Brain Volume Loss in Patients With Multiple Sclerosis, CNS Drugs. (2017) 31, no. 4, 289–305, 10.1007/s40263-017-0415-2, 2-s2.0-85014033727, 28247239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Narayana P. A., Coronado I., Sujit S. J., Wolinsky J. S., Lublin F. D., and Gabr R. E., Deep Learning for Predicting Enhancing Lesions in Multiple Sclerosis From Noncontrast MRI, Radiology. (2020) 294, no. 2, 398–404, 10.1148/radiol.2019191061. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Zou K. H., Warfield S. K., Bharatha A., Tempany C. M. C., Kaus M. R., Haker S. J., Wells W. M., Jolesz F. A., and Kikinis R., Statistical Validation of Image Segmentation Quality Based on a Spatial Overlap index, Academic Radiology. (2004) 11, no. 2, 178–189, 14974593, 10.1016/S1076-6332(03)00671-8, 2-s2.0-0942265691. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Rocca M. A., Battaglini M., Benedict R. H., De Stefano N., Geurts J. J., Henry R. G., Horsfield M. A., Jenkinson M., Pagani E., and Filippi M., Brain MRI Atrophy Quantification in MS: From Methods to Clinical Application, Neurology. (2017) 92, no. 12, 403–413, 10.1212/WNL.0000000000003542, 2-s2.0-85010875136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Geraldes R., Ciccarelli O., Barkhof F., De Stefano N., Enzinger C., Filippi M., Hofer M., Paul F., Preziosa P., Rovira A., and DeLuca G. C., The Current Role of MRI in Differentiating Multiple Sclerosis From Its Imaging Mimics, Nature Reviews Neurology. (2020) 16, no. 4, 198–213, 10.1038/nrneurol.2018.14, 2-s2.0-85044502766. [DOI] [PubMed] [Google Scholar]
  • 48. Nakagawa Y. and Rosenfeld A., A Note on Polygonal and Elliptical Approximation of Mechanical Parts, Pattern Recognition. (1979) 11, no. 2, 133–142, 10.1016/0031-3203(79)90059-1, 2-s2.0-0018729373. [DOI] [Google Scholar]
  • 49. Meys K. M. E., de Vries L. S., Groenendaal F., Vann S. D., and Lequin M. H., The Mammillary Bodies: A Review of Causes of Injury in Infants and Children, American Journal of Neuroradiology. (2022) 43, no. 6, 802–812, 10.3174/ajnr.A7463, 35487586. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Falahati F., Westman E., and Simmons A., Multivariate Data Analysis and Machine Learning in Alzheimer′s Disease With a Focus on Structural Magnetic Resonance Imaging, Journal of Alzheimer′s Disease. (2014) 41, no. 3, 685–708, 10.3233/JAD-131928, 2-s2.0-84921491253. [DOI] [PubMed] [Google Scholar]
  • 51. Kalincik T., Manouchehrinia A., Sobisek L., Jokubaitis V., Spelman T., Horakova D., Havrdova E., Trojano M., Izquierdo G., Lugaresi A., and Girard M., Towards Personalized Therapy for Multiple Sclerosis: Prediction of Individual Treatment Response, Brain. (2017) 140, no. 9, 2426–2443, 10.1093/brain/awx185, 2-s2.0-85031816766. [DOI] [PubMed] [Google Scholar]
  • 52. Rathmann E., Hemkemeier P., Raths S., Grothe M., Mankertz F., Hosten N., and Flessa S., Changes in MRI Workflow of Multiple Sclerosis After Introduction of an AI-Software: A Qualitative Study, Healthcare. (2024) 12, no. 10, 10.3390/healthcare12100978. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Gros C., De Leener B., Badji A., Maranzano J., Eden D., Dupont S. M., Talbott J., Zhuoquiong R., Liu Y., Granberg T., and Ouellette R., Automatic Segmentation of the Spinal Cord and Intramedullary Multiple Sclerosis Lesions With Convolutional Neural Networks, Neuroimage. (2019) 184, 901–915, 10.1016/j.neuroimage.2018.09.081, 2-s2.0-85054688847, 30300751. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Dahan A., Pereira R., Malpas C. B., Kalincik T., and Gaillard F., PACS Integration of Semiautomated Imaging Software Improves Day-to-Day MS Disease Activity Detection, American Journal of Neuroradiology. (2019) 40, no. 10, 1624–1629, 10.3174/ajnr.A6195, 2-s2.0-85073082573, 31515214. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Dikici E., Bigelow M., Prevedello L. M., White R. D., and Erdal B. S., Integrating AI Into Radiology Workflow: Levels of Research, Production, and Feedback Maturity, Journal of Medical Imaging. (2020) 7, no. 1, 016502, 10.1117/1.JMI.7.1.016502, 32064302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Jorg T., Halfmann M. C., Stoehr F., Arnhold G., Theobald A., Mildenberger P., and Müller L., A Novel Reporting Workflow for Automated Integration of Artificial Intelligence Results Into Structured Radiology Reports, Insights Into Imaging. (2024) 15, no. 1, 10.1186/s13244-024-01660-5, 38502298. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Research data are not shared.


Articles from International Journal of Biomedical Imaging are provided here courtesy of Wiley

RESOURCES