Skip to main content
PLOS One logoLink to PLOS One
. 2025 Aug 14;20(8):e0330044. doi: 10.1371/journal.pone.0330044

Predicting achievement of clinical goals using machine learning in myasthenia gravis

Hiroyuki Akamine 1,¤a, Akiyuki Uzawa 1,*, Satoshi Kuwabara 1, Shigeaki Suzuki 2,¤b, Yosuke Onishi 1, Manato Yasuda 1, Yukiko Ozawa 1, Naoki Kawaguchi 3, Tomoya Kubota 4, Masanori P Takahashi 4, Yasushi Suzuki 5, Genya Watanabe 5, Takashi Kimura 6, Takamichi Sugimoto 7, Makoto Samukawa 8, Naoya Minami 9, Masayuki Masuda 10, Shingo Konno 11, Yuriko Nagane 12, Kimiaki Utsugisawa 12
Editor: Pinyi Lu13
PMCID: PMC12352761  PMID: 40811489

Abstract

Background

Myasthenia Gravis (MG) is an autoimmune disease characterized by the production of autoantibodies against neuromuscular junctions, leading to varying degrees of severity and outcomes among patients. This variability makes clinical evaluation crucial for determining appropriate treatment targets. However, accurately assessing Minimal Manifestation (MM) status is challenging, requiring expertise in MG management. Therefore, this study aims to develop a diagnostic model for MM in MG patients by leveraging their clinical scores and machine learning approaches.

Methods

This study included 1,603 MG patients enrolled from the Japan MG Registry in the 2021 survey. We employed non-negative matrix factorization to decompose three MG clinical scores (MG composite score, MGADL scale, and MG quality of life (QOL) 15r) into four distinct modules: Diplopia, Ptosis, Systemic symptoms, and QOL. We developed a machine learning model with the four modules to predict MM or better status in MG patients. Using 414 registrants from the Japan MG Registry in the 2015 survey, we validated the model’s performance using various metrics, including area under the receiver operating characteristic curve (AUROC), sensitivity, specificity, accuracy, F1 score, and Matthews Correlation Coefficient (MCC).

Results

The ensemble model achieved an AUROC of 0.94 (95% CI: 0.94–0.94), accuracy of 0.87 (95% CI: 0.86–0.88), sensitivity of 0.85 (95% CI: 0.85–0.86), specificity of 0.89 (95% CI: 0.88–0.91), precision of 0.93 (95% CI: 0.92–0.94), F1 score of 0.89 (95% CI: 0.88–0.89), and MCC of 0.74 (95% CI: 0.72–0.75) on the validation dataset.

Conclusions

The developed MM diagnostic model can effectively predict MM or better status in MG patients, potentially guiding clinicians in determining treatment objectives and evaluating treatment outcomes.

Introduction

Myasthenia Gravis (MG) is an autoimmune disease that produces autoantibodies targeting acetylcholine receptors (AChR) at neuromuscular junctions [1]. Approximately 80% of patients have antibodies against AChR, while 5–8% produce muscle-specific kinase (MuSK) antibodies [2]. Around 10% of MG cases are called double seronegative MG, as they do not have detectable levels of either pathogenic antibodies [3]. MG is a disease characterized by muscle weakness affecting various parts of the body, leading to symptoms such as diplopia, ptosis, dysarthria, dysphagia, respiratory distress, limb weakness, and fatigue [1]. A notable feature of this disease is the fluctuation of these symptoms within a day or across different days. Several standardized scales are widely used to evaluate the clinical symptoms of MG patients: the quantitative MG (QMG) scale [4] for objective assessment of muscle weakness severity, the MG activities of daily living (MGADL) score [5] for evaluating functional limitations in daily activities, and the MG composite (MGC) scale [6] as a comprehensive measure combining both objective and subjective assessments. MG patients often experience not only motor symptoms but also difficulties in psychological and social well-being [7]. However, it is challenging to assess these burdens. MG Quality of Life 15 (MGQOL15) [8] and its revised version (MGQOL15r) [9] are able to capture a more holistic, patient-centered assessment [10]. The MG Foundation of America (MGFA) classification is widely used to evaluate treatment outcomes by categorizing pre-treatment conditions into five stages and post-treatment status into eight stages [11]. In clinical practice, many physicians set Minimal Manifestation (MM) status as a treatment goal based on international guidelines [12]. Jaretzki III et al. defined “MM status: The patient has no symptoms or functional limitations from MG but has some weakness on examination of some muscles” [11]. This ambiguity makes it challenging to establish clear cutoff values for defining MM in MG clinical scores. There is no clear definition to identify minimal muscle weakness that is clinically important, which can result in variability in how physicians interpret and apply the MM criteria. This inconsistency may lead to different treatment approaches and creates challenges in comparing outcomes across medical institutes. By providing visualization of MG symptoms in MM determination and standardizing the assessment process, our approach aims to support more unified treatment selection criteria and help identify unmet clinical needs in MG management. We aimed to develop a diagnostic model for MM using MG clinical scores from patients enrolled in the MG registry, with the goal of standardizing MM assessment.

Methods

Data collection and access

The Japan MG Registry 2021 survey [13] was conducted between April and October 2021 across 13 medical institutions, which collaboratively gathered clinical data from MG patients. Written informed consent was obtained from all participants. Each participant was assigned a study-specific identification number by one (Y.N.) of co-authors, and personally identifiable information was anonymized. The first author (H.A.) accessed the dataset on December 10, 2021. The clinical assessments were performed by neurologists with extensive experience in MG diagnosis and management.

Data preprocessing

A total of 1,603 MG patients were included in the analysis after excluding cases with missing data from the 1,710 MG patients enrolled in the Japan MG Registry in the 2021 survey [13]. Their clinical scores, including MGC scale, MGADL score, and MGQOL15r, were collected. The evaluation of these clinical scores, as well as the determination of whether a patient had reached MM or better status (MM, Complete Stable Remission, or Pharmacologic Remission), was conducted by neurologists with expertise in MG management. To ensure consistency in feature scaling across the dataset, numerical columns were normalized to a range between 0 and 1. We utilized clinical data from 414 patients in the 2015 survey as the validation dataset. This ensured temporal independence between the training and validation datasets, allowing for a more robust evaluation of model. Fig 1 shows an overall view of Methods.

Fig 1. Overview of research methods.

Fig 1

The outline of the research was visualized by classifying it into four categories: data collection, NMF, machine learning, and validation.

Non-Negative Matrix Factorization (NMF)

Following the initial data preprocessing, we applied NMF [14] to decompose the dataset into latent modules. NMF was selected due to its ability to produce interpretable, non-negative components that align with the underlying structure of the data. We initialized the NMF model with four components using the nndsvd initialization method and a maximum of 200 iterations to ensure convergence. The decomposition resulted in two matrices: the W matrix (representing the latent variables for each sample) and the H matrix (representing the loadings of each feature on these latent variables). Both matrices were saved for downstream analysis and clustering.

Hierarchical clustering and visualization

To further analyze the relationships between clinical features and their respective NMF-derived components, hierarchical clustering was performed on both the rows and columns of the H matrix. Using complete linkage and Euclidean distance as the metric, we reordered the H matrix based on the hierarchical clustering results. This reordered matrix was then visualized as a heatmap, which allowed us to visually inspect the clustering patterns and relationships between the clinical features and the four NMF modules.

Model training and cross-validation

The W matrix obtained from the NMF decomposition was used to train machine learning models for predicting patient outcomes, the binary classification of achieving MM or better status. The dataset was split using 5-fold cross-validation, implemented via the K Fold method from scikit-learn. For each fold, the training and test sets were extracted, and four machine learning models were trained: Support Vector Machine (SVM), Logistic Regression, Random Forest, and Naive Bayes. The models were optimized using GridSearchCV for SVM, Logistic Regression, and Random Forest, while Naive Bayes was trained directly without parameter tuning. Specifically, for SVM, parameters such as ‘C’, ‘kernel’, and ‘gamma’ were tuned; for Logistic Regression, different configurations of ‘penalty’, ‘C’, ‘fit_intercept’, and ‘solver’ were explored, including support for elastic net regularization; and for Random Forest, ‘n_estimators’, ‘max_depth’, ‘min_samples_split’, and ‘min_samples_leaf’ were optimized. For each dataset split, the best parameters and scores were recorded. The final performance was assessed based on the average Area Under the Receiver Operating Characteristic Curve (AUROC) across five iterations. 95% confidence intervals were calculated with each model from 5-fold cross-validation.

Model ensembling

To improve prediction accuracy, we used an ensemble learning approach with four models: SVM, Logistic Regression, Random Forest, and Naive Bayes. For performance evaluation including AUROC, we applied a soft voting strategy, averaging the predicted probabilities from all four models to generate the final prediction scores.

Validation on independent dataset

To evaluate the generalization capability of the trained models, an independent dataset from 414 registrants from the Japan MG Registry in the 2015 survey was introduced for validation. The baseline characteristics of the validation cohort are presented in S1 Table. The validation dataset was preprocessed similarly to the training data, including feature scaling and variable selection. Using the H matrix obtained from NMF on the training data, we calculated the W matrix for the validation set with multiplicative updates, preserving the latent structure. The trained models were then applied to the validation data, and predictions were made using a majority voting ensemble approach. We evaluated the model’s performance using various metrics, including AUROC, sensitivity, specificity, accuracy, F1 score, and Matthews Correlation Coefficient (MCC).

Ethical publication statement

We confirm that we have read the Journal’s position on issues involved in ethical publication and affirm that this report is consistent with those guidelines.

Ethics approval

This study was approved by the ethics committee of each neurological center (Institutional Review Board No. R3-1 at Hanamaki General Hospital, the primary investigating institute) and written informed consent was obtained from all study participants.

Results

Data profiles

The profiles of the 1,603 MG patients included in this study are shown in Table 1. A total of 942 patients from the 1,603 MG patients achieved MM or better. The patients who achieved MM or better were more often male (46% vs. 30%), had a higher age of onset (51 [36, 64] vs. 46 [33, 60] years), and tended to be ocular type (29% vs. 11%). Whereas patients with SNMG and MuSK MG were less likely to achieve MM or better. Median MG clinical scores [IQR] were 5 [4, 8] vs. 1 [0, 2] for MGADL, 7 [4, 11] vs. 1 [0, 3] for MGC, and 12 [7, 18] vs. 3 [1, 7] for MGQOL15r in the non MM group and the MM or better group, significantly lower scores in the MM group (p < 0.001). MGFA PIS was 7.5%, PR 11%, and MM 81% in the MM or better group. The median PSL dose [IQR] was 5 mg [0, 8] in the non-MM group and 2 mg [0, 5] in the MM group, significantly lower in the MM group (p < 0.001).

Table 1. The profiles of 1,603 MG patients included in the study.

Non MM MM or better p-value
N = 661 N = 942
Sex (Male) 197 (30%) 436 (46%) < 0.001
Age 58 (48, 72) 62 (49, 73) 0.014
Onset age 46 (33, 60) 51 (36, 64) < 0.001
MG subtypes < 0.001
 EOMG 148 (22%) 195 (21%)
 LOMG 135 (20%) 190 (20%)
 MuSK 30 (4.5%) 18 (1.9%)
 OMG 70 (11%) 275 (29%)
 SNMG 127 (19%) 56 (5.9%)
 TAMG 151 (23%) 208 (22%)
MGFA PIS < 0.001
 CSR 0 (0%) 71 (7.5%)
 PR 0 (0%) 106 (11%)
 MM 0 (0%) 765 (81%)
 E 9 (1.4%) 0 (0%)
 I 485 (73%) 0 (0%)
 U 138 (21%) 0 (0%)
 W 27 (4.1%) 0 (0%)
MGC total score 7 (4, 11) 1 (0, 3) < 0.001
MGADL total score 5 (4, 8) 1 (0, 2) < 0.001
MGQOL15r total score 12 (7, 18) 3 (1, 7) < 0.001
PSL dose 5.0 (0.0, 8.0) 2.0 (0.0, 5.0) < 0.001

The patients with MGFA PIS of CSR, PR, or MM at the time of the survey were defined as MM or better group; Median (IQR), n (%) is shown and statistical tests were performed with Wilcoxon rank sum test and Pearson’s Chi-squared test. minimal symptoms; MM, MG Foundation of America; MGFA, MG activities of daily living; MGADL, MG composite; MGC, MG quality of life 15 revised; MGQOL15r.

NMF decomposition and feature clustering

NMF was applied to the dataset of 1,603 patients, resulting in the decomposition of clinical features into four modules. The W matrix, which represents patient scores for latent variables, and the H matrix, which shows the contributions of clinical features to each latent variable, were generated. Hierarchical clustering of the H matrix revealed distinct patterns among the clinical features. The heatmap visualization of the reordered H matrix showed distinct patterns, with four main clusters corresponding to the four NMF modules. These modules were designated as “Module QOL,” “Module Diplopia,” “Module Ptosis,” and “Module Systemic,” based on the clinical features in each cluster (Fig 2). The clustering results demonstrated that NMF effectively captured the underlying structure of the data, with clear separations observed across features. Comparing the W matrix between the non-MM and MM or better groups revealed that all Module scores were significantly lower in the MM or better group (Fig 3), similar to the original MG clinical score.

Fig 2. The H matrix obtained from NMF decomposition.

Fig 2

The H matrix represents the four modules and their coefficients for each clinical score. Clinical parameters were clustered into modules defined as QOL, systemic symptoms, diplopia, and ptosis. The color of each cell indicates the coefficient value: cells with a red color have higher coefficient values, indicating that the clinical score is more strongly associated with that module.

Fig 3. The distribution of each module in the non-MM and MM or better groups.

Fig 3

The four modules of the two groups are compared in the boxplot. Each box represents the distribution of a feature for a group. The central line within each box indicates the median value, and the box’s length represents the interquartile range, giving an idea of the data’s spread. The statistical significance of the differences in distribution for each feature between the two groups were tested using the Mann−Whitney U test.

Model training and cross-validation

Using the W matrix from NMF, four machine learning models—SVM, Logistic Regression, Random Forest, and Naive Bayes—were trained to predict whether patients had achieved a state of MM or better. Hyperparameter tuning via GridSearchCV improved the performance of all models. After tuning, the Logistic Regression model achieved the highest AUROC score among the models (Fig 4, S1 and S2 Figs).

Fig 4. The Receiver Operating Characteristic (ROC) curves comparing the predictive performance of four models on the test dataset.

Fig 4

The ROC curves for each model are shown, with corresponding the area under the curve (AUC) presented in the legend. The colors of the lines represent different models: SVM (blue), logistic regression (yellow), random forest (green), and naïve Bayes (red). The optimal cutoff point for each model is marked with a black “×”. The dashed diagonal line represents the performance of a random classifier (AUC = 0.5).

Evaluation of ensemble learning model

The majority voting ensemble, which combined predictions from all four models, resulted in a further improvement in prediction performance. The ensemble model achieved an AUROC of 0.96 across the cross-validation folds, outperforming any individual model (S3 Fig). This demonstrated the strength of ensemble learning in aggregating the diverse strengths of the individual classifiers, leading to a more accurate and robust prediction. The generalization ability of the ensemble model was further evaluated using an independent dataset of 414 patients who had participated in the 2015 survey. The ensemble model achieved an AUROC of 0.94 (95% CI: 0.94–0.94), accuracy of 0.87 (95% CI: 0.86–0.88), sensitivity of 0.85 (95% CI: 0.85–0.86), specificity of 0.89 (95% CI: 0.88–0.91), precision of 0.93 (95% CI: 0.92–0.94), F1 score of 0.89 (95% CI: 0.88–0.89), and MCC of 0.74 (95% CI: 0.72–0.75) on the validation dataset, confirming its strong generalization performance (Fig 5). Furthermore, the ensemble model outperformed models based on individual clinical scores in predicting MM or better status, achieving an AUC of 0.94, compared to 0.92 for MGADL, 0.91 for MGC, and 0.85 for MGQOL15 (Fig 6). The baseline characteristics of the validation cohort are shown in S1 Table. The validation cohort demonstrated similar demographic and clinical patterns to the training dataset, with comparable distributions of MG subtypes and clinical severity scores. The consistently high AUROC values indicates that the model can reliably distinguish between patients who have achieved MM or better and those who have not, even when applied to a temporally independent dataset.

Fig 5. The Receiver Operating Characteristic (ROC) curve illustrating the performance of the ensemble model on the validation dataset.

Fig 5

The ROC curve (blue line) represents the relationship between sensitivity and 1-specificity at various threshold settings. The area under the curve (AUC) is shown in the lower right. The dashed diagonal line represents the performance of a random classifier (AUC = 0.5).

Fig 6. Comparison of ROC curves between the ensemble model and cutoff approaches for MM or better status prediction.

Fig 6

The ROC curves compare the performance of our ensemble machine learning model against traditional rule-based approaches using individual clinical scores. The ensemble model (blue line) achieved an AUC of 0.94, demonstrating superior discriminative ability compared to rule-based methods using MGADL alone (AUC = 0.92, green line), MGC alone (AUC = 0.91, orange line), and MGQOL15 alone (AUC = 0.85, red line). The diagonal dashed line represents the performance of a random classifier (AUC = 0.5).

Discussion

Assessing MG symptoms often relies on subjective patient reports and clinical judgment. Therefore, the appropriateness and validity of these clinical scores may be subject to personal bias, especially for physicians unfamiliar with MG practice. Even though clinical scores are applied as treatment benchmarks, the choice of suitable scores and the determination of definitive cutoff values still leave room for further discussion. Research on thresholds for MG clinical scores using the PASS (Patient-acceptable symptom state) have indicated that patients feel themselves to be in an acceptable state if they score within 0–2 on the MGADL scale, within 3 on the MGC, and within 8 on the MGQOL15 questionnaire [15]. Also, MGADL scores of 3 or higher were determined to be unsatisfactory status in a Swedish cohort study [16]. However, in clinical practice, we often find that it is not possible to reliably predict a state of MM or better using single score thresholds. For instance, there are cases where the MGADL score is below the threshold while the MGQOL15r score exceeds it, or the opposite can also occur. Additionally, especially with MGADL, using a simple clinical score cutoff can result in patients exceeding the threshold based solely on ocular symptoms. In other words, the cutoff for each clinical score, or the determination of MM or better based on a combination of scores, can often be inconsistent.

In our study, a large dataset containing 1603 cases assessed by physicians with expertise in MG management was analyzed. We employed NMF as an effective technique for understanding the relationships across our cases and various score parameters. NMF is a matrix decomposition method that decomposes our original matrix (V) of cases and score parameters into the product of two non-negative matrices (W and H), following the formula V ≈ WH. The resulting decomposition, represented by V ≈ WH, provides a concise and easily interpretable representation of the data and revealed potential structures within the data set. We have showed that MG clinical scores can potentially be categorized into Ptosis, Diplopia, Systemic symptoms, and Quality of Life components, and utilized these modules as features in machine learning to develop a model that integrates the patient’s clinical scores to determine if treatment goals; MM or better, have been achieved. Our model breaks down and visualizes the patient’s condition into four simple modules to support the physician’s judgment and also provides a prediction of MM or better based on our cohort.

The absence of established biomarkers in the clinical evaluation of MG presents a challenge in determining treatment effectiveness in clinical trials and daily clinical practice. Present assessments mainly rely on individual physician judgment. Implementing a model such as the one proposed would allow for referencing a shared evaluative standard, potentially simplifying the assessment process and fostering more consistent outcomes globally. Recent advancements in the treatment of MG have been remarkable. This progress has been marked by the development of a range of molecularly targeted drugs, including therapies against complement, neonatal FcRn, and CD20 [17]. By setting unified MM or better clinical decisions as the treatment objectives, we can guide the selection of these novel therapies for suitable patients, thus further improving patient outcomes.

This study has several limitations which should be acknowledged. It is essential to ensure that the criteria used in this model rely on reliable data since their accuracy depends on the data quality used in the training process. Also, this data reflects the practice of MG in Japan, as it was determined by physicians familiar with the practice of MG at multiple facilities in Japan. Our model may not be easily adapted as-is for different countries and regions, and the model might need to be retrained.

Conclusions

This study has presented a machine-learning model that provides a comprehensive and uniform method for assessing treatment success in MG patients. This model can reduce personal bias and promote more consistent treatment outcomes by utilizing dimensionality reduction techniques and integrating various clinical scores. As research and treatment for MG continue to advance, incorporating decision aids like this diagnostic model becomes increasingly important.

Supporting information

S1 Fig. ROC curves for machine learning model performance on training data.

ROC curves and AUC values for four machine learning models evaluated on training data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

(PDF)

pone.0330044.s001.pdf (491.9KB, pdf)
S2 Fig. ROC curves for machine learning model performance on test data.

ROC curves and AUC values for four machine learning models evaluated on test data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

(PDF)

pone.0330044.s002.pdf (461.4KB, pdf)
S3 Fig. ROC curves for machine learning model performance on validation data.

ROC curves and AUC values for four machine learning models evaluated on validation data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

(PDF)

pone.0330044.s003.pdf (454.8KB, pdf)
S1 Table. The profiles of 414 MG patients in validation data.

(CSV)

pone.0330044.s004.csv (624B, csv)

Acknowledgments

We would like to thank Dr. Hiroyuki Murai (Department of Neurology, International University of Health and Welfare, Narita) for the collection of data from patients.

Data Availability

As stated in our Data Sharing Statement, the clinical data are not publicly available due to privacy concerns. However, data may be available upon reasonable request and subject to review within the MGR group to establish the viability of sharing. To request access to the data, please contact the Ethics Committee of Hanamaki General Hospital (4-56 Odaya-chō, Hanamaki, Iwate 025-0082, Japan; phone: +81-198-23-3311), or the Japan MG Registry website (http://mg.kenkyuukai.jp/information/information_detail.asp?id=67084).

Funding Statement

This work was partly supported by the Health and Labor Sciences Research Grant on Intractable Diseases (Neuroimmunological Diseases) from the Ministry of Health, Labor, and Welfare of Japan (23FC1009). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Gilhus NE. Myasthenia Gravis. N Engl J Med. 2016;375(26):2570–81. doi: 10.1056/NEJMra1602678 [DOI] [PubMed] [Google Scholar]
  • 2.Carr AS, Cardwell CR, McCarron PO, McConville J. A systematic review of population-based epidemiological studies in myasthenia gravis. BMC Neurology. 2010;10:46. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Cortés-Vicente E, Gallardo E, Martínez MÁ, et al. Clinical characteristics of patients with double-seronegative myasthenia gravis and antibodies to cortactin. JAMA Neurol. 2016;73(9):1099–104. [DOI] [PubMed] [Google Scholar]
  • 4.Barohn RJ, McIntire D, Herbelin L, Wolfe GI, Nations S, Bryan WW. Reliability testing of the quantitative myasthenia gravis score. Ann N Y Acad Sci. 1998;841:769–72. doi: 10.1111/j.1749-6632.1998.tb11015.x [DOI] [PubMed] [Google Scholar]
  • 5.Wolfe GI, Herbelin L, Nations SP, Foster B, Bryan WW, Barohn RJ. Myasthenia gravis activities of daily living profile. Neurology. 1999;52(7):1487–9. doi: 10.1212/wnl.52.7.1487 [DOI] [PubMed] [Google Scholar]
  • 6.Burns TM. The MG composite: an outcome measure for myasthenia gravis for use in clinical trials and everyday practice. Ann N Y Acad Sci. 2012;1274:99–106. doi: 10.1111/j.1749-6632.2012.06812.x [DOI] [PubMed] [Google Scholar]
  • 7.Hoffmann S, Ramm J, Grittner U, Kohler S, Siedler J, Meisel A. Fatigue in myasthenia gravis: risk factors and impact on quality of life. Brain Behav. 2016;6(10):e00538. doi: 10.1002/brb3.538 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Burns TM, Grouse CK, Wolfe GI, Conaway MR, Sanders DB, MG Composite and MG-OL15 Study Group. The MG-QOL15 for following the health-related quality of life of patients with myasthenia gravis. Muscle Nerve. 2011;43(1):14–8. doi: 10.1002/mus.21883 [DOI] [PubMed] [Google Scholar]
  • 9.Burns TM, Sadjadi R, Utsugisawa K, Gwathmey KG, Joshi A, Jones S, et al. International clinimetric evaluation of the MG-QOL15, resulting in slight revision and subsequent validation of the MG-QOL15r. Muscle Nerve. 2016;54(6):1015–22. doi: 10.1002/mus.25198 [DOI] [PubMed] [Google Scholar]
  • 10.Diez Porras L, Homedes C, Alberti MA, Velez Santamaria V, Casasnovas C. Quality of life in myasthenia gravis and correlation of MG-QOL15 with other functional scales. J Clin Med Res. 2022;11(8). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Jaretzki A 3rd, Barohn RJ, Ernstoff RM, Kaminski HJ, Keesey JC, Penn AS, et al. Myasthenia gravis: recommendations for clinical research standards. Task Force of the Medical Scientific Advisory Board of the Myasthenia Gravis Foundation of America. Neurology. 2000;55(1):16–23. doi: 10.1212/wnl.55.1.16 [DOI] [PubMed] [Google Scholar]
  • 12.Sanders DB, Wolfe GI, Benatar M, et al. International consensus guidance for management of myasthenia gravis: Executive summary. Neurology. 2016;87(4):419–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Suzuki S, Masuda M, Uzawa A, et al. Japan MG registry: Chronological surveys over 10 years. Clin Exp Neuroimmunol. 2022;14(1):5–12. [Google Scholar]
  • 14.Lee DD, Seung HS. Learning the parts of objects by non-negative matrix factorization. Nature. 1999;401(6755):788–91. doi: 10.1038/44565 [DOI] [PubMed] [Google Scholar]
  • 15.Mendoza M, Tran C, Bril V, Katzberg HD, Barnett C. Patient-acceptable symptom states in myasthenia gravis. Neurology. 2020;95(12):e1617–28. doi: 10.1212/WNL.0000000000010574 [DOI] [PubMed] [Google Scholar]
  • 16.Petersson M, Feresiadou A, Jons D, Ilinca A, Lundin F, Johansson R, et al. Patient-Reported Symptom Severity in a Nationwide Myasthenia Gravis Cohort: Cross-sectional Analysis of the Swedish GEMG Study. Neurology. 2021;97(14):e1382–91. doi: 10.1212/WNL.0000000000012604 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Uzawa A, Utsugisawa K. Biological therapies for myasthenia gravis. Expert Opin Biol Ther. 2023;23(3):253–60. doi: 10.1080/14712598.2023.2184257 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Pinyi Lu

29 Apr 2025

Dear Dr. Uzawa,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jun 13 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Pinyi Lu, Ph.D.

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1.Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that funding information should not appear in any section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript.

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

4. Thank you for stating the following in the Competing Interests section:

“AU has received honoraria from Alexion Pharmaceuticals and Argenx. SS received speakers’ fees from Alexion Pharmaceuticals, the Japan Blood Products Organization, Asahi Kasei Medical and Argenx and participated on advisory board meeting of Alexion Pharmaceuticals and Argenx. SK has received Research Grant on Intractable Diseases (Neuroimmunological Diseases) from the Ministry of Health, Labour and Welfare of Japan (20FC1030) and honoraria from CSL Behring. TK reports honorarium for lectures from Alexon Pharmaceuticals.

MPT reports unrestricted research grants from Japan Blood Products Organization, Astellas Pharma, Mitsubishi Tanabe Pharma and Pfizer, outside the submitted work, and has served as a paid Consultant for Alexion, Argenx, Sanofi, and UCB Pharma and received honorarium for lectures from Argenx, Alexion Pharmaceuticals and UCB Pharma. MM has received speaker honoraria from Argenx, Asahi Kasei Medical, Alexion Pharmaceutical and participated on advisory board meeting of Alexion Pharmaceutical. KU has served as a paid Consultant for UCB Pharma, Janssen Pharma, Horizon Therpeutics (Viela Bio), Chugai Pharma, Hanall BioPharma and Mitsubishi Tanabe Pharma, and has received speaker honoraria from Argenx, Alexion Pharmaceuticals, UCB Pharma, and the Japan Blood Products Organisation. Other authors declare no financial or other conflicts of interest.

YN has received speaker honoraria from argenx, Alexion Pharmaceuticals, Japan Blood Products Organization, and UCB Pharma.”

We note that you received funding from a commercial source: Alexion Pharmaceuticals, Argenx, Japan Blood Products Organization, Astellas Pharma, Mitsubishi Tanabe Pharma, Pfizer and UCB Pharma

Please provide an amended Competing Interests Statement that explicitly states this commercial funder, along with any other relevant declarations relating to employment, consultancy, patents, products in development, marketed products, etc.

Within this Competing Interests Statement, please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your amended Competing Interests Statement within your cover letter. We will change the online submission form on your behalf.

5. In this instance it seems there may be acceptable restrictions in place that prevent the public sharing of your minimal data. However, in line with our goal of ensuring long-term data availability to all interested researchers, PLOS’ Data Policy states that authors cannot be the sole named individuals responsible for ensuring data access (http://journals.plos.org/plosone/s/data-availability#loc-acceptable-data-sharing-methods).

Data requests to a non-author institutional point of contact, such as a data access or ethics committee, helps guarantee long term stability and availability of data. Providing interested researchers with a durable point of contact ensures data will be accessible even if an author changes email addresses, institutions, or becomes unavailable to answer requests.

Before we proceed with your manuscript, please also provide non-author contact information (phone/email/hyperlink) for a data access committee, ethics committee, or other institutional body to which data requests may be sent. If no institutional body is available to respond to requests for your minimal data, please consider if there any institutional representatives who did not collaborate in the study, and are not listed as authors on the manuscript, who would be able to hold the data and respond to external requests for data access? If so, please provide their contact information (i.e., email address). Please also provide details on how you will ensure persistent or long-term data storage and availability.

6. Your ethics statement should only appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please ensure that your ethics statement is included in your manuscript, as the ethics statement entered into the online submission form will not be published alongside your manuscript.

7. We note that Figure 1 in your submission contain map/satellite images which may be copyrighted. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For these reasons, we cannot publish previously copyrighted maps or satellite images created using proprietary data, such as Google software (Google Maps, Street View, and Earth). For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

 We require you to either (a) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (b) remove the figures from your submission:

 a. You may seek permission from the original copyright holder of Figure 1 to publish the content specifically under the CC BY 4.0 license. 

 We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

 Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

 In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

 b. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

The following resources for replacing copyrighted map figures may be helpful:

 USGS National Map Viewer (public domain): http://viewer.nationalmap.gov/viewer/

The Gateway to Astronaut Photography of Earth (public domain): http://eol.jsc.nasa.gov/sseop/clickmap/

Maps at the CIA (public domain): https://www.cia.gov/library/publications/the-world-factbook/index.html and https://www.cia.gov/library/publications/cia-maps-publications/index.html

NASA Earth Observatory (public domain): http://earthobservatory.nasa.gov/

Landsat: http://landsat.visibleearth.nasa.gov/

USGS EROS (Earth Resources Observatory and Science (EROS) Center) (public domain): http://eros.usgs.gov/#

Natural Earth (public domain): http://www.naturalearthdata.com/

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

Reviewer #1: Comment#1

The introduction effectively outlines the pathophysiology, classification, and assessment tools for Myasthenia Gravis (MG), making it informative for readers. Highlighting the challenges in defining Minimal Manifestation (MM) status aligns well with the study's objective of standardization, making the research question clear. Also, the transition from general MG characteristics to the specific issue of MM definition is well-structured.

Comment#2

"However, there is no clear definition to identify minimal muscle weakness that is clinically important, which can result in variability in how physicians interpret and apply the MM criteria."

Comment: The variability in MM criteria is a key gap, but the impact of this inconsistency on clinical decision-making should be more explicitly stated. Does it lead to differences in treatment approaches, misclassification of patients, or difficulties in research comparability?

Comment#3

"The quantitative MG (QMG) scale, the MG activities of daily living (MGADL) score, the MG composite (MGC) scale are widely used to evaluate the clinical symptoms of MG patients."

Comment: The sentence is dense and would benefit from breaking down each scale's purpose briefly (e.g., QMG for objective severity, MGADL for daily function, MGC as a composite measure).

Comment#4

Use "Myasthenia Gravis (MG)" at first mention, then "MG" thereafter for consistency.

Reviewer #2: This study developed and assessed machine learning models to predict clinical goal achievement in myasthenia gravis. I commend the authors for utilizing artificial intelligence to help monitor management in this patient population. Some comments for consideration:

1. Throughout the results (AUROC, sensitivity, specificity) results are given without 95% CI. I think it would be valuable to assess the variation in these results based on the 95% CI range and the authors should include them for each result, both in the abstract and the main results.

2. The authors state that the clinical status was assessed by physicians expert in MG treatment- were these physicians primary care, rheumatologists, or any other?

3. What were the baseline characteristics of patients in the 2015 validation cohort? A table summarizing them, perhaps in the supplement, will enhance the findings in the study.

Reviewer #3: This manuscript presents a diagnostic model that predicts NM or better status in MG patients. The model achieves good performance on both the training and validation datasets. However, several points need to be addressed before the manuscript can be considered for publication:

1. The authors should provide a thorough review of existing models that predict NM status in MG patients.

2. If other models for NM prediction exist, the authors need to clearly highlight the advantages of their approach in comparison.

3. (Line 249): The legend for Figure 2 appears after the legend for Figure 3. Please correct the order.

4. (Line 250): Change "he H matrix" to "The H matrix".

5. (Line 261): The authors should report both training and test results for all four models, including SVM, logistic regression, random forest, and Naive Bayes.

6. (Line 261): Since a 5-fold cross-validation was used, the variability across the five folds should be reported.

7. (Line 274): Similarly, for the ensemble model, both training and test performance should be shown, along with the variability across the five cross-validation folds.

8. (Line 277): Given that five iterations were conducted, was an ensemble model generated for each iteration? If so, how were the five ensemble models applied to the external validation dataset? Please clarify.

9. Figures 4 and 5: The ROC curves shown in Figures 4 and 5 use inconsistent font sizes, legend locations, and titles. Please ensure uniform formatting.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: Yes:  Ahmed Kamal Siddiqi

Reviewer #2: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

PLoS One. 2025 Aug 14;20(8):e0330044. doi: 10.1371/journal.pone.0330044.r002

Author response to Decision Letter 1


3 Jul 2025

Response to Reviewers

Reviewer #1:

Comment #1

The introduction effectively outlines the pathophysiology, classification, and assessment tools for Myasthenia Gravis (MG), making it informative for readers. Highlighting the challenges in defining Minimal Manifestation (MM) status aligns well with the study's objective of standardization, making the research question clear. Also, the transition from general MG characteristics to the specific issue of MM definition is well-structured.

Response:

We thank the reviewer for these positive comments about our introduction and the presentation of the research question. In the introduction section, we described the pathophysiology of MG, outlined the various clinical assessment tools (QMG, MGADL, MGC, and MGQOL15r), and highlighted the challenges in MM status determination.

Comment #2

"However, there is no clear definition to identify minimal muscle weakness that is clinically important, which can result in variability in how physicians interpret and apply the MM criteria." Comment: The variability in MM criteria is a key gap, but the impact of this inconsistency on clinical decision-making should be more explicitly stated. Does it lead to differences in treatment approaches, misclassification of patients, or difficulties in research comparability?

Response:

We thank the reviewer for this important and insightful comment. We completely agree that the clinical impact of MM criteria variability should be more clearly described, as this is fundamental to understanding the significance of our research. We have revised the manuscript to better explain how this inconsistency may affect clinical practice.

Revised text (lines 121-128):

"There is no clear definition to identify minimal muscle weakness that is clinically important, which can result in variability in how physicians interpret and apply the MM criteria. This inconsistency may lead to different treatment approaches and creates challenges in comparing outcomes across medical centers. By providing visualization of MG symptoms in MM determination and standardizing the assessment process, our approach aims to support more unified treatment selection criteria and help identify unmet clinical needs in MG management."

Comment #3

"The quantitative MG (QMG) scale, the MG activities of daily living (MGADL) score, the MG composite (MGC) scale are widely used to evaluate the clinical symptoms of MG patients." Comment: The sentence is dense and would benefit from breaking down each scale's purpose briefly (e.g., QMG for objective severity, MGADL for daily function, MGC as a composite measure).

Response:

We thank the reviewer for this helpful and constructive suggestion. We have revised the text to clearly describe the distinct roles of these clinical assessment tools.

Revised text (line 106-111):

"Several standardized scales are widely used to evaluate the clinical symptoms of MG patients: the quantitative MG (QMG) scale for objective assessment of muscle weakness severity, the MG activities of daily living (MGADL) score for evaluating functional limitations in daily activities, and the MG composite (MGC) scale as a comprehensive measure combining both objective and subjective assessments."

Comment #4

Use "Myasthenia Gravis (MG)" at first mention, then "MG" thereafter for consistency.

Response:

We thank the reviewer for pointing out this important inconsistency in our terminology usage. We have carefully revised the entire manuscript to ensure consistent terminology throughout, using "Myasthenia Gravis (MG)" at the first mention and "MG" thereafter.

Reviewer #2:

Comment #1

Throughout the results (AUROC, sensitivity, specificity) results are given without 95% CI. I think it would be valuable to assess the variation in these results based on the 95% CI range and the authors should include them for each result, both in the abstract and the main results.

Response:

We thank the reviewer for this excellent and important suggestion. We agree that reporting 95% confidence intervals is essential for properly assessing the reliability and precision of our results. We have comprehensively revised the manuscript to include 95% confidence intervals calculated from the 5-fold cross-validation results for AUC values, and comprehensive metrics with 95% CIs for the final ensemble model validation results throughout both the abstract and main results sections.

Comment #2

The authors state that the clinical status was assessed by physicians expert in MG treatment- were these physicians primary care, rheumatologists, or any other?

Response:

We thank the reviewer for this important clarification request, as the expertise level of assessing physicians is crucial information for evaluating the validity of our clinical assessments. The clinical status assessments were conducted by neurologists with extensive experience in MG diagnosis and management at specialized neuromuscular centers across Japan. We have added this clarification to the main text in the Data Collection and Access section (lines 139-140), and in the Data Processing section (lines 145-148).

Added text: line 139-140

“The clinical assessments were performed by neurologists with extensive experience in MG diagnosis and management.”

Added text: line 145-148

“The evaluation of these clinical scores, as well as the determination of whether a patient had reached MM or better status (MM, Complete Stable Remission, or Pharmacologic Remission), was conducted by neurologists with expertise in MG management.”

Comment #3

What were the baseline characteristics of patients in the 2015 validation cohort? A table summarizing them, perhaps in the supplement, will enhance the findings in the study.

Response:

We thank the reviewer for this valuable and insightful suggestion. We have added a comprehensive supplementary table (Supplementary Table S1) summarizing the demographic and clinical characteristics of the validation dataset.

Added text: line 293-296

"The baseline characteristics of the validation cohort are presented in Supplementary Table S1. The validation cohort demonstrated similar demographic and clinical patterns to the training dataset, with comparable distributions of MG subtypes and clinical severity scores."

Reviewer #3:

Comment #1

The authors should provide a thorough review of existing models that predict MM status in MG patients.

Response:

We thank the reviewer for this important suggestion. After conducting a comprehensive literature review, we found that, to our knowledge, no previous studies have developed machine learning models specifically for MM status prediction in MG patients. This represents a significant gap in the literature that our study addresses.

Comment #2

If other models for MM prediction exist, the authors need to clearly highlight the advantages of their approach in comparison.

Response:

We thank the reviewer for this excellent point. While no machine learning models exist specifically for MM prediction, existing clinical approaches rely on single clinical score cutoffs, which have inherent limitations. To demonstrate the superiority of our approach, we compared our ensemble model performance with individual score cutoff methods in the validation dataset. Our ensemble model achieved an AUC of 0.94, demonstrating clear superiority over rule-based methods using individual scores: MGADL alone (AUC = 0.92), MGC alone (AUC = 0.91), and MGQOL15 alone (AUC = 0.85). This comparison clearly demonstrates the advantage of our integrated machine learning approach over traditional single-score cutoff methods (Fig 6).

Comment #3

(Line 249): The legend for Figure 2 appears after the legend for Figure 3. Please correct the order.

Response:

We thank the reviewer for identifying this formatting error. We have carefully corrected the figure legend order throughout the manuscript to ensure proper sequential numbering and presentation.

Comment #4

(Line 250): Change "he H matrix" to "The H matrix".

Response:

We thank the reviewer for catching this typographical error. This has been corrected.

Comment #5

(Line 261): The authors should report both training and test results for all four models, including SVM, logistic regression, random forest, and Naive Bayes.

Response:

We thank the reviewer for this important suggestion. We completely agree that comprehensive reporting of both training and test results for all models enhances the transparency and completeness of our evaluation. We have added detailed ROC curves showing both training and test results for all four individual models (SVM, logistic regression, random forest, and Naive Bayes) in the supplementary figures (S1-S3 Figures).

Comment #6

(Line 261): Since a 5-fold cross-validation was used, the variability across the five folds should be reported.

Response:

We thank the reviewer for this excellent point about reporting cross-validation variability. We have comprehensively reported the variability across all five cross-validation folds by showing individual fold results in our ROC curve figures (Supplementary Figures 1-3). Given the manageable number of folds (n=5), we present all individual AUC values for each fold rather than just summary statistics, providing complete transparency about model performance variability.

Comment #7

(Line 274): Similarly, for the ensemble model, both training and test performance should be shown, along with the variability across the five cross-validation folds.

Response:

We thank the reviewer for this important suggestion. We have included comprehensive results showing both training and test performance for the ensemble model, with individual fold results displayed in the ROC curves to demonstrate the complete range of performance across all five cross-validation iterations (Fig 5).

Comment #8

(Line 277): Given that five iterations were conducted, was an ensemble model generated for each iteration? If so, how were the five ensemble models applied to the external validation dataset? Please clarify.

Response:

We thank the reviewer for requesting this important methodological clarification. For external validation, we used all five trained models from each cross-validation fold: for each validation sample, we calculated predictions from all five folds and used soft voting (averaging probabilities) to generate the final ensemble prediction.

line 193-195:

“For performance evaluation including AUROC, we applied a soft voting strategy, averaging the predicted probabilities from all four models to generate the final prediction scores.”

Comment #9

Figures 4 and 5: The ROC curves shown in Figures 4 and 5 use inconsistent font sizes, legend locations, and titles. Please ensure uniform formatting.

Response:

We thank the reviewer for this attention to detail regarding figure presentation. We have carefully reviewed and uniformly corrected the formatting of all ROC curve figures, ensuring consistent font sizes, legend locations, and titles throughout the manuscript.

Attachment

Submitted filename: Response to Reviewers.docx

pone.0330044.s006.docx (19.8KB, docx)

Decision Letter 1

Pinyi Lu

25 Jul 2025

Predicting achievement of clinical goals using machine learning in myasthenia gravis

PONE-D-24-57934R1

Dear Dr. Uzawa,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support .

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Pinyi Lu, Ph.D.

Academic Editor

PLOS ONE

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions??>

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #2: Yes

Reviewer #3: Yes

**********

Reviewer #2: The authors have adequately addressed my comments and revised the manuscript accordingly. The manuscript is acceptable for publication.

Reviewer #3: (No Response)

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #2: No

Reviewer #3: No

**********

Acceptance letter

Pinyi Lu

PONE-D-24-57934R1

PLOS ONE

Dear Dr. Uzawa,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Pinyi Lu

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. ROC curves for machine learning model performance on training data.

    ROC curves and AUC values for four machine learning models evaluated on training data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

    (PDF)

    pone.0330044.s001.pdf (491.9KB, pdf)
    S2 Fig. ROC curves for machine learning model performance on test data.

    ROC curves and AUC values for four machine learning models evaluated on test data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

    (PDF)

    pone.0330044.s002.pdf (461.4KB, pdf)
    S3 Fig. ROC curves for machine learning model performance on validation data.

    ROC curves and AUC values for four machine learning models evaluated on validation data using 5-fold cross-validation. (a) Logistic Regression, (b) Naive Bayes, (c) Random Forest, (d) SVM. Individual fold ROC curves (colored lines) and mean ROC curve (thick line) are shown with corresponding AUC values.

    (PDF)

    pone.0330044.s003.pdf (454.8KB, pdf)
    S1 Table. The profiles of 414 MG patients in validation data.

    (CSV)

    pone.0330044.s004.csv (624B, csv)
    Attachment

    Submitted filename: Response to Reviewers.docx

    pone.0330044.s006.docx (19.8KB, docx)

    Data Availability Statement

    As stated in our Data Sharing Statement, the clinical data are not publicly available due to privacy concerns. However, data may be available upon reasonable request and subject to review within the MGR group to establish the viability of sharing. To request access to the data, please contact the Ethics Committee of Hanamaki General Hospital (4-56 Odaya-chō, Hanamaki, Iwate 025-0082, Japan; phone: +81-198-23-3311), or the Japan MG Registry website (http://mg.kenkyuukai.jp/information/information_detail.asp?id=67084).


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES