Skip to main content
Frontiers in Cellular and Infection Microbiology logoLink to Frontiers in Cellular and Infection Microbiology
. 2026 Mar 12;16:1787545. doi: 10.3389/fcimb.2026.1787545

Fungal recognition in vaginal discharge using deep learning analysis of mobile device-acquired microscopic images

Monsicha Pongpom 1, Siriwoot Sookkhee 1, Siriporn Chongkae 1, Sara Wattanasombat 1, Kornprom Pikulkaew 2, Narin Lawan 3, Phit Upaphong 4,*, Tanaporn Wangsanut 1,*
PMCID: PMC13017809  PMID: 41909846

Abstract

Background

Vulvovaginal candidiasis (VVC) is a common fungal infection that is frequently diagnosed through manual microscopic examination of vaginal discharge. Artificial Intelligence (AI)-assisted analysis of microscopic images enables rapid and accurate diagnosis, supporting timely and effective antifungal therapeutic interventions. However, conventional light microscopy often lacks cameras, limiting digital image analysis and AI applications. While mobile devices offer a practical alternative, no AI tools currently exist for the automated detection of fungal cellular morphology in microscopic images captured by smartphones and tablets. In this study, we developed deep learning models to segment fungal morphologies in microscopic images of vaginal discharge acquired with smartphones and tablets.

Methods

Three models were developed: ResNet18 for binary classification (Candida presence/absence), YOLOv5 for detection, and YOLOv11 for segmentation. Models were trained using 1,259 microscopy images of Gram-stained vaginal discharge acquired with smartphones or tablets, along with 67 images obtained from conventional microscopes. These images were divided into training, validation, and test sets. Annotated microscopic images for fungal elements were used to train YOLO models in a two-stage approach: Stage 1 utilized 687 annotated images of yeast infections to learn general fungal morphology, comprising 266 bounding box–annotated images sourced from Roboflow and 421 segmentation-labeled images manually annotated from the open-access dataset. Stage 2 fine-tuned the models on the annotated mobile device-acquired dataset. Metrics included F1-score, area under the curve (AUC), precision, recall, and mean average precision at 50% intersection over union (mAP50). Experts assessed segmentation outputs for diagnostic utility, providing explainability to the AI results.

Results

ResNet18 achieved F1-score=0.986, AUC = 0.99. YOLOv5 performed best at IoU=0.50 (precision=0.812, recall=0.622, mAP50 = 0.730); YOLOv11 at IoU=0.25 (precision=0.766, recall=0.700, mAP50 = 0.727). Expert ratings averaged 4.25/5. Only 3.68% of images were rated as inappropriate due to false negative or false positive segmentations.

Conclusion

ResNet18 accurately classified microscopic images for fungal elements, while the YOLOv11 model effectively delineated Candida morphologies, including yeasts, budding yeasts, and filamentous forms from clinical specimens. The high accuracy and positive expert feedback demonstrate the feasibility of integrating AI-assisted mobile microscopy into routine workflows, thereby advancing digital analysis of microbial infections using conventional light microscopy. With further clinical validation and expansion to include other infections, this approach holds great potential to establish robust real-world utility.

Keywords: Artificial intelligence, Candida, digital health, fungal image recognition, medical image analysis, mobile device, smartphone AI, vulvovaginal candidiasis

1. Introduction

Microscopic examination plays a central role in the clinical identification of microbial pathogens from patient-derived specimens. However, camera-equipped microscopes remain unavailable in many research facilities, large teaching laboratories, and healthcare centers, particularly in underserved areas, limiting access to digital image analysis and remote consultation. The widespread availability of mobile devices offers a practical, low-cost alternative for capturing high-resolution microscopy images. For example, mobile microscopy was successfully integrated into telemedicine for parasitic infection quantification in remote areas and into medical education through real-time image sharing and feedback (Dacal et al., 2021; Xu et al., 2025). At our institution, mobile microscopy is routinely used in teaching laboratories for reports, enhancing hands-on learning and mirroring real-world applications. This growing reliance on mobile devices presents an opportunity to overcome the limitations imposed by the lack of camera-equipped microscopes, enabling digital workflow integration and remote consultation within routine microscopic practice.

Vaginal discharge syndrome results from alterations in vaginal secretions, commonly due to infection of the vagina or cervix (Workowski et al., 2021) (Shroff, 2023). Most symptomatic cases of vaginal infections are caused by bacterial vaginosis (40–50%), vulvovaginal candidiasis (VVC; 20–25%), or trichomoniasis (15–20%) (Sobel, 1997; Paladine and Desai, 2018). Cervicitis, which frequently occurs secondary to sexually transmitted infections, may also present with vaginal discharge in conjunction with other clinical manifestations. Notably, Neisseria gonorrhoeae and Chlamydia trachomatis account for 30% to 50% of infectious cervicitis case (Shroff, 2023). Importantly, abnormal vaginal discharge must be distinguished from physiological vaginal discharge associated with normal vaginal conditions, which are characterized by a predominance of Lactobacillus species adherent to vaginal epithelial cells. Therefore, the ability to accurately distinguish among these various infections and non-infectious conditions is crucial for the effective clinical management of vaginal discharge syndrome (Shroff, 2023; Workowski et al., 2021).

Given the wide range of potential causes of vaginal discharge, this study focuses on VVC. Most women experience at least one episode, and up to 45% develop recurrent disease (Anderson et al., 2004; Sobel, 1997; Gonçalves et al., 2016). VVC typically presents with pruritus, soreness, vulvar erythema, and thick white discharge. Accurate Candida identification is essential to prevent misdiagnosis, persistent symptoms, and inappropriate treatment. Microscopic examination remains central to diagnosis, with Gram staining considered particularly reliable for distinguishing infection from commensal colonization (Qi et al., 2021). Candida albicans exhibits polymorphism, appearing as yeast, pseudohyphae, or true hyphae, with filamentous forms closely associated with virulence (Sudbery, 2011; Sudbery et al., 2004; Noble et al., 2017; Bettauer et al., 2022). Yeast cells alone may represent asymptomatic colonization, whereas budding yeasts or (pseudo)hyphae are required for laboratory confirmation of VVC (Workowski et al., 2021; Wang et al., 2025). Therefore, effective clinical diagnosis relies on the recognition of the various morphological forms of Candida species (Roselletti et al., 2019; Wang et al., 2025).

Recent advances in AI, particularly Convolutional Neural Networks (CNNs), have enabled automated fungal classification, detection, and segmentation (Supplementary Table 1) (Kim et al., 2024). However, most prior studies relied on high-quality images acquired under controlled conditions, using camera-equipped microscope, which limits their generalizability (Supplementary Table 1) (Kim et al., 2024). In real-world resource-limited settings, mobile-acquired images tend to display lower and more variable quality. Moreover, few studies have evaluated performance on images captured by non-experts such as medical students or frontline healthcare personnel. Robust models tailored to these conditions are therefore needed.

To address this gap, this study developed and validated deep learning models for fungal analysis using Gram-stained vaginal discharge samples acquired by mobile devices under realistic conditions. Our framework includes a ResNet18 classification model, a YOLOv5 detection model, and a YOLOv11 segmentation model. Our deep learning models can directly identify fungal morphology from a wide range of Gram-stained vaginal discharge smears, enabling timely and appropriate treatment decisions. By emphasizing heterogeneous imaging conditions and practical deployment, this work aims to bridge advanced AI techniques with routine mycology practice that may help reduce antifungal drug resistance and improve patient outcomes.

2. Materials and methods

2.1. Microscopic slide preparation of vaginal discharge specimens

Permanent Gram-stained slides of vaginal discharge were prepared for use in teaching laboratories. Vaginal discharge samples were collected anonymously from patients at Maharaj Nakorn Chiang Mai Hospital (Chiang Mai, Thailand) in accordance with relevant guidelines for human research, with approval from the institutional ethics committee (Approval No. 441/2568).

2.2. Study design and image collection

In total, 1,259 microscopic images of Gram-stained vaginal discharge were collected, encompassing the full spectrum of conditions seen in vaginal discharge syndrome. Specifically, the dataset included cases of bacterial vaginosis (BV), gonococcal infection (GU; Neisseria gonorrhoeae), chlamydial infection (NGU; Chlamydia trachomatis), trichomoniasis (TV; Trichomonas vaginalis), vulvovaginal candidiasis (VVC), as well as physiological (normal) vaginal samples dominated by Lactobacillus. This heterogeneous collection was used to train deep learning models for three tasks: binary classification of fungal presence (ResNet18), fungal object detection (YOLOv5), and fungal instance segmentation (YOLOv11). The number of images used for training each model is detailed in Table 1. An independent test set smartphone-captured images (not used in model training) was set aside to evaluate the final performance of each model (Table 1).

Table 1.

Datasets used for the ResNet18, YOLOv5 and YOLOv11 models.

Dataset ResNet18 YOLOv5 YOLOv11
Training set 570 (65%) Set 1: 271 and Set 2: 244 (67%) 362 (60%)
Validation set 143 (17%) Set 1: 32 and Set 2:62 (12%) 91 (15%)
Test set 159 (18%) 165 (21%) 165 (25%)
Total 872 774 618

Of the 1,259 mobile-acquired images, 894 were collected during a medical school laboratory session (Diagnostic of Vaginal Leucorrhea and Sexually Transmitted Infections, course 330329 PMLI II) from September to November 2025 (Tables 2, 3). A total of 235 students each captured images from three different slides under 1000X magnification using their smartphones or tablets held to the microscope ocular (Supplementary Figure 1A). Students submitted images (.jpg or.heic format) via the university’s learning management system as routine lab reports. All images were downloaded, anonymized, and converted to .jpg format.

Table 2.

Characteristics and number of images used in this study.

Type Image # Source
Mobile device
Mobile device Camera-equipped microscope
Total
894
298
67
1259
PMLI II Lab
Research Lab
Research Lab
Camera-equipped microscope
Camera-equipped microscope
Total
266
421
687
Roboflow (YOLOv5)
Nguyen et al., 2025

Table 3.

Number of images categorized by infections.

Group BV TV GU NGU VVC Transition Healthy
F (n = 362) 28 188 146
NF (n = 351) 152 44 83 72

*There are 146 yeast-containing images (only yeast forms present) and 216 pseudohypha-containing images (at least one pseudohyphal form present, regardless of other morphologies). Of the pseudohypha-containing images, 28 lack Lactobacillus. Based on fungal morphology and Lactobacillus presence (Supplementary Table 2), the dataset includes 146 healthy images, 188 transition images, and 23 VVC images.

An additional 298 images were acquired in a controlled setting using a Celestron NexYZ universal adapter (Celestron, LLC; https://www.celestron.com) mounted on microscope eyepieces, allowing smartphones to function as digital microscopes (Supplementary Figure 1B). To replicate student image variability, at least two laboratory personnel captured images using three smartphone models (iPhone 8, 13, 15) on three Olympus CX31 microscopes. A subset of 67 images was also captured using a dedicated digital microscope (Nikon DS-Fi1 microscope model) to provide high-quality training and validation samples.

2.3. Image annotation

All images were annotated and reviewed by at least two medical mycology experts based on established diagnostic characteristics (Supplementary Table 2). Annotation involved drawing bounding boxes and free-form segmentation masks around fungal elements using the Labelbox platform (Labelbox, Inc.). A portion of images (10%) was annotated independently by at least two labelers to ensure consistency. One expert annotator (T.W.) then reviewed and reconciled any discrepancies to establish a consensus ground truth across all annotations. The finalized annotations were exported from Labelbox in JSON format and converted to the required YOLOv5 and YOLOv11 training formats using custom Python scripts. For the YOLOv11 segmentation format, pixel masks were converted into polygon coordinates using OpenCV’s approxPolyDP function with an epsilon value of 0.001, enabling the model to utilize polygon mask labels.

2.4. Image preprocessing

For the classification task, Contrast Limited Adaptive Histogram Equalization (CLAHE) was applied to improve image quality (Yadav et al., 2014; Lidong et al., 2015). CLAHE enhances local contrast and can reveal subtle features while minimizing noise amplification, thus facilitating more accurate detection and segmentation of fungal elements.

2.5. Model architecture and training

2.5.1. Classification (presence/absence of fungus)

For binary classification, multiple CNN architectures were evaluated for efficient mobile deployment. MobileNetV2 (Sandler et al., 2018), EfficientNetB0 (Tan and Le, 2019), and ResNet18 (He et al., 2016) (all ImageNet-pretrained) were selected as candidates due to their compact size and fast inference. ImageNet is a widely used, large-scale visual database (https://www.image-net.org/) containing over 14 million annotated images across more than 20,000 categories. Each model was fine-tuned to classify images as fungus-positive or fungus-negative. ResNet18 achieved highest validation accuracy and was selected as the final classification model.

A ResNet18 CNN classifier was then trained to categorize images as containing fungal elements (F) or not (NF). The F group included samples with Candida (either VVC cases or asymptomatic Candida colonization; n = 362 images), while the NF group comprised samples from BV, TV, GU, and NGU, reflecting other causes of discharge (n = 351 images) (Table 3). The network used ImageNet-pretrained weights with the final layer modified for binary classification. Images were resized to 224×224 pixels with standard ImageNet normalization. Data augmentation during training involved resizing, random horizontal flipping (probability 0.5), and random rotation up to ±10° for each image per epoch. These transformations did not increase the number of images but enhanced dataset diversity throughout training. Training used Adam optimizer (learning rate 1×10-4), cross-entropy loss, and early stopping (patience=5 epochs). The best model was selected based on validation accuracy.

Model performance was evaluated on validation and independent test sets (159 images). The best checkpoint was loaded for inference. Test images underwent identical normalization and resizing as training. Predictions used softmax over two outputs, classifying images as F or NF based on higher probability. Standard metrics (precision, recall, F1-score, accuracy) were calculated using scikit-learn. Confusion matrices and ROC curves visualized performance. Gradient-weighted Class Activation Mapping (Grad-CAM++) (Selvaraju et al., 2017; Chattopadhay et al., 2018) was applied to representative images, generating heatmaps highlighting influential regions for predictions. All code was implemented in PyTorch and executed in Google Colab with GPU acceleration.

2.5.2. Object detection and segmentation

For fungal object detection and segmentation, YOLO-based models were employed. Specifically, YOLOv5 (for detection) and YOLOv11 (for segmentation) models, pretrained by Ultralytics on the COCO (Common Objects in Context) dataset (Lin et al., 2014), were used as starting points. YOLOv5 was chosen for its lightweight, fast architecture, suitable for eventual mobile deployment, while YOLOv11 was used to showcase instance segmentation of fungal elements on mobile-acquired images.

For YOLOv5 detection, the small variant (YOLOv5s) was fine-tuned starting from the bestYeast701.pt model (Wangsanut et al., 2025), previously trained on general yeast and pseudohyphae images (https://universe.roboflow.com/yeast-rrffa/yeast-5arid/dataset/1). Although this base model recognized fungal morphology, it was not specialized for mobile-acquired Gram-stained images. The model was therefore fine-tuned on the 609-annotated mobile dataset (Table 1), producing bestM2.2.pt. Since YOLOv5 requires bounding boxes, filamentous forms (hyphae) were annotated with multiple small boxes along their length to minimize background inclusion during training.

For YOLOv11, a two-stage training strategy was employed to address the domain gap between natural and microscopy images. First, YOLOv11 was trained on an external public dataset of wet mount preparation of vaginal discharge images (Nguyen et al., 2025). A total of 421 images (Table 2) were annotated for fungal elements (yeast and hyphae), converted to segmentation masks, then polygon format, yielding an initial model (bestVVC1.pt). This model was then fine-tuned on the mobile-acquired dataset (453 images, Table 1), producing the final bestF1.1.pt model. This transfer learning approach enabled the model to learn general fungal morphology from high-quality images before adapting to smartphone image variability. All images were resized to 640×640 pixels. Training used 100 epochs, batch size 8, with PyTorch implementation of YOLOv5/YOLOv11 (built-in augmentation) on Google Colab GPUs.

For YOLOv5 hyperparameter tuning, Ultralytics’ automated evolution method was used (https://docs.ultralytics.com/yolov5/tutorials/hyperparameter_evolution). This genetic algorithm iteratively optimized hyperparameters (learning rate, batch size, augmentation settings) based on performance. Evolution was performed on an HPE Apollo 6500 system with AMD EPYC 7742 CPU and NVIDIA A100 GPU. Both YOLO models were evaluated on validation sets at multiple intersection over union (IoU) thresholds (0.25, 0.50, 0.75) following established methodologies (Koo et al., 2021; Ren et al., 2025).

2.5.3. Data separation strategy

To prevent data leakage, training and test datasets were separated at the patient level by design, based on non-overlapping data source, slide identity, and acquisition date. Slides collected during different laboratory sessions originated from different patients. Images acquired on 16 September and 2 October 2025 were used for training, whereas images from 14 October 2025 were reserved exclusively for testing. Research laboratory slides, which yielded 50–100 images per slide, were obtained from a separate patient cohort and were distinct from those used in the teaching laboratory.

2.5.4. Calculation of intersection-over-union and Dice coefficient

To quantitatively evaluate segmentation performance at the pixel level, IoU and Dice coefficient were calculated between the predicted segmentation masks and the corresponding ground-truth annotations. For each image, the binary mask generated by the YOLOv11 model was compared with the manually annotated reference mask. IoU was defined as the ratio of the overlapping area between the predicted and ground-truth masks to their union, while the Dice coefficient, which emphasizes spatial overlap, is commonly used for biomedical image segmentation. IoU and Dice scores were computed for each image and averaged across the test dataset. These metrics were applied consistently to both the internal mobile device–acquired images and the external open-source images to enable direct comparison of segmentation performance across datasets.

2.5.5. Expert evaluation of the segmentation model

The YOLOv11 segmentation model was evaluated by an expert panel of clinicians and microbiologists specializing in medical mycology, each with >5 years of experience (Supplementary Tables S3-S5). Thirty-four images were selected: 19 positive (containing fungi) and 15 negative cases. YOLOv11-predicted segmentation masks were overlaid on original images and presented to experts. Each expert independently assessed: (1) appropriateness of predicted segmentation (whether highlighted areas matched actual fungal elements), and (2) satisfaction with model performance. Responses used a five-point Likert scale. A total of 272 ratings were collected (34 images × 8 experts). The frequency of “very inappropriate” ratings was recorded to identify failures. Rating distributions were analyzed separately for positive and negative images. Experts reported whether inaccuracies would impact clinical use (from “no impact” to “major deterrent”) and provided overall satisfaction ratings (1=very dissatisfied to 5=very satisfied).

3. Results

3.1. Classification task (presence vs absence of fungus)

To fully represent the spectrum of vaginal discharge syndromes encountered in clinical settings, the mobile device-acquired images included cases of VVC, BV, TV, GU, NGU, as well as physiologic discharge associated with a normal vaginal microbiome. All fungal morphologies, including yeast, budding yeast, and hyphae, were grouped into a single “fungus” category for classification and detection in this study; this approach was adopted for proof-of-concept purposes. On the independent test set of 159 unseen images, ResNet18 demonstrated superior performance among evaluated CNN architectures for binary classification, achieving an F1-score of 0.986 and AUC of 0.99 in distinguishing fungal from non-fungal images (Figures 1A, B; Table 1). Alternative models showed lower performance: MobileNetV2 and EfficientNetB0 achieved F1-scores of 0.882 and 0.889, respectively (Supplementary Figure 2; Table 4).

Figure 1.

Panel A shows a confusion matrix with high classification accuracy for two classes labeled F and NF. Panel B displays an ROC curve for class F with an area under the curve of zero point nine nine. Panel C contains three rows of paired microscopic images, left showing original stained slides, and right showing GradCAM++ heatmaps with classification scores and predicted labels.

Confusion matrix, Receiver-Operating characteristic (ROC) curve, and Grad-CAM illustrate the classification performance of ResNet18 model on mobile-acquired microscopic images of Gram-stained vaginal discharge samples. (A.) The confusion matrix of model trained using pre-trained architecture: ResNet18. (B.) The ROC curves for the F (left) and NF (right) classes produced by the ResNet18 classification model on test images (n = 159). The area under the curve (AUC) is annotated for each class. (C.) Example of mobile device-microscopic images and the corresponding explanatory maps. Abbreviations: F, The presence of fungal elements, and NF, The absence of fungal elements.

Table 4.

Comparison of classification and object detection models for fungal recognition in Gram-stained vaginal discharge samples.

Model Precision Recall F1 score
Classification
MobileNetV2
- F
- NF
Average
0.8684
0.8971
0.9041
0.8592

0.8859
0.8777
0.8819
EfficientNetB0
- F
- NF
Average
0.8608
0.9231
0.9315
0.8451

0.8947
0.8824
0.8889
ResNet18
- F
- NF
Average
0.9733
1.0000
1.0000
0.9718

0.9865
0.9857
0.9857
Detection
YOLOv5
Fungal elements
0.812 0.622 mAP50 = 0.730
Segmentation
YOLOv11
Fungal elements
0.766 0.700 mAP50 = 0.727

F, The presence of fungal elements, NF, The absence of fungal elements, mAP50 = Mean average precision calculated at an intersection over union (IoU) threshold of 0.50.

Examples of prediction results on test images are illustrated in Figures 1-4. The YOLO models have been sequentially trained twice with human annotated images (first with conventional camera-acquired microscopic images and second with mobile device-acquired microscopic images), and then fine-tuned hyperparameter. The precision, recall and mAP50 values for the YOLOv5 and YOLOv11 models were reported at IoU of 0.50 and 0.25, respectively.

To generate visual explanation maps for ResNet18 models, Grad-CAM++ was applied to produce heatmaps highlighting the regions of input images that contributed most significantly to the model’s predictions. As depicted in Figure 1C, heatmaps consistently highlighted fungal cell clusters in Candida-positive images, demonstrating model focus on relevant microscopic features. In positive (F) images, overlays corresponded to yeast clusters or hyphae regions, while in negative (NF) images, attention was diffusely distributed or focused on background elements.

3.2. Fungal detection with YOLOv5

While the yeast forms of Candida species exhibit round to oval shapes, their filamentous forms appear as curved linear structures, which can pose challenges for fungal image recognition (Nguyen et al., 2025; Wang et al., 2025). As the YOLOv5 model detects objects using bounding boxes, filamentous form was manually annotated by drawing multiple small boxes along the length of each hypha to minimize background presented in vaginal discharge samples. A total of 609 microscopic images were manually annotated for the presence of fungal elements (Tables 1, 2). To reduce the false-negative rate and ensure the model detects as many suspected hyphal regions as possible, we evaluated segmentation performance across a range of IoU thresholds (0.1, 0.25, 0.5, and 0.75), following recommendations from Koo et al. (2021) (Table 3). After fine-tuning, YOLOv5 demonstrated strong object-detection performance, achieving optimal results at IoU=0.50 with precision=0.812, recall=0.622, and mAP50 = 0.730 (Tables 1, 5). At the more lenient IoU=0.25, precision remained high (>0.80) with increased recall, while at the stricter IoU=0.75, recall decreased despite maintained precision. IoU=0.50 was selected as the optimal threshold, balancing localization accuracy and false positive minimization.

Table 5.

Performance of object detection and segmentation models across IoU thresholds.

IoU Box
Precision
Box
Recall
Box
mAP50
Mask
Precision
Mask
Recall
Mask
mAP50
YOLOv5
0.10
0.25
0.50
0.75
0.758
0.743
0.812
0.750
0.687
0.691
0.622
0.583
0.714
0.719
0.730
0.668
-
-
-
-
-
-
-
-
-
-
-
-
YOLOv11
0.10
0.25
0.50
0.75
0.809
0.803
0.788
0.752
0.739
0.740
0.739
0.711
0.765
0.771
0.764
0.741
0.775
0.766
0.767
0.741
0.697
0.700
0.690
0.656
0.724
0.727
0.723
0.704

For model detection visualization, fine-tuning on mobile device-acquired microscopic images substantially improved detection accuracy. In one test image, the retrained YOLOv5 detected 57 fungal elements versus 49 by the pre-trained model (Figure 2B, top). The fine-tuned model also reduced false positives, better distinguishing fungi from artifacts such as epithelial cell edges or microscope pointer arrows (green arrows, Figure 2B, bottom). These improvements demonstrate how transfer learning with domain-specific annotations enhanced Candida cellular morphology recognition.

Figure 2.

Panel A contains 10 graphs showing loss, precision, recall,and mean average precision metrics for a model's training and validation processes overone hundred epochs. Panel B presents six circular microscopic images in two rows, withoriginal images on the left, and two different detection models’ results on the right,where detected objects are marked with blue boxes and text annotations.

YOLOv5 training with transfer learning for mobile-acquired microscopic images of Gram-stained vaginal discharge samples. The bestYeast701.pt (yolov5s.pt-based model available via Hugging Face) was used as a base model and fine-tuned on mobile-acquired microscopic images using a transfer learning method, resulting in bestM2.2.pt model. (A) Training matrices from fine-tuning bestYeast701.pt to mobile-acquired images. (B) Visualization of model prediction after fine-tuning. Green arrow indicates background artifacts.

3.3. Fungal segmentation with YOLOv11

To address the challenge of image recognition of filamentous forms, the YOLOv11 segmentation model was trained to enable precise annotation of the curved, linear structures characteristic of fungal hyphae. A total of 874 microscopic images were manually annotated for the presence of fungal elements using free-form segmentation (Tables 1, 2). As shown in Figure 3, the YOLOv11 segmentation model accurately outlined individual yeast and hyphal structures even in challenging, variably stained fields. By contrast, the initial YOLOv11 model (before fine-tuning on our mobile device-acquired microscopic data) often misclassified background artifacts as fungal elements. For example, as shown in Figure 3B, a piece of background elements (yellow arrow, Figure 3B middle panel) was mistakenly highlighted as a fungus by the initial model but correctly ignored by the fine-tuned model (Figure 3B right panel). On a validation set of 91 images containing 1,132 annotated fungal instances, YOLOv11 achieved optimal performance at IoU=0.25, with precision=0.766, recall=0.700, and mAP50 = 0.727 (Tables 1, 5). When evaluated as bounding boxes (ignoring mask boundaries), precision exceeded 0.80 at IoU=0.25, indicating accurate localization of fungal objects.

Figure 3.

Panel A presents sixteen line charts showing training and validationlosses, precision, recall, and mean average precision metrics over 100 epochs, alldisplaying downward loss trends and increasing performance metrics. Panel B showsthree circular microscopic images: the left is the original stained cell sample, the centerdemonstrates YOLOv11 bestVVC1 model predictions with green markings and yellowarrows, and the right shows YOLOv11 bestF1.1 results with green-highlighted features

YOLOv11 training with transfer learning for mobile-acquired microscopic images of Gram-stained vaginal discharge samples. The yolov11s-seg.pt model was used as a base model and fine-tuned on mobile-acquired microscopic images using a transfer learning method, resulting in bestF1.1.pt model. (A) Training matrices from fine-tuning YOLOV11 segmentation model to mobile-acquired microscopic images with manual annotation of fungal elements (bestF1.pt base model). (B) Visualization of model prediction after fine-tuning, confident threshold was set to 0.1 for display fungal elements. Yellow arrow indicates background artifacts.

Both YOLO models (v5 and v11) exhibited high precision (0.7–0.8), indicating that detected fungal elements were usually correct. However, recall was more modest (0.6–0.7), likely due to imaging challenges in the heterogeneous mobile-captured dataset. Missed detections occurred predominantly in out-of-focus or low-contrast regions, particularly at image peripheries where microscope fields often blur with smartphone or tablet cameras, while centrally focused regions showed reliable detection.

Despite imaging variability, YOLOv5 and YOLOv11 effectively localized and delineated fungal elements in mobile device-acquired microscope images, performing robustly even on blurred, low-contrast, or artifact-laden images (Figure 4A). Importantly, both models correctly handled fungus-negative cases by producing no detections, even with complex backgrounds such as white blood cell-dominated slides (gonococcal/chlamydial infections) or Gardnerella-coated clue cells (BV), demonstrating crucial clinical specificity (Figure 4B).

Figure 4.

Panel A displays four circular microscopy images: the original sample, a ResNet18 heatmap overlay, a YOLOv5 detection with blue bounding boxes, and a YOLOv11 detection with green highlights. Panel B shows four images: the original sample, a ResNet18 heatmap overlay, and two nearly identical images without detected features. Both rows compare model-based visualizations of cellular structures under a microscope.

Visualization of fungal recognition on fine-tuned models using mobile device-acquired images of Gram-stained vaginal discharge samples. Images were acquired by using a smartphone (A) or a tablet (B). The examples illustrate mobile device-acquired images that were correctly predicted as showing the presence (F, A.) or the absence (NF, B.) of fungal elements by ResNet18 classification (left panel), YOLOv5 fungal detection, (version M2.2 middle panel), and YOLOv11 fungal segmentation (version F1.1 right panel) in agreement with expert human ground truth.

3.3.1. Expert evaluation of the AI segmentation model

Expert evaluation focused on the fungal segmentation models, as this image recognition task offers the greatest potential for explainable AI. A total of 34 mobile device-acquired microscopic images were included, consisting of 19 images containing fungal elements and 15 images without fungal elements. First, mask-based overlap metrics were calculated on this test set. The YOLOv11 model achieved an average IoU of 0.49 and an average Dice coefficient of 0.65, indicating moderate segmentation accuracy at the pixel level (Table 6, Supplementary Figure 3). Second, The same set of images was independently reviewed by eight experts, resulting in 272 total ratings. Across all ratings, the rate of “very inappropriate” responses was notably low, with only 10 ratings (3.68%) out of 272 assigned to this category, and these were confined to only two images (Supplementary Table 3; Supplementary Figure 4). For the 19 positive images (152 ratings), expert evaluations indicated high model performance, with 49.34% of responses rated as “mostly appropriate” and 25% as “very appropriate” (Supplementary Table 3; Supplementary Figure 5). For the 15 negative images (120 ratings), all expert responses (100.00%) rated the absence of a predicted mask as “very appropriate,” demonstrating that the model correctly recognized the background in every case (Supplementary Table 4). These findings highlight the high overall appropriateness and reliability of the AI model’s segmentation performance.

Table 6.

Mask-based segmentation performance of the YOLOv11 model on the internal mobile device-acquired test set (n = 34; number of fungal-positive image = 19).

Metric *Mean ± SD
IoU 0.49 ± 0.11
Dice 0.65 ± 0.10

* Values are reported as mean ± standard deviation.

Experts noted that the model performed well overall, effectively identifying fungal elements with only minor instances of under- or over- segmentation. These segmentation area issues were not considered significant enough to deter clinical or educational implementation: 37.5% stated minor errors would not affect willingness to use the model, 25% reported slight impact, and none indicated strong deterrence (Supplementary Table 5). Experts suggested that the target area or fungal elements should be centered within the image, as peripheral regions were often out of focus and frequently exhibited missing fungal segmentation masks. Overall satisfaction was strongly positive. The mean satisfaction was 4.25/5.

Together, the quantitative mask-based metrics and expert evaluations support the overall reliability and appropriateness of the YOLOv11 model’s segmentation performance on our mobile device–acquired images.

3.3.2. External evaluation and generalizability

In the absence of external clinical datasets of mobile device–acquired microscopic images, we performed a preliminary, proof-of-concept external evaluation using 21 openly available Gram-stained microscopic images of vaginal discharge (Supplementary Table 6) to test the YOLOv11 model (bestF1.1.pt). The model delineated pseudohyphae and yeast morphologies in images obtained from outside our institution (Supplementary Figure 6; Beder et al., 2025; Nambiar et al. 2021), achieving an average IoU of 0.44 and Dice score of 0.60, indicating limited but measurable generalizability. Some false-positive and false-negative predictions were observed. Segmentation performance was slightly higher on internal mobile-acquired images (IoU = 0.49, Dice = 0.65), consistent with model optimization for mobile microscopy. Together, these results demonstrate the feasibility of morphology-based fungal detection on external images while highlighting the need for future validation using clinically sourced, mobile-acquired datasets.

4. Discussion

In this proof-of-concept study, we demonstrated that AI-based image recognition is a promising technology for the automatic identification of fungal elements using mobile device-acquired microscopic images of Gram-stained vaginal discharge. Our classification model, ResNet18, exhibited strong performance, as reflected by high F1 scores and AUC values, both matrices exceeding 0.90. The performance of our model is comparable to previous classification models trained on high quality images obtained from camera-equipped microscopes (Supplementary Table 1; (Wang et al., 2025; Nguyen et al., 2025). In a previous study, Nguyen et al. (2025) employed the MobileNetv2 model to screen wet-prep vaginal discharge samples for vaginitis diagnosis. Additionally, Wang et al., 2025 applied YOLOv5 for the automated diagnosis of VVC from Gram-stained vaginal discharge samples. Notably, none of these studies utilized a fungal segmentation model to delineate fungal cellular morphology from microscopic images. While our YOLOv11 segmentation model may occasionally over- or under-segment fungal elements, expert evaluation demonstrated these outputs remain clinically acceptable. This human-in-the-loop assessment provides practical validation that minor imperfections do not substantially impair diagnostic utility, supporting real-world applicability. Overall, our approach offers a range of image recognition tasks (ResNet18, YOLOv5, and YOLOv11) for the automatic identification of Candida morphology using mobile devices coupled with a conventional light microscope.

To our knowledge, this is the first study to develop and validate deep learning models for recognizing fungal elements in microscopic images acquired directly via mobile devices under everyday conditions. Although smartphone photography of microscopic fields has been increasingly explored for various biomedical applications (Soleimani et al., 2025; Hui et al., 2025; Soe et al., 2025, Soe et al., 2023), translating mobile microscopy to clinical practice remains challenging due to the substantial variability in how images are captured. In real clinical and educational settings, smartphone and tablet microscope images are often taken freehand (without a stabilizing adaptor), resulting in inconsistent focus, illumination, magnification, and field alignment. These issues are compounded by the wide range of device models, camera specs, and microscope optics that might be used across different settings. Prior studies have mitigated these factors by using phone-to-microscope adaptors to stabilize image capture (Dacal et al., 2021). In contrast, our study intentionally embraced the variability of true point-of-care usage: images were generated by many individuals (medical students) using unassisted handheld devices during routine lab sessions. This has direct implications for clinical implementation in low- and middle-income countries, remote clinics, and other point-of-care environments where standard digital microscopy and expert mycologists may not be available. Importantly, leveraging ubiquitous mobile devices means our approach can be adopted without the need for specialized hardware. Our work has the potential to strengthen global health diagnostics by improving accessibility and bridging gaps in infrastructure and expertise.

It should be emphasized that detection of fungal elements by our current model does not equate to a clinical diagnosis; it does not differentiate between colonization and infection, nor does it quantify fungal burden in a clinically actionable way. Instead, these findings should be regarded as a screening aid to inform, but not determine, clinical decision-making. Clinicians are advised to interpret the results in conjunction with clinical context and further diagnostic workup. Future studies could expand this fungal detection foundation into broader diagnostic applications. We are currently developing a separate deep learning model designed to distinguish between healthy and infected states, as well as other causes of vaginal discharge syndrome. For instance, integrating our model with algorithms recognizing other vaginal discharge features—such as clue cells (BV), motile Trichomonas (trichomoniasis), or intracellular diplococci (gonorrhea)—could enable comprehensive vaginal infection screening. Additionally, combining fungal detection with Lactobacillus quantification could potentially improve diagnostic accuracy, as Lactobacilli are typically abundant in healthy samples but diminished in VVC. This multimodal approach could enhance discrimination between VVC, normal physiology, and other infections.

Beyond vaginal candidiasis, the trained model weights offer transfer learning potential for other fungal imaging tasks. The progressive training strategy—from high-quality to mobile images—produced robust features adaptable to different clinical contexts, such as detecting Fusarium or Aspergillus in corneal smears (fungal keratitis), or dermatophytes in skin/nail scrapings (ringworm, onychomycosis). Future research could leverage these models as starting points for other mycoses, potentially accelerating AI development across medical mycology applications.

We acknowledge several limitations in our study. First, the sample collection was from a single center and primarily involved images taken by students in an educational setting. While this provided a convenient diversity of image conditions, it may not capture the full variability of clinical practice (e.g., images from other hospitals or regions). Nevertheless, within a teaching context, our approach met its goal of exposing students to authentic clinical specimens and preparing them for real-world scenarios. Second, we only used Gram-stained slides; in practice, wet-mount (saline or KOH prep) microscopy is also common for diagnosing vaginal infections. Future work should include other preparation methods to broaden the model’s applicability. Third, our binary classification and detection tasks treated all fungal morphologies (yeast, budding yeast, hyphae) as one “fungus” category versus background. While this simplification was useful for an initial proof-of-concept, it does not differentiate between the forms of Candida. In future iterations, the framework could be refined to classify specific fungal morphologies or even identify species of Candida (Cuevas-Tello et al., 2025), given sufficient data. Transitioning this platform from an educational prototype to a clinical diagnostic tool will also involve collecting more image data that reflect patient-level variability (different patient populations, varying specimen types, use of both Gram stain and wet prep, etc.). Expanding the training data in this manner would likely improve the robustness, generalizability, and clinical utility of the deep learning models.

5. Conclusion

We present a pioneering deep learning framework for morphology-based recognition of fungal elements in mobile device-acquired microscopic images. The fungal classification, detection, and segmentation models demonstrate strong performance on images captured by students in medical training settings, indicating potential for clinical application. This work establishes a foundation for broader multimodal AI platforms to aid diagnosis of vaginal discharge syndromes and other fungal infections. Conventional light microscopes can be effectively integrated with smartphone or tablet cameras and AI algorithms, bridging traditional microscopy with modern digital diagnostics. This advancement is especially valuable in resource-limited environments, enabling applications that span from medical education to clinical diagnostics. Ultimately, our approach contributes to a practical solution that supports improved surveillance, patient care, and research of pathogenic fungi. Our work aligns closely with the collective goal of developing innovative tools and strategies to confront the growing global challenge of fungal infections.

Acknowledgments

We would like to thank ERAWAN HPC, Information Technology Service Center (ITSC) of Chiang Mai University, for their support in the high-performance computing resources.

Funding Statement

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by Learning Innovation Research – Teaching and Learning Innovation Center, Chiang Mai University, grant number R68IN00537 to TW, the Faculty of Medicine Research Fund, grant no. 138–2568 to TW.

Footnotes

Edited by: Mariusz Dyląg, University of Warsaw, Poland

Reviewed by: Pierre R. Marcoux, CEA LETI, France

Jumana Waleed, University of Diyala, Iraq

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material.

Ethics statement

The studies involving humans were approved by institutional ethics committee (Approval No. 441/2568). The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements.

Author contributions

MP: Methodology, Conceptualization, Writing – review & editing, Investigation. SS: Methodology, Writing – review & editing, Data curation, Investigation, Conceptualization. SC: Writing – review & editing, Data curation. SW: Writing – review & editing, Methodology. KP: Methodology, Writing – review & editing, Conceptualization. NL: Methodology, Writing – review & editing, Validation. PU: Investigation, Supervision, Methodology, Writing – review & editing, Conceptualization, Writing – original draft. TW: Investigation, Conceptualization, Funding acquisition, Writing – original draft, Writing – review & editing, Visualization, Methodology, Formal analysis, Data curation, Supervision.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fcimb.2026.1787545/full#supplementary-material

Image1.pdf (11.8MB, pdf)
Image2.pdf (5.2MB, pdf)
Table1.docx (37.2KB, docx)
Table2.pdf (7.1MB, pdf)

References

  1. Anderson M. R., Klink K., Cohrssen A. (2004). Evaluation of vaginal complaints. Jama 291, 1368–1379. doi:  10.1001/jama.291.11.1368, PMID: [DOI] [PubMed] [Google Scholar]
  2. Beder D., Esenkaya Taşbent F., Kiliç Hamzaoğlu F., Türen Demir E., Özdemir M., Arslan G. K. (2025). Investigation of causative agents of vaginitis in symptomatic and asymptomatic women in Konya, Turkey. Parasitologia 5, 15. doi:  10.3390/parasitologia5020015, PMID: 41725453 [DOI] [Google Scholar]
  3. Bettauer V., Costa A., Omran R. P., Massahi S., Kirbizakis E., Simpson S., et al. (2022). A Deep learning approach to capture the essence of Candida albicans morphologies. Microbiol. Spectr. 10, e0147222. doi:  10.1128/spectrum.01472-22, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Chattopadhay A., Sarkar A., Howlader P., Balasubramanian V. N. (2018). “ Grad-cam++: Generalized gradient-based visual explanations for deep convolutional networks,” in 2018 IEEE winter conference on applications of computer vision (WACV). 839–847 (Piscataway, NJ: IEEE; ). [Google Scholar]
  5. Cuevas-Tello J. C., Monreal-De La Rosa A., Quistian-Navarro J. L., Martinez-Gamboa A., González-Lara M. F., López-García N. I., et al. (2025). Deep learning for the identification of Candida spp. directly from blood culture gram stains from candidemia patients. Med. Mycol 63, myaf097. doi:  10.1093/mmy/myaf097, PMID: [DOI] [PubMed] [Google Scholar]
  6. Dacal E., Bermejo-Peláez D., Lin L., Álamo E., Cuadrado D., Martínez Á., et al. (2021). Mobile microscopy and telemedicine platform assisted by deep learning for the quantification of Trichuris trichiura infection. PLoS Negl. Trop. Dis. 15, e0009677. doi:  10.1371/journal.pntd.0009677, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Gonçalves B., Ferreira C., Alves C. T., Henriques M., Azeredo J., Silva S. (2016). Vulvovaginal candidiasis: Epidemiology, microbiology and risk factors. Crit. Rev. Microbiol. 42, 905–927. doi:  10.3109/1040841X.2015.1091805, PMID: [DOI] [PubMed] [Google Scholar]
  8. He K., Zhang X., Ren S., Sun J. (2016). “ Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Piscataway, NJ, USA: IEEE; ). 770–778. [Google Scholar]
  9. Hui S., Xie J., Dong L., Wei L., Dai W., Li D. (2025). Deep learning-based mobile application for efficient eyelid tumor recognition in clinical images. NPJ Digit Med. 8, 185. doi:  10.1038/s41746-025-01539-9, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Kim J., Boo J., Park C. O. (2024). Artificial intelligence applications in medical mycology: Current and Future. J. Mycology Infection 29, 85–91. doi:  10.17966/JMI.2024.29.3.85 [DOI] [Google Scholar]
  11. Koo T., Kim M. H., Jue M. S. (2021). Automated detection of superficial fungal infections from microscopic images through a regional convolutional neural network. PLoS One 16, e0256290. doi:  10.1371/journal.pone.0256290, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Lidong H., Wei Z., Jun W., Zebin S. (2015). Combination of contrast limited adaptive histogram equalisation and discrete wavelet transform for image enhancement. IET image Process. 9, 908–915. doi:  10.1049/iet-ipr.2015.0150, PMID: 40688697 [DOI] [Google Scholar]
  13. Lin T.-Y., Maire M., Belongie S., Hays J., Perona P., Ramanan D., et al. (2014). “ Microsoft COCO: Common objects in context,” in Computer Vision – ECCV 2014. Lecture Notes in Computer Science, vol. 8693, eds. Fleet D., Pajdla T., Schiele B., Tuytelaars T., 740–755. doi:  10.1007/978-3-319-10602-1_48, PMID: [DOI] [Google Scholar]
  14. Nambiar M., Varma S. R., Jaber M., Sreelatha S. V., Thomas B., Nair A. S. (2021). Mycotic infections - mucormycosis and oral candidiasis associated with Covid-19: a significant and challenging association. J. Oral. Microbiol. 13, 1967699. doi:  10.1080/20002297.2021.1967699, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Nguyen T. B., Nguyen H. B., Le T. X., Bui T. H., Nguyen L. S., Nguyen T. H., et al. (2025). Applying machine learning with MobileNetV2 model for rapid screening of vaginal discharge samples in vaginitis diagnosis. Sci. Rep. 15, 19171. doi:  10.1038/s41598-025-04626-9, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Noble S. M., Gianetti B. A., Witchley J. N. (2017). Candida albicans cell-type switching and functional plasticity in the mammalian host. Nat. Rev. Microbiol. 15, 96–108. doi:  10.1038/nrmicro.2016.157, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Paladine H. L., Desai U. A. (2018). Vaginitis: diagnosis and treatment. Am. Fam Physician 97, 321–329. [PubMed] [Google Scholar]
  18. Qi W., Li H., Wang C., Li H., Zhang B., Dong M., et al. (2021). Recent advances in presentation, diagnosis and treatment for mixed vaginitis. Front. Cell Infect. Microbiol. 11, 759795. doi:  10.3389/fcimb.2021.759795, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Ren R., Tan W., Chen S., Xu X., Zhang D., Chen P., et al. (2025). Deep learning application to hyphae and spores identification in fungal fluorescence images. Sci. Rep. 15, 27222. doi:  10.1038/s41598-025-11228-y, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Roselletti E., Perito S., Sabbatini S., Monari C., Vecchiarelli A. (2019). Vaginal epithelial cells discriminate between yeast and hyphae of Candida albicans in women who are colonized or have vaginal candidiasis. J. Infect. Dis. 220, 1645–1654. doi:  10.1093/infdis/jiz365, PMID: [DOI] [PubMed] [Google Scholar]
  21. Sandler M., Howard A., Zhu M., Zhmoginov A., Chen L.-C. (2018). “ Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Piscataway, NJ, USA: IEEE; ). 4510–4520. [Google Scholar]
  22. Selvaraju R. R., Cogswell M., Das A., Vedantam R., Parikh D., Batra D. (2017). “ Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE international conference on computer vision. (Piscataway, NJ, USA: IEEE; ). 618–626. [Google Scholar]
  23. Shroff S. (2023). Infectious vaginitis, cervicitis, and pelvic inflammatory disease. Med. Clin. North Am. 107, 299–315. doi:  10.1016/j.mcna.2022.10.009, PMID: [DOI] [PubMed] [Google Scholar]
  24. Sobel J. D. (1997). Vaginitis. N Engl. J. Med. 337, 1896–1903. doi:  10.1056/NEJM199712253372607, PMID: [DOI] [PubMed] [Google Scholar]
  25. Soe N. N., Latt P. M., Lee D. M., Yu Z., Schmidt M., Bissessor M., et al. (2025). Using deep learning systems for diagnosing common skin lesions in sexual health. Commun. Med. (Lond) 5, 452. doi:  10.1038/s43856-025-01144-7, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Soe N., Yu Z., Latt P., Lee D., Ong J., Ge Z., et al. (2023). “ Image capture: AI-assisted sexually transmitted infection diagnosis tool for clinicians in a clinical setting,” in Australasian Sexual and Reproductive Health Conference, Sydney. [Google Scholar]
  27. Soleimani M., Cheung A. Y., Rahdar A., Kirakosyan A., Tomaras N., Lee I., et al. (2025). Diagnosis of microbial keratitis using smartphone-captured images; a deep-learning model. J. Ophthalmic Inflammation Infect. 15, 8. doi:  10.1186/s12348-025-00465-x, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Sudbery P. E. (2011). Growth of Candida albicans hyphae. Nat. Rev. Microbiol. 9, 737–748. doi:  10.1038/nrmicro2636, PMID: [DOI] [PubMed] [Google Scholar]
  29. Sudbery P., Gow N., Berman J. (2004). The distinct morphogenic states of Candida albicans. Trends Microbiol. 12, 317–324. doi:  10.1016/j.tim.2004.05.008, PMID: [DOI] [PubMed] [Google Scholar]
  30. Tan M., Le Q. (2019). “ EfficientNet: Rethinking model scaling for convolutional neural networks,” in Proceedings of the 36th International Conference on Machine Learning (ICML 2019), vol. 97 of Proceedings of Machine Learning Research. (Long Beach, CA, USA: PMLR; ) 6105–6114. Available online at: https://proceedings.mlr.press/v97/tan19a.html [Google Scholar]
  31. Wang Z., Wang R., Guo H., Zhao Q., Ren H., Niu J., et al. (2025). AI-assisted diagnosis of vulvovaginal candidiasis using cascaded neural networks. Microbiol. Spectr. 13, e0169124. doi:  10.1128/spectrum.01691-24, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Wangsanut T., Pongpom M., Upaphong P. (2025). YOLOv5 Model for Fungal Element Detection (bestYeast701.pt). Hugging Face. Available online at: https://huggingface.co/GiftWangsanut/bestYeast701 (Accessed February 26, 2026). [Google Scholar]
  33. Workowski K. A., Bachmann L. H., Chan P. A., Johnston C. M., Muzny C. A., Park I., et al. (2021). Sexually transmitted infections treatment guidelines 2021. MMWR Recomm Rep. 70, 1–187. doi:  10.15585/mmwr.rr7004a1, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Xu J., Sha J., Jia S., Li J., Xu L., Shao Z. (2025). Effectiveness of a 5G local area network-based digital microscopy interactive system: Quasi-experimental design. JMIR Med. Educ. 11, e70256. doi:  10.2196/70256, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Yadav G., Maheshwari S., Agarwal A. (2014). Contrast limited adaptive histogram equalization based enhancement for real time video system. 2014 international conference on advances in computing, communications and informatics (ICACCI) (Piscataway, NJ: IEEE; ), 2392–2397. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Image1.pdf (11.8MB, pdf)
Image2.pdf (5.2MB, pdf)
Table1.docx (37.2KB, docx)
Table2.pdf (7.1MB, pdf)

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material.


Articles from Frontiers in Cellular and Infection Microbiology are provided here courtesy of Frontiers Media SA

RESOURCES