Abstract
Quantifying ear deformity using linear measurements and mathematical modeling is difficult due to the ear’s complex shape. Machine learning techniques, such as convolutional neural networks (CNNs), are well-suited for this role. CNNs are deep learning methods capable of finding complex patterns from medical images, automatically building solution models capable of machine diagnosis. In this study, we applied CNN to automatically identify ear deformity from 2D photographs. Institutional review board (IRB) approval was obtained for this retrospective study to train and test the CNNs. Photographs of patients with and without ear deformity were obtained as standard of care in our photography studio. Profile photographs were obtained for one or both ears. A total of 671 profile pictures were used in this study including: 457 photographs of patients with ear deformity and 214 photographs of patients with normal ears. Photographs were cropped to the ear boundary and randomly divided into training (60%), validation (20%), and testing (20%) datasets. We modified the softmax classifier in the last layer in GoogLeNet, a deep CNN, to generate an ear deformity detection model in Matlab. All images were deemed of high quality and usable for training and testing. It took about 2 hours to train the system and the training accuracy reached almost 100%. The test accuracy was about 94.1%. We demonstrate that deep learning has a great potential in identifying ear deformity. These machine learning techniques hold the promise in being used in the future to evaluate treatment outcomes.
Subject terms: Diagnosis, Paediatric research
Introduction
Congenital auricular deformities occur in 5% of the pediatric population. Patients with ear deformity may undergo neonatal ear molding or surgical correction to improve ear aesthetics and quality of life1,2. Plastic surgeons often rely on 2D photography to document the severity of disease and assess treatment outcomes.
There are several approaches to treat patients with ear deformity including surgical3,4 and nonsurgical intervention5,6. Otoplasty is often performed during childhood through adulthood or when the ear has reached its full size. Nonsurgical methods, such as splinting or rigid ear molding, take advantage of the plasticity of auricular cartilage during the neonatal period, correcting the deformity over the first month of life5. Early identification of ear deformity is crucial for the success of non-invasive ear molding7. At birth, high levels of circulating maternal estrogens result in increased levels of hyaluronic acid in ear cartilage, which increases the ear’s malleability and plasticity. These levels are highest in the 72 hours after birth and drop rapidly over the first 6 weeks of life8. As the estrogen level drops, the auricular cartilage stiffens, which prevents neonatal ear molding in achieving a long-term improvement. An alternative approach would require surgical correction1,2. To date, there are no reliable objective methods1,2,9 to evaluate ear shape anomalies and evaluation relies on subjective assessment. Therefore, an objective measure to assess ear deformity during the first month of life would be useful for practitioners and parents.
There are several types of congenital ear deformity, including Stahl’s ear, cup ear, and cryptotia, that can affect one or several structures of the ear. Developing mathematical modeling to identify ear deformity can be difficult due to the ear’s complex shape and composition. One solution is machine learning, which automatically recognizes patterns in the training data and builds models to predict future outputs. It is an innovative technique that has been shown to augment human intuition in data analysis. In particular, machine learning has been applied to several divisions of medicine including radiology10–13, ophthalmology14,15, dermatology16,17, and plastic surgery18,19 to detect patterns in data and assist in predicting disease or treatment outcome.
Deep learning, a branch of machine learning, has been equipped with feasible computational algorithms and its application has been successful in ear recognition20,21. The success comes from its deep layer structure. Among deep learning methods, convolutional neural networks (CNNs) excel in image analysis22 as they can capture local information within an image while reducing the complexity of the model. Therefore, CNNs have been successful in medical image analysis to automatically classify disease16,23,24 and segment anatomical regions25,26.
Training CNNs can be performed from scratch or by transfer learning. Training from scratch requires a large set of labeled training data, on the order of 1000 images per class, which is often lacking in the healthcare domain. In addition, training data this large is expensive and cumbersome to build27. In subspecialized medicine, such as plastic surgery, there is a lack of large number of standardized and annotated datasets due to the low number of patients seen at these clinics. Transfer learning allows for training with a source task if the target task is similar27,28, shrinking the necessary labeled training data. GoogLeNet has been pre-trained with 1.2 million images with more than 1000 object categories, and, therefore, has learned extensive features for a wide range of images.
In this study, we optimized a pre-trained CNN model, GoogLeNet29, to classify ears from 2D photographs as normal or abnormal. This study aims to evaluate the performance of CNN in classifying ear abnormality when compared to ground truth clinical diagnosis.
Results
GoogLeNet was trained with 60% of the 671 photographs (274 abnormal vs 128 normal ears). Sample photographs of the training data can be seen in Fig. 1. It took about 2 hours to train GoogLeNet CNNs. After the training was completed, we tested our CNN model using the remaining 92 abnormal and 43 normal ears (Fig. 2).
Figure 1.
Sample photographs used to train the CNN model. A total of 274 photographs of abnormal ears (top panel) and 128 photographs of normal ears (bottom panel) were used.
Figure 2.
Sample photographs used to validate the CNN model. A total of 92 photographs of abnormal ears (top panel) and 43 photographs of normal ears (bottom panel) were used.
Overall, our deep CNN model achieved 94.1% accuracy. In addition, the model achieved high precision 93.8%, sensitivity 97.8% and specificity 86.0%.
Of the 92 abnormal ears, the model classified 90 photographs correctly but misclassified 2 photographs. In addition, the model correctly classified 37 out the 43 normal ears. The misclassified photographs can be seen in Fig. 3.
Figure 3.
Misclassified photographs. The CNN model misdiagnosed 2 abnormal ears as normal (top row) and 6 normal ears as abnormal (bottom rows). The CNN model achieved 94.1% accuracy.
Discussion
In this study, we modified the GoogLeNet model to identify ear abnormalities from 2D photographs obtained for clinical practices. While these photographs are taken according to standard clinical photography30, they exhibit variability in factors such as the angle and zoom. Our CNN model identified ear abnormality with a high accuracy of 94.1%.
To our knowledge, this is the first study to use CNN to identify congenital ear abnormality. However, several mathematical models (edge detection, shape model and iterative closest point) have been developed for healthy ear detection or ear biometric recognition in adults20,31. While some models require user interaction for ear detections32,33, others developed automatic ear identification using image ray transform34. In addition, 3D images have been used for ear detection based on contour matching (achieved a 90.9% detection rate)35 or histograms of categorized shapes (achieved a 100% detection rate)36.
CNN has also been used for ear recognition from 2D photographs20,37. CNN achieved better accuracy (84.8%) when compared to traditional computer vision systems that uses feature extraction algorithm such as principal component analysis (accuracy 76.8%) and speeded up robust features (accuracy 76.8%)20.
The ability of artificial neural networks (ANN) to identify disease has been compared to classic methods in neurosurgery, oncology, and plastic surgery. A study using CNN to classify skin cancer achieved similar accuracy to board-certified dermatologists16. In 2018, the FDA approved the first medical device using deep learning to detect diabetic retinopathy. In neurosurgery, ANN accurately predicted endoscopic 3rd ventriculostomy success38, patient satisfaction after laminectomy for lumbar spinal stenosis39, and in-hospital mortality in patients with traumatic brain injury40. In plastic surgery research, ANN predicted (accuracy 96%) whether a burn would heal within 14 days41.
This study shows that deep learning, such as a CNN, can provide objective assessment of ear deformity during the first month of life. A shared algorithm, if universally deemed accurate in identifying pathology and assessing treatment outcome, would be invaluable in making a diagnosis. Furthermore, computer programs or mobile applications equipped with deep learning models can potentially benefit early detection of disease (during the first month of life) outside of the clinic. For example, a parent or medical provider of a neonate who was unsure of whether there was an abnormality could submit a photograph for automated diagnosis.
This study’s neural network placed ears in the category of “normal” or “abnormal” as we do not have large annotated data available at our center to classify more specific pathology with respect to location on the auricle (e.g. lobule deformity, prominent ears, ear clefts, cup ear deformities, etc.)42. A future study might require a multicenter collaboration to obtain a sufficiently large series of data to train and test a neural network.
Conclusion
The machine learning algorithm we used can extract features from pictures of ears and diagnose normal vs. abnormal with high accuracy. With further research, this could be a standardized tool to objectively evaluate ear intervention outcomes.
Materials and Methods
This study was approved by the Institutional Review Board (IRB) at UT Southwestern Medical Center, and it was carried out in accordance with IRB guidelines and regulations. The IRB approved a waiver of informed consent given that our study is a retrospective review. After obtaining Institutional Review Board (IRB) approval, we retrieved 2D photographs of normal and abnormal ears that were taken as standard of care in our plastic and reconstructive surgery photography studio between 2009 and 2017. The photographs were retrieved using our database search engine43 to train and test the deep learning networks. Ears were labeled as “normal” or “abnormal” for the training and validation sets, based on the practitioner’s documentation of an ear molding intervention from the cohort of children who visit the Fogelson Plastic and Craniofacial surgery clinic for ear molding evaluation. Lateral view photographs were taken of one or both ears using a Nikon D90 with a Nikkor 24–85 mm f/3.5–4.5 lens. A PocketWizard transmitter and receiver were used to trigger the strobes.
A total of 671 ears (left or right side) were used, classified as follows: 457 photographs of patients with ear deformity and 214 photographs of patients with normal ears. Photographs were cropped to the ear boundary at a 4:5 aspect ratio and randomly divided into training (60%), validation (20%), and testing (20%) datasets. The model uses the pre-trained GoogLeNet architecture in Matlab. We modified the softmax classifier in the last layer of GoogleNet to perform binary classification (normal and abnormal). During the training, each training image was randomly scaled and translated per epoch to overcome an overfitting problem often caused by small training datasets. We optimized the batch size, number of epochs, and learning rate hyperparameters using the validation images. After testing several combinations of hyperparameters, we selected hyperparameters of batch size = 50; number of epochs = 300; and learning rate = 1e-4. Machine learning analyses were performed using an EVGA GeForce GTX 1080 with 8 GB onboard memory. The sensitivity, specificity, accuracy and precision were calculated.
Institutional review board statement
IRB has approved this study.
Acknowledgements
The study was supported in part by infrastructure provided by the Children’s Analytical Imaging and Modeling Center Research Program. J.L. was supported by the Cancer Prevention Research Institute (CPRIT) (RP150596).
Author contributions
R.H. conceptualized and designed the study, collected data, developed the computer algorithms, wrote the manuscript, and approved the final manuscript as submitted. J.L. developed the computer algorithms, revised the manuscript, and approved the final manuscript as submitted. M.P., J.S. and A.K. interpreted the data, revised the manuscript, and approved the final manuscript as submitted.
Data availability
Data from this study are available to interested readers upon reasonable request.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Byrd HS, Langevin C-J, Ghidoni LA. Ear Molding in Newborn Infants with Auricular Deformities. Plastic and Reconstructive Surgery. 2010;126:1191–1200. doi: 10.1097/PRS.0b013e3181e617bb. [DOI] [PubMed] [Google Scholar]
- 2.Daniali LN, et al. Classification of Newborn Ear Malformations and their Treatment with the EarWell Infant Ear Correction System. Plastic and Reconstructive Surgery. 2017;139:681–691. doi: 10.1097/prs.0000000000003150. [DOI] [PubMed] [Google Scholar]
- 3.Stewart KJ, Lancerotto L. Surgical Otoplasty: An Evidence-Based Approach to Prominent Ears Correction. Facial Plastic Surgery Clinics of North America. 2018;26:9–18. doi: 10.1016/j.fsc.2017.09.002. [DOI] [PubMed] [Google Scholar]
- 4.van Hövell tot Westerflier CVA, van Wijk MP, Kon M. Surgical Correction of the “Sunken Ear”: An Auricular Positional Change after Canal Wall Down Mastoidectomy. Otolaryngology–Head and Neck Surgery. 2016;154:1161–1163. doi: 10.1177/0194599816644589. [DOI] [PubMed] [Google Scholar]
- 5.Schultz K, Guillen D, Maricevich RS. Newborn Ear Deformities: Early Recognition and Novel Nonoperative Techniques. Seminars in plastic surgery. 2017;31:141–145. doi: 10.1055/s-0037-1603958. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.van Wijk MP, Breugem CC, Kon M. Non-surgical correction of congenital deformities of the auricle: A systematic review of the literature. Journal of Plastic, Reconstructive & Aesthetic Surgery. 2009;62:727–736. doi: 10.1016/j.bjps.2009.01.020. [DOI] [PubMed] [Google Scholar]
- 7.Doft MA, et al. The Newborn Butterfly Project: A Shortened Treatment Protocol for Ear Molding. Plastic and Reconstructive Surgery. 2015;135:577e–583e. doi: 10.1097/prs.0000000000000999. [DOI] [PubMed] [Google Scholar]
- 8.Anstadt EE, Johns DN, Kwok AC-M, Siddiqi F, Gociman B. Neonatal Ear Molding: Timing and Technique. Pediatrics. 2016;137:e20152831. doi: 10.1542/peds.2015-2831. [DOI] [PubMed] [Google Scholar]
- 9.Petersson RS, Recker CA, Martin JRK, Driscoll CLW, Friedman O. Identification of congenital auricular deformities during newborn hearing screening allows for non-surgical correction: A Mayo Clinic pilot study. International Journal of Pediatric Otorhinolaryngology. 2012;76:1406–1412. doi: 10.1016/j.ijporl.2012.06.011. [DOI] [PubMed] [Google Scholar]
- 10.Cheng PM, Malhi HS. Transfer Learning with Convolutional Neural Networks for Classification of Abdominal Ultrasound Images. Journal of Digital Imaging. 2017;30:234–243. doi: 10.1007/s10278-016-9929-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Chen Y, et al. Automated quantification of cerebral edema following hemispheric infarction: Application of a machine-learning algorithm to evaluate CSF shifts on serial head CTs. NeuroImage: Clinical. 2016;12:673–680. doi: 10.1016/j.nicl.2016.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Tixier F, et al. Intratumor heterogeneity characterized by textural features on baseline (18)F-FDG PET images predicts response to concomitant radiochemotherapy in esophageal cancer. Journal of Nuclear Medicine. 2011;52:369–378. doi: 10.2967/jnumed.110.082404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.McBee MP, et al. Deep Learning in Radiology. Academic Radiology. 2018 doi: 10.1016/j.acra.2018.02.018. [DOI] [PubMed] [Google Scholar]
- 14.Gulshan V, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316:2402–2410. doi: 10.1001/jama.2016.17216. [DOI] [PubMed] [Google Scholar]
- 15.Abràmoff MD, et al. Improved Automated Detection of Diabetic Retinopathy on a Publicly Available Dataset Through Integration of Deep Learning. Investigative Ophthalmology & Visual Science. 2016;57:5200–5206. doi: 10.1167/iovs.16-19964. [DOI] [PubMed] [Google Scholar]
- 16.Esteva A, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542:115. doi: 10.1038/nature21056. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Martini JT, Andreas T, Mette M, Peter B, B.E. JG. Machine-learning classification of non-melanoma skin cancers from image features obtained by optical coherence tomography. Skin Research and Technology. 2008;14:364–369. doi: 10.1111/j.1600-0846.2008.00304.x. [DOI] [PubMed] [Google Scholar]
- 18.Cho M-J, et al. Quantifying Normal Craniofacial Form and Baseline Craniofacial Asymmetry in the Pediatric Population. Plastic and Reconstructive Surgery. 2018;141:380e–387e. doi: 10.1097/prs.0000000000004114. [DOI] [PubMed] [Google Scholar]
- 19.Cho M-J, Kane AA, Seaward JR, Hallac RR. Metopic “ridge” vs. “craniosynostosis”: Quantifying severity with 3D curvature analysis. Journal of Cranio-Maxillofacial Surgery. 2016;44:1259–1265. doi: 10.1016/j.jcms.2016.06.019. [DOI] [PubMed] [Google Scholar]
- 20.Galdámez PL, Raveane W, González Arrieta A. A brief review of the ear recognition process using deep neural networks. Journal of Applied Logic. 2017;24:62–70. doi: 10.1016/j.jal.2016.11.014. [DOI] [Google Scholar]
- 21.Tian, L. & Mu, Z. Ear recognition based on deep convolutional network. 9th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), 437–441, 10.1109/CISP-BMEI.2016.7852751 (2016).
- 22.Tajbakhsh N, et al. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Transactions on Medical Imaging. 2016;35:1299–1312. doi: 10.1109/TMI.2016.2535302. [DOI] [PubMed] [Google Scholar]
- 23.Anthimopoulos M, Christodoulidis S, Ebner L, Christe A, Mougiakakou S. Lung Pattern Classification for Interstitial Lung Diseases Using a Deep Convolutional Neural Network. IEEE Transactions on Medical Imaging. 2016;35:1207–1216. doi: 10.1109/TMI.2016.2535865. [DOI] [PubMed] [Google Scholar]
- 24.Cheng, D. & Liu, M. 106–113 (Springer International Publishing).
- 25.Moeskops P, et al. Automatic Segmentation of MR Brain Images With a Convolutional Neural Network. IEEE Transactions on Medical Imaging. 2016;35:1252–1261. doi: 10.1109/TMI.2016.2548501. [DOI] [PubMed] [Google Scholar]
- 26.Havaei M, et al. Brain tumor segmentation with Deep Neural Networks. Medical Image Analysis. 2017;35:18–31. doi: 10.1016/j.media.2016.05.004. [DOI] [PubMed] [Google Scholar]
- 27.Pan SJ, Yang Q. A Survey on Transfer Learning. IEEE Transactions on Knowledge and Data Engineering. 2010;22:1345–1359. doi: 10.1109/tkde.2009.191. [DOI] [Google Scholar]
- 28.Lisa, T. & Jude, S. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques (eds Olivas Emilio Soria et al.) 242–264 (IGI Global, 2010).
- 29.Szegedy, C. et al. Going deeper with convolutions. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1–9, 10.1109/CVPR.2015.7298594 (2015).
- 30.Yavuzer R, Smirnes S, Jackson IT. Guidelines for Standard Photography in Plastic Surgery. Annals of Plastic Surgery. 2001;46:293–300. doi: 10.1097/00000637-200103000-00016. [DOI] [PubMed] [Google Scholar]
- 31.Pflug A, Busch C. Ear biometrics: a survey of detection, feature extraction and recognition methods. IET Biometrics. 2012;1:114–129. doi: 10.1049/iet-bmt.2011.0003. [DOI] [Google Scholar]
- 32.Burge M, Burger W. Ear biometrics in computer vision. Proceedings 15th International Conference on Pattern Recognition. ICPR-2000. 2000;2:822–826. doi: 10.1109/ICPR.2000.906202. [DOI] [Google Scholar]
- 33.Ping, Y. & Bowyer, K. Empirical Evaluation of Advanced Ear Biometrics. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops, 41–41 (2005).
- 34.Cummings, A. H., Nixon, M. S. & Carter, J. N. A novel ray analogy for enrolment of ear biometrics. Fourth IEEE International Conference onBiometrics: Theory, Applications and Systems (BTAS), 1–6 (2010).
- 35.Chen H, Bhanu B. Contour Matching for 3D Ear Recognition. Seventh IEEE Workshops on Applications of Computer Vision (WACV/MOTION'05) - Volume 1. 2005;1:123–128. doi: 10.1109/ACVMOT.2005.38. [DOI] [Google Scholar]
- 36.Zhou, J., Cadavid, S. & Abdel-Mottaleb, M. Histograms of Categorized Shapes for 3D ear detection. Fourth IEEE International Conference onBiometrics: Theory, Applications and Systems (BTAS), 1–6 (2010).
- 37.Ž, E., Štepec, D., Štruc, V. & Peer, P. Training Convolutional Neural Networks with Limited Training Data for Ear Recognition in the Wild. 12thIEEE International Conference on Automatic Face & Gesture Recognition (FG2017), 987–994, 10.1109/FG.2017.123 (2017).
- 38.Azimi Parisa, Mohammadi Hasan Reza. Predicting endoscopic third ventriculostomy success in childhood hydrocephalus: an artificial neural network analysis. Journal of Neurosurgery: Pediatrics. 2014;13(4):426–432. doi: 10.3171/2013.12.PEDS13423. [DOI] [PubMed] [Google Scholar]
- 39.Azimi P, Benzel EC, Shahzadi S, Azhari S, Mohammadi HR. Use of artificial neural networks to predict surgical satisfaction in patients with lumbar spinal canal stenosis: clinical article. Journal of neurosurgery. Spine. 2014;20:300–305. doi: 10.3171/2013.12.spine13674. [DOI] [PubMed] [Google Scholar]
- 40.Shi HY, Hwang SL, Lee KT, Lin CL. In-hospital mortality after traumatic brain injury surgery: a nationwide population-based comparison of mortality predictors used in artificial neural network and logistic regression models. Journal of neurosurgery. 2013;118:746–752. doi: 10.3171/2013.1.jns121130. [DOI] [PubMed] [Google Scholar]
- 41.Yeong EK, Hsiao TC, Chiang HK, Lin CW. Prediction of burn healing time using artificial neural networks and reflectance spectrometer. Burns: journal of the International Society for Burn Injuries. 2005;31:415–420. doi: 10.1016/j.burns.2004.12.003. [DOI] [PubMed] [Google Scholar]
- 42.Weerda H. Classification of congenital deformities of the auricle. Facial plastic surgery: FPS. 1988;5:385–388. doi: 10.1055/s-2008-1064778. [DOI] [PubMed] [Google Scholar]
- 43.Hallac Rami R., Dumas Blaike M., Seaward James R., Herrera Royiel, Menzies Christopher, Kane Alex A. Digital Images in Academic Plastic Surgery: A Novel and Secure Methodology for Use in Clinical Practice and Research. The Cleft Palate-Craniofacial Journal. 2018;56(4):552–555. doi: 10.1177/1055665618787399. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data from this study are available to interested readers upon reasonable request.



