Abstract
Purpose:
To develop an algorithm to automate the organization of large photo databases using the Haar cascade algorithm for face and eye detection and machine learning tools in Python.
Design:
Retrospective study for the purposes of clinical tool development.
Methods:
We developed an algorithm, termed FaceFinder, to identify front facing images in a large dataset of facial, orthodontal and miscellaneous images. FaceFinder works by detecting the presence of faces and at least two eyes using the Haar cascade. Execution time was recorded using different-sized datasets. A total of 895 images were analyzed by FaceFinder using various thresholds for face and eye detection. Precision, recall, specificity, accuracy, and F1 score were computed by comparison to ground truth labels of the images as determined by a human grader.
Results:
Using medium thresholds for face and eye detection, FaceFinder reached recall, accuracy, and F1 score of 89.3%, 91.6%, and 92.9%, respectively with an execution time per image was 0.91 s. Using the highest threshold for face and eye detection, FaceFinder achieved precision and specificity values of 98.3% and 99.2% respectively.
Conclusions:
FaceFinder is capable of sorting through a heterogenous dataset of photos of patients with craniofacial disease and identifying high-quality front-facing facial images. This capability allows for automated sorting of large databases that can facilitate and expedite data preparation for further downstream analyses involving artificial intelligence and facial landmarking.
Keywords: Artificial intelligence, Face detection, Computer vision, Oculoplastic surgery
Precis
FaceFinder is a machine learning based tool for the sorting of large medical facial datasets.
Introduction
Two- and three-dimensional imaging is widely utilized in surgical specialties for diagnostic, planning, and postoperative assessment purposes. Sophisticated techniques, such as photogrammetry, machine learning, and artificial intelligence, facilitate these by creating predictive models, and providing personalized surgical plans.1–4 From pre-operative simulations to post-operative evaluations, computational image analysis is driving a more precise and individualized approach to patient care.3,5 For instance, automated detection of facial landmarks has been leveraged in craniofacial surgery. However, the utility of artificial intelligence (AI) far exceeds the surgical domains.6–8 RetinaFace and MediaPipe are two algorithms developed for robust face detection and facial landmark identification.9,10 Previous reports have demonstrated the use of models like OpenFace to detect pain action units for automatic assessment and grading of pain.11,12 Similar technology has been implemented for quantification of disease severity in Parkinson’s disease, Alzheimer’s disease, and cerebral palsy.13–15 With respect to ophthalmology, much work has been devoted to automating the prediction of periorbital distances and quantifying the state of disease using facial scene recognition via AI.16–19
Despite the promise of AI with regard to image analysis in these fields, the difficulty in streamlined organization and curation of large image datasets still hinders the efficiency of research. Given the volume and variety of facial images obtained in the craniofacial and oculoplastic clinic, considerable manual effort and time is required to prepare the dataset for computational analysis, thereby slowing down the research process and potentially leading to inaccuracies. Authors have previously proposed database schemes for the storage of large datasets.20 Additionally, various AI systems exist to help curate image datasets from large open-source data repositories or via scraping the web.21,22 However, these models were not developed for use in a clinical setting which encompasses thousands of pictures, ranging from intraoperative photos to paperwork. To date, no automated system exists which can parse through and organize facial photographs of patients with craniofacial pathologies.
To address this issue of dataset management for downstream artificial intelligence experiments, we designed an open-source, novel machine-learning pipeline, FaceFinder, to automate the identification and categorization of front-facing facial images from large datasets. FaceFinder is a time-saving tool that will expedite the development of novel insights and therapeutic strategies in the craniofacial and oculoplastic fields via the reduction of data preparation time. FaceFinder has been open sourced to allow further engagement and utility within the community.
Methods
Dataset
This is a retrospective study conducted at a single institution. Approval was obtained from the Institutional Review Board and the study was performed in adherence with the Health Insurance Portability and Accountability Act guidelines and tenets of the Declaration of Helsinki (IRB Number: STUDY2022-0030). All images used in this study were collected from the University of Illinois at Chicago. The original dataset (n = 895) was manually sorted into true front-facing images (n = 388), and negative images (n = 507). The dataset contained photographs from clinical visits of healthy patients mixed with those having craniofacial pathologies affecting the eyes (Supplemental Table 1). The patients were majority female (55%) with an average age of 17.4 ± 5.4 years old. The other images from orthodontia visits, paperwork, intraoperative photographs, and other non-facial images were included. The criteria for a true front facing (positive image) were the presence of a single face and two eyes being observable within the image. An image not meeting these criteria was defined as a negative image. Three human graders created the designations of ground truth labels. A subset of representative images showcasing positive and negative images from the dataset is shown in Fig. 1. An additional dataset of healthy faces, the Chicago Facial Dataset, was also used for external validation.23–25 This open-source dataset consisted of 827 true positive and zero true negative images.
Fig. 1.

A-C) Cropped examples of human designated true front facing (positive image) and d-F) negative images. The criteria for a positive image were the presence of a single face and two eyes being observable within the image. An image not meeting these criteria was defined as a negative image.
Algorithm development
FaceFinder is a Python-based image processing tool developed to automatically identify and manipulate images of faces from a large dataset, streamlining the process of data preparation for further computational analysis. In addition to standard Python libraries, FaceFinder primarily leverages the Haar Cascade for face and eye detection.26–28 The Haar cascade is one of the original machine learning face detection techniques proposed in 2001.29 It works by detecting faces and eyes using the handcrafted Haar features. Through a cascading mechanism described previously, highly accurate face detection can be achieved.30 While convolutional neural networks (CNN) have emerged as state-of-the-art for face detection, the Haar cascade is still widely used due to its high accuracy and low computational cost.21,23,24,31
A complete flow diagram of FaceFinder can be seen in Fig. 2A. The user first defines a source directory containing the images and two destination directories to save the predicted facial images as well as predicted non face images. FaceFinder removes all non-alphanumeric characters from the folder and file names to standardize the inputs to the algorithm. It then identifies photos with faces in them via face detection using the Haar cascade.29 If one or more faces are detected, FaceFinder then applies eye detection via the Haar cascade to each identified face. If at least two eyes are detected in the image, it is deemed as a true front-facing image. For all images meeting these criteria, the face bounding box is expanded by 20% in all directions to include a broader area around the face and the image is cropped using this box, working on the assumption that more relevant content (such as the forehead or ears) is located outside the initial bounding box. Importantly, no further training of the Haar Cascade was performed, other than what was previously published.29 As such, the entire dataset was treated as a testing set and individual train and test sets were not used which is common practice in other machine learning settings.
Fig. 2.

A Flow Diagram of the complete FaceFinder algorithm. A) The files are standardized to have appropriate naming. The algorithm predicts the location of the face and the eyes. It will subsequently crop the image and then copy the EXIF metadata and save the image in the appropriate folder. B) The graphic user interface to set up FaceFinder experiments. Home directory designates the directory containing the image database. A popup window to select the directory is displayed when ‘Select Directory’ is clicked. The minimum eye and face neighbors can be chosen from the dropdown window and ‘Run FaceFinder’ initiates the algorithm. The ‘Stop Processing’ Terminates the program execution.
In the Haar cascade, detections are grouped by proximity, and this hyperparameter determines the number of detections per group required to deem a detection valid. That is, for every group of faces or eyes detected, if there are less than the specified minimum neighbor value of detections, the candidate object will be discarded.28,32 Multiple combinations of face and eye minimum neighbors were evaluated, and the results were recorded. This hyperparameter can be tuned by others based on their specific use case and database size.
The processed images are then saved to the output folder, with the directory structure and original EXIF data of each image preserved. Additionally, if an image has EXIF data, it is used to orient the image in the ‘portrait’ orientation. This protocol ensures the provenance and other contextual information of the images are maintained for subsequent analysis. A simple user interface was designed to input the paths to folders and modify the minimum neighbors value for both eye and face detection (Fig. 2B). This interface can be expanded to include other hyperparameters of the Haar cascade as the code is freely available.
Evaluation metrics
FaceFinder was evaluated by comparing the predictions to the ground truth label of every image in the initial dataset (n = 895). For all combinations of thresholds evaluated, precision, recall, specificity, accuracy, and F1-score were calculated. These metrics were calculated according to Eq. (1). To evaluate the speed of FaceFinder, the algorithm was run on multiple datasets of different sizes and the time to completion was logged.
| (1) |
Additionally, to assess whether or not the predictions from FaceFinder were likely to be achieved by chance, Cohen’s Kappa was calculated according to Eq. (2).33
| (2) |
Hardware
All timing experiments were performed on a 2020 M1 MacBook air using the M1 chip (8 core CPU) with 8GB of RAM.
Results
FaceFinder processed a total of 895 images successfully using different combinations of threshold values. A combination of 1,3,5 and 8 minimum neighbors were evaluated to explore the effect of a large range of values on FaceFinder’s performance. In the standard implementation of the Haar Cascade, the default minimum neighbor value is 3.34 The maximum precision (98.3%) and specificity (99.2%) was achieved when using the highest threshold (minimum neighbors of 8 for both face and eye detection). The maximum recall (89.3%), accuracy (91.6%), and F1-Score (92.9%) was achieved using eye and face minimum neighbor values of 3 and 5 respectively. At this combination of thresholds, the precision and specificity were also reasonably high at 96.8% and 95.3% respectively (Table 1). The eye and face minimum neighbor combination of 3 and 5 also achieved the highest Cohen Kappa value of 0.878.
Table 1:
Results of FaceFinder using various minimum neighbor values in the Haar Cascade. Bold indicates the highest percentage for any given category. A higher minimum neighbor value increases the threshold of detection. We evaluated combinations of 1,3,5 and 8 minimum neighbors to explore the effect of a large range of values on FaceFinder’s performance. In the standard implementation of the Haar Cascade, the default minimum neighbor value is 3. The utilization of 3 minimum eye neighbors and 5 minimum 5 face neighbors resulted in the highest recall, accuracy, and F1 Score. F1 score is the harmonic mean of precision and recall. The equations for all metrics can be seen in Equation 1 and 2.
| Eye Min Neighbors | 1 | 1 | 3 | 5 | 8 | 8 |
|---|---|---|---|---|---|---|
| Face Min Neighbors | 1 | 8 | 5 | 3 | 1 | 8 |
| Precision | 80.6% | 97.9% | 96.8% | 96.1% | 95.1% | 98.3% |
| Recall | 82.8% | 68.8% | 89.3% | 67.1% | 68.1% | 58.6% |
| Specificity | 83.8% | 98.8% | 95.3% | 97.8% | 97.2% | 99.2% |
| Accuracy | 83.4% | 85.4% | 91.6% | 84.0% | 84.1% | 81.0% |
| F1 Score | 81.7% | 80.8% | 92.9% | 79.0% | 79.4% | 73.4% |
| Cohen’s Kappa | .671 | .701 | .878 | .672 | .675 | .606 |
To further validate FaceFinder, the pipeline on an open-source dataset (Chicago Facial Dataset) of healthy faces was evaluated with all in the front-facing position.23–25 Minimum neighbor values of 3 and 5 were used for eye and face detection, respectively. With this, the FaceFinder precision, recall, and accuracy score was 1, 0.99, and 0.99, respectively (Supplemental Figure 1).
A selection of false positives is shown in Fig. 3 with the misidentified face and eye regions marked. When using the lowest thresholds (minimum neighbors of 1 for both face and eyes) for detection, features such as braces were detected to be faces and eyes within the image (Fig. 3A). Using minimum neighbors of 3 and 5 for face and eye detection respectively, scars and corners of mouths were erroneously predicted to be eyes resulting in the misclassification of side profile images (Fig. 3B) as front-facing.
Fig. 3.

Examples of false positive images from FaceFinder. A) Orthodontal image where braces were predicted to be faces (green box) and eyes (blue box). 1 minimum neighbor (lowest threshold) was used with for face detection on this image. (B-C) Side profile photographs where face prediction was accurate, but where the corners of the mouth, chin, and a scar were predicted to be eyes (blue box) A minimum neighbors value of 3 for eye detection was used on this image. (blue box),.
Following successful face detection, FaceFinder executed facial cropping and metadata preservation with 100% success rate. FaceFinder had an average processing time of 0.91 s per image determined by evaluating the time to completion on multiple sized datasets (Fig. 4).
Fig. 4.

Time to completion for FaceFinder process datasets with various size image datasets. Average processing rate was 0.91 s/image.
Discussion
FaceFinder demonstrates the capability to efficiently sort through large, heterogeneous datasets of craniofacial photographs with a high degree of accuracy, specificity, precision, and recall. In clinical research settings, relevant front-facing photographs are often mixed in with other miscellaneous biomedical image data acquired over multiple clinical visits. Even in scenarios where clinical datasets only contain pictures of the face, the logistic process of renaming files, identifying only front facing images (as is standard in much of the literature related to facial scene understanding via AI), and organizing file hierarchies still presents a large time burden.16–19 The ability to parse these datasets facilitates more efficient data management and research, especially for analysis of soft tissue facial features.
Multiple studies have evaluated and validated the Haar cascade for facial detection.29,31 Here, we validate these previous studies and show that highly accurate parsing of patients with rare craniofacial pathology can be achieved using simple feature extraction techniques from the Haar cascade, which permit short run times and avoid the reliance on graphic processing units (GPUs) or additional model development. Furthermore, we demonstrate the utility on craniofacial images, a subset that differs from standard evaluation cases in the literature. Also in line with prior literature, we have validated FaceFinder on an open-source dataset of healthy eyes and achieved near perfect performance. FaceFinder is, to our knowledge, the first tool purpose-built to facilitate data for craniofacial and oculoplastic research. Once a dataset has been properly curated and cleaned, further use of AI has been described in the detection and analysis of craniofacial and oculoplastic disease.2,17,35–37 Given the wide variety of dysmorphology present in these patients, development of a program to accurately identify patient images regardless of underlying condition is of paramount importance.
The best results for precision and specificity using FaceFinder were obtained using the largest threshold combination of minimum neighbors for face and eye detection. This is an intuitive finding as higher thresholds will increase the chance of returned predicted positive images being truly positive while avoiding the misclassification of negative images. However, this high threshold combination does come at the cost of recall (58.6%) as many true positive images were not detected. Of all the threshold combinations evaluated, we determined that a lower threshold for eye detection with a medium threshold for face detection (minimum neighbors of 3 and 5 respectively) is the optimal combination. FaceFinder achieved the maximum recall (96.8%), accuracy (91.6%), and F1-Score (92.9%) while still maintaining very high precision and specificity when run with these values. However, the best hyperparameters for FaceFinder will depend on the research needs of downstream users, and the thresholds can easily be modified within the graphical user interface (GUI). If having a very pure end dataset is of importance, a higher minimum neighbors values would be better, but for best overall performance, values in the middle of the range may be best. In this study, only the minimum neighbors hyperparameter was explored as previous literature has suggested this is sufficient to control the number of true positives and negatives in face and eye detection.38 However, there are other hyperparameters in the Haar cascade which can be tuned on an individual basis to achieve optimal results on user-specific datasets in the future.
There are certain limitations in this algorithm that should be discussed. From a clinical logistic point of view, storage strategies of imaging data may need to be altered to be most effectively compatible with FaceFinder. Integration with tools such as FiftyOne could further increase the ease of use the algorithm across multiple clinical sites.39 From an algorithmic point of view, the simple feature extraction leveraged by the Haar cascade makes it prone to misidentifying objects such as gaps in teeth, braces, or scars (Fig. 3). This is most pronounced at lower threshold values, but we observed false positives even at the best threshold combination evaluated. More sophisticated methods, such as CNNs, could result in higher accuracy compared to the Haar cascade which has been reported previously.30 However, the training of CNNs is more computationally expensive and relies on training or finetuning a network on representative image data.30,40 As training data for patients with rare craniofacial pathology is generally limited and GPUs are not always accessible in clinical environments, FaceFinder can serve as a preprocessing step to facilitate these research goals through curation of high-quality representative datasets without the need for excessive finetuning. Future experiments could be performed to incorporate CNNs trained for face detection that can be run on a CPU, such as RetinaFace or Mediapipe, into the pipeline to organize and curate large medical datasets.9,10 Usage of more advanced algorithms may allow for further options in the process of dataset creation such as only selecting side profile images. It would also be interesting to study potential time savings from using such algorithms. Additional limitations are that FaceFinder has not been validated on images of oculoplastic diseases not included in our dataset and we did not evaluate the robustness of FaceFinder to noise that may be present in images captured from different clinical imaging devices.
In conclusion, FaceFinder can aid in efficiently sorting through large, heterogeneous datasets with a high degree of accuracy, specificity, precision, and recall. As such, FaceFinder may be leveraged as a clinical tool for automated sorting of large databases to facilitate downstream analysis of large datasets and expediting data preparation for further analyses involving artificial intelligence and facial landmarking.
Supplementary Material
Funding
This project was supported by the Prevention of Blindness Grant, NIH Grant P30EY001792.
Footnotes
Prior presentations
This work was presented as a poster at the 2023 Association for Research in Vision and Ophthalmology.
Supplemental material
All code is available at https://github.com/monkeygobah/facefinder
Declaration of competing interest
AT is a consultant for Genentech. PS is a consultant for Oyster Point Pharma and the Owner of Lodestone Ophthalmics. The other authors have no financial disclosures or conflicts of interest.
Supplementary materials
Supplementary material associated with this article can be found, in the online version, at doi:10.1016/j.ajoint.2024.100083.
References
- 1.Jarvis T, Thornburg D, Rebecca AM, Teven CM. Artificial intelligence in plastic surgery: current applications, future directions, and ethical implications. Plast Reconstr Surg Glob Open. 2020;8(10):e3200. 10.1097/GOX.0000000000003200. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Cho MJ, Hallac RR, Effendi M, Seaward JR, Kane AA. Comparison of an unsupervised machine learning algorithm and surgeon diagnosis in the clinical differentiation of metopic craniosynostosis and benign metopic ridge. Sci Rep. 2018; 8(1):6312. 10.1038/s41598-018-24756-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Hashimoto DA, Rosman G, Rus D, Meireles OR. Artificial intelligence in surgery: promises and perils. Ann Surg. 2018;268(1):70–76. 10.1097/SLA.0000000000002693. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Douglas TS. Image processing for craniofacial landmark identification and measurement: a review of photogrammetry and cephalometry. Comput Med Imaging Graph. 2004;28(7):401–409. 10.1016/j.compmedimag.2004.06.002. [DOI] [PubMed] [Google Scholar]
- 5.Sun Y, Huang X, Zhang Q, et al. A fully automatic postoperative appearance prediction system for blepharoptosis surgery with image-based deep learning. Ophthalmology Science. 2022;2(3), 100169. 10.1016/j.xops.2022.100169. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Nishimoto S, Saito T, Ishise H, Fujiwara T, Kawai K, Kakibuchi M. Three-dimensional craniofacial landmark detection in series of CT slices using multiphased regression networks. Diagnostics. 2023;13(11):1930. 10.3390/diagnostics13111930. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Mak ML, Al-Shaqsi SZ, Phillips J. Prevalence of machine learning in craniofacial surgery. J Craniofac Surg. 2020;31(4):898–903. 10.1097/SCS.0000000000006234. [DOI] [PubMed] [Google Scholar]
- 8.El Rhazi M, ZARGHILI A, Majda A Automated detection of craniofacial landmarks on a 3D facial mesh. In:; 2020. doi: 10.1007/978-3-030-62199-5_47. [DOI] [Google Scholar]
- 9.Deng J, Guo J, Zhou Y, Yu J, Kotsia I, Zafeiriou S RetinaFace: single-stage dense face localisation in the wild. Published online May 4, 2019. Accessed October 12, 2024. http://arxiv.org/abs/1905.00641. [Google Scholar]
- 10.Kartynnik Y, Ablavatski A, Grishchenko I, Grundmann M Real-time facial surface geometry from monocular video on mobile GPUs. Published online July 15, 2019. doi: 10.48550/arXiv.1907.06724. [DOI] [Google Scholar]
- 11.Baltrusaitis T, Robinson P, Morency LP. OpenFace: an open source facial behavior analysis toolkit. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE; 2016:1–10. 10.1109/WACV.2016.7477553. [DOI] [Google Scholar]
- 12.Nerella S, Cupka J, Ruppert M, Tighe P, Bihorac A, Rashidi P. Pain action unit detection in critically Ill patients. Proc COMPSAC. 2021;2021:645–651. 10.1109/compsac51774.2021.00094. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Nam U, Lee K, Ko H, Lee JY, Lee EC. Analyzing facial and eye movements to screen for Alzheimer’s disease. Sensors (Basel). 2020;20(18):5349. 10.3390/s20185349. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Jin B, Qu Y, Zhang L, Gao Z. Diagnosing Parkinson Disease Through Facial Expression Recognition: video Analysis. J Med Internet Res. 2020;22(7):e18697. 10.2196/18697. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Kim J, Jeong H, Cho J, et al. Numerical approach to facial palsy using a novel registration method with 3D facial landmark. Sensors. 2022;22(17):6636. 10.3390/s22176636. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Rana K, Beecher M, Caltabiano C, et al. Artificial intelligence to automate assessment of ocular and periocular measurements. Eur J Ophthalmol. 2024. 10.1177/11206721241249773. Published online May 611206721241249773. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Chen HC, Tzeng SS, Hsiao YC, Chen RF, Hung EC, Lee OK. Smartphone-based artificial intelligence–assisted prediction for eyelid measurements: algorithm development and observational validation study. JMIR Mhealth Uhealth. 2021;9(10): e32444. 10.2196/32444. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Van Brummen A, Owen JP, Spaide T, et al. PeriorbitAI: artificial intelligence automation of eyelid and periorbital measurements. Am J Ophthalmol. 2021;230:285–296. 10.1016/j.ajo.2021.05.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Nahass GR, Yazdanpanah G, Cheung M, et al. State-of-the-art periorbital distance prediction and disease classification using periorbital features. Published online September 27, 2024. doi: 10.48550/arXiv.2409.18769. [DOI] [Google Scholar]
- 20.Mi W, Jianya G, Deren LI. Design and implementation of large-scale image database management system. whdxxbxxkxb. 2003;28(3):294–300. [Google Scholar]
- 21.Narayanan T, Krishnan A, Koul A, Ganju S Curator: creating large-scale curated labelled datasets using self-supervised learning. Published online December 28, 2022. doi: 10.48550/arXiv.2212.14099. [DOI] [Google Scholar]
- 22.Arsenovic M, Sladojevic S, Anderla A, Stefanovic D, Lalic B. Deep learning powered automated tool for generating image based datasets. In: 2017 IEEE 14th International Scientific Conference on Informatics. 2017:13–17. 10.1109/INFORMATICS.2017.8327214. [DOI] [Google Scholar]
- 23.Ma DS, Correll J, Wittenbrink B. The Chicago face database: a free stimulus set of faces and norming data. Behav Res. 2015;47(4):1122–1135. 10.3758/s13428-014-0532-5. [DOI] [PubMed] [Google Scholar]
- 24.Ma DS, Kantner J, Wittenbrink B. Chicago Face Database: multiracial expansion. Behav Res. 2021;53(3):1289–1300. 10.3758/s13428-020-01482-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Lakshmi A, Wittenbrink B, Correll J, Ma DS. The India face set: international and cultural boundaries impact face impressions and perceptions of category membership. Front Psychol. 2021;12. 10.3389/fpsyg.2021.627678. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Hunter JD. Matplotlib: a 2D graphics environment. Comput Sci Eng. 2007;9(3):90–95. 10.1109/MCSE.2007.55. [DOI] [Google Scholar]
- 27.Harris CR, Millman KJ, van der Walt SJ, et al. Array programming with NumPy. Nature. 2020;585(7825):357–362. 10.1038/s41586-020-2649-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Bradski G The OpenCV Library. Dr. Dobb’s Journal of Software Tools. 2000;120:122–125. - References -Accessed November 18, 2023 https://www.scirp.org/(S(351jmbntvnsjt1aadkposzje))/reference/ReferencesPapers.aspx?ReferenceID=1692176. [Google Scholar]
- 29.Viola P, Jones M. Rapid object detection using a boosted cascade of simple features. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 20011. [Google Scholar]
- 30.Andrie Asmara R, Ridwan M, Budiprasetyo G. Haar cascade and convolutional neural network face detection in client-side for cloud computing face recognition. In: 2021 International Conference on Electrical and Information Technology (IEIT). 2021:1–5. 10.1109/IEIT53149.2021.9587388. [DOI] [Google Scholar]
- 31.Feng Y, Yu S, Peng H, Li YR, Zhang J. Detect faces efficiently: a survey and evaluations. IEEE Trans Biom Behav Identity Sci. 2022;4(1):1–18. 10.1109/TBIOM.2021.3120412. [DOI] [Google Scholar]
- 32.Guennouni S, Ahaitouf A, Mansouri A. A comparative study of multiple object detection using Haar-like feature selection and local binary patterns in several platforms. Modelling and Simulation in Engineering. 2015;2015, e948960. 10.1155/2015/948960. [DOI] [Google Scholar]
- 33.Chicco D, Warrens MJ, Jurman G. The matthews correlation coefficient (MCC) is more informative than Cohen’s kappa and brier score in binary classification Assessment. IEEE Access. 2021;9:78368–78381. 10.1109/ACCESS.2021.3084050. [DOI] [Google Scholar]
- 34.Cascade classification — OpenCV 2.4.13.7 documentation. Accessed December 14, 2023. https://docs.opencv.org/2.4/modules/objdetect/doc/cascade_classification.html#cascadeclassifier-detectmultiscale.
- 35.Geisler EL, Agarwal S, Hallac RR, Daescu O, Kane AA. A role for artificial intelligence in the classification of craniofacial anomalies. J Craniofac Surg. 2021;32 (3):967–969. 10.1097/SCS.0000000000007369. [DOI] [PubMed] [Google Scholar]
- 36.Agarwal S, Hallac RR, Mishra R, Li C, Daescu O, Kane A. Image based detection of craniofacial abnormalities using feature extraction by classical convolutional neural network. In: 2018 IEEE 8th International Conference on Computational Advances in Bio and Medical Sciences (ICCABS). 2018:1–6. 10.1109/ICCABS.2018.8541948. [DOI] [Google Scholar]
- 37.Bi S, Chen R, Zhang K, et al. Differentiate cavernous hemangioma from schwannoma with artificial intelligence (AI). Ann Transl Med. 2020;8(11):710. 10.21037/atm.2020.03.150. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Schmidt A, Kasiński A. The performance of the Haar cascade classifiers applied to the face and eyes detection. In: Kurzynski M, Puchala E, Wozniak M, Zolnierek A, eds. Computer Recognition Systems 2. Springer; 2007:816–823. 10.1007/978-3-540-75175-5_101. [DOI] [Google Scholar]
- 39.Moore BE, Corso JJ FiftyOne. GitHub Note: https://github.com/voxel51/fiftyone. Published online 2020. [Google Scholar]
- 40.Lu Z, Rallapalli S, Chan K, La Porta T. Modeling the resource requirements of convolutional neural networks on mobile devices. In: Proceedings of the 25th ACM International Conference on Multimedia. 2017:1663–1671. 10.1145/3123266.3123389. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
