Skip to main content
British Journal of Cancer logoLink to British Journal of Cancer
. 2021 Apr 19;124(12):1934–1940. doi: 10.1038/s41416-021-01386-x

Artificial Intelligence-based methods in head and neck cancer diagnosis: an overview

Hanya Mahmood 1,, Muhammad Shaban 2, Nasir Rajpoot 2, Syed A Khurram 3
PMCID: PMC8184820  PMID: 33875821

Abstract

Background

This paper reviews recent literature employing Artificial Intelligence/Machine Learning (AI/ML) methods for diagnostic evaluation of head and neck cancers (HNC) using automated image analysis.

Methods

Electronic database searches using MEDLINE via OVID, EMBASE and Google Scholar were conducted to retrieve articles using AI/ML for diagnostic evaluation of HNC (2009–2020). No restrictions were placed on the AI/ML method or imaging modality used.

Results

In total, 32 articles were identified. HNC sites included oral cavity (n = 16), nasopharynx (n = 3), oropharynx (n = 3), larynx (n = 2), salivary glands (n = 2), sinonasal (n = 1) and in five studies multiple sites were studied. Imaging modalities included histological (n = 9), radiological (n = 8), hyperspectral (n = 6), endoscopic/clinical (n = 5), infrared thermal (n = 1) and optical (n = 1). Clinicopathologic/genomic data were used in two studies. Traditional ML methods were employed in 22 studies (69%), deep learning (DL) in eight studies (25%) and a combination of these methods in two studies (6%).

Conclusions

There is an increasing volume of studies exploring the role of AI/ML to aid HNC detection using a range of imaging modalities. These methods can achieve high degrees of accuracy that can exceed the abilities of human judgement in making data predictions. Large-scale multi-centric prospective studies are required to aid deployment into clinical practice.

Subject terms: Head and neck cancer, Head and neck cancer

Background

Head and neck cancers: incidence and diagnosis

Head and neck cancers (HNC) comprise a heterogeneous group of cancers, most commonly squamous cell carcinomas (SCC), that typically arise from the epithelial lining of the oral cavity, sinonasal tract, pharynx, larynx and salivary glands.1 Most HNC are already at an advanced stage when diagnosed, which significantly reduces the survival rate, even after curative treatment.2 Major risk factors include tobacco smoking/chewing,3 excessive alcohol consumption,4 areca (betel) nut, paan masala (Gutkha),5 gamma and ultraviolet radiation, overexposure to sunlight, a family history of cancer and increasing age.6 The role of human papillomavirus (HPV)7 and Epstein–Barr virus (EBV) has also been implicated in the development of oropharyngeal and nasopharyngeal SCC.3,8 The global incidence of HNC continues to rise9,10 with more than half a million cases annually11 and ~12,000 new cases in the UK each year, an increase of 20% in the last decade.12 Prognosis remains poor, with a 28–67% chance of survival at five years, depending upon the stage at presentation.12

Public health screening/awareness programmes, withdrawal of environmental carcinogens and early detection of precancerous lesions remain the focus for primary and secondary prevention.3 However, early detection of some HNC can be difficult due to vague histories and indistinctive diagnostic features. Conventional diagnosis of HNC is based on histopathological evaluation of tissue sections from biopsies or surgical resections, in addition to clinical and radiological examinations. These methods can be time-consuming and are prone to errors in observation or variations in interpretation,1315 which can result in inconsistencies in cancer grading and prognostication.16 Consequently, this can cause delays and/or inaccuracies in diagnosis, which can have significant implications on patient management and survival. Indeed, improvements in HNC prediction accuracy and disease outcomes could greatly assist healthcare professionals in the early detection and planning of patient-specific optimal treatments to reduce the disease burden.

Artificial intelligence: machine learning and medical image analysis

Recent technological advancements in Artificial Intelligence (AI) algorithms, computer hardware and big medical imaging datasets have enabled computer scientists and healthcare researchers to collaborate closely to improve consistency in cancer risk stratification over the use of multi-factor analysis, conventional logistic regression and Cox analyses.17

The recent advent of Machine Learning (ML) has seen a surge of interest with the exponential growth of evidence to support its wide applications in a range of cancers.1824 ML is a branch of AI that uses computational methods to detect patterns, gather insight and make predictions about new data by using historical information that has been ‘learnt’. As the volume of training data increases, ML algorithms can produce more accurate and efficient predictions.25 Deep learning (DL) is a subfield of ML in which algorithms are structured to create artificial neural networks with multiple hidden layers. These methods have also gained significant popularity in recent years due to their achieving relatively high accuracy of prediction.

Much of the recent focus in cancer diagnostics has centred on digital image analysis and processing, which involves extraction of meaningful information from images to enable delineation of features of clinical interest (segmentation) or description of labels (classification).26,27 A number of ad hoc (or hand-crafted) feature analysis-based ML approaches have been shown to be successful in different diagnostic applications, by explicitly defining a prior set of features and processing steps28 (Fig. 1). Detection of HNC can be achieved using these ML methods by obtaining clinically important information from primary diagnostic imaging modalities, in which high-dimensional, mineable images can be input to train algorithms. For example, radiomic data can be derived from radiographs, ultrasound (US), computed tomography (CT), magnetic resonance (MR), positron emission tomography (PET) and nuclear medicine imaging methods, such as single-photon emission computed tomography (SPECT). Similarly, histological, cytological and immunohistochemical data can be obtained from high-resolution whole-slide images (WSI) of stained tissue sections from biopsies or surgical resections. Other emerging tools for non-invasive detection of HNC include multispectral narrow-band imaging, Raman spectroscopy, confocal laser endomicroscopy (CLE) and infrared thermal imaging.

Fig. 1. Feature extraction from primary diagnostic imaging modalities to train ML/DL algorithms to aid outcome prediction.

Fig. 1

Source: The schematic diagrams were prepared in line with the journal’s artwork guidelines; the clinical/histological/radiological images were obtained with appropriate consent from Sheffield Teaching Hospitals NHS Foundation Trust.

This paper seeks to provide an overview of the recently published literature relating to the application of AI/ML methods to aid diagnostic evaluation of HNC. An outline of the different anatomical sites for HNC lesions, the type of diagnostic imaging modality and the AI/ML method used will be presented.

Methods

Literature search

Electronic database searches using MEDLINE via OVID, EMBASE and Google Scholar were conducted to retrieve articles published in the English language over the last eleven years (2009–2020). This period was chosen due to the evolving application of AI/ML methods in diagnostic cancer research over the last decade.

The search strategy was developed in collaboration with a medical information specialist (Health Sciences Library, University of Sheffield, UK) to ensure keywords were appropriately chosen for optimal identification of articles. A combination of tailored search strings containing database-specific medical subject headings (MeSH) and controlled vocabulary was used (see Supplementary Information), and grey literature screened. Whilst not intended as a formal systematic review, the recommendations of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement and checklist were followed where possible.

Study selection

The selection criterion was jointly developed by the authorship team. The principal inclusion criteria were studies applying AI/ML methods to aid diagnosis of HNC using image analysis, with no restrictions placed on the types of methods or imaging modalities used. Due to the anticipated small number of studies in the field, a broad range of studies such as those using AI/ML to identify or differentiate between benign/pre-malignant/malignant pathology, classify disease subtype, segment cancer regions or predict disease outcome.

Studies using AI/ML to predict cancer susceptibility, metastasis, recurrence, survival or treatment efficacy were not included in this review. Studies focussing solely on the evaluation of oesophageal or thyroid cancers were excluded, unless they were included as part of a larger study that included other HNC lesions. Narrative reviews, letters to editors, commentaries, conference abstracts and animal studies were also excluded. All articles were independently screened by two authors (H.M. and M.S.). The first screen involved the assessment of study title and abstracts and the removal of duplicate articles. The second screen involved comprehensive full-text examination against the predefined criterion. In the case of author discrepancy, two further authors (NMR, SAK) were consulted to make a final decision on article inclusion.

Data capture and synthesis

Relevant data were extracted, tabulated and processed in Microsoft Excel® (Microsoft Corporation, Washington, USA). Data collection included

  • Study details (date of publication, authors, study location and aims)

  • Study methods (anatomical sites for HNC lesions, diagnostic imaging modality, dataset sizes and application of AI/ML methods to aid cancer diagnosis/outcome)

  • AI/ML algorithm performance (reported accuracy measures)

A narrative synthesis with the relevant graphical display is presented. Due to the variations in study outcomes and heterogeneity of data, a meta-analysis for the calculation of adjusted pool estimates was not performed.

Results

The electronic database search retrieved 771 scientific articles. After the removal of duplicates and screening of study titles and abstracts, 698 articles were excluded. Detailed full-text examination of remaining articles excluded a further 41 studies, resulting in 32 articles for inclusion (see Supplementary Table). Among the selected articles, 9 were published between 2009 and 2014, and the remaining 23 articles published between 2015 and 2020. The primary outcomes of interest were anatomical sites of the HNC lesions, diagnostic imaging modalities used for algorithm training/optimisation and the type of AI/ML method.

Anatomical sites for HNC lesions

Figure 2 illustrates the different anatomical sites for the HNC lesions in the selected studies, with the largest proportion (16 studies) involving the oral cavity. Amongst these, nine studies focussed on the assessment of oral squamous cell carcinoma (OSCC) and seven studies focussed on the evaluation of, or differentiation between, oral potentially malignant disorders (OPMD) and OSCC. The remaining studies focussed on assessment of nasopharyngeal SCC (n = 3),2931 laryngeal SCC (n = 2),32,33 oropharyngeal SCC (n = 3),3436 parotid gland neoplasms (n = 2)37,38 and differentiation between sinonasal SCC from inverted papilloma (n = 1).39 In four studies,4042 tissue sections of HNC from various different sites (tongue, floor of mouth, soft palate, mandible, gingivae, alveolar ridge, supraglottis, maxillary sinus, nose, thyroid and parotid gland) were evaluated. One of these studies did not specify the anatomical sites of the HNC lesions.43

Fig. 2. HNC anatomical site distribution.

Fig. 2

A bar chart showing the proportion of included studies based on head and neck cancer anatomical subtype.

Diagnostic imaging modalities for algorithm training/optimisation

Histology WSI of haematoxylin and eosin (H&E)-stained tissue sections was used to train AI/ML algorithms in nine studies (Fig. 3). Radiology image data were used in eight studies and obtained from dynamic contrast-enhanced MRI (DCE-MRI) (n = 3),31,39,43 CT (n = 2),36,37 PET/CT (n = 1),29 US (n = 1)38 and plain film intraoral radiographs (n = 1).44 Other imaging modalities included hyperspectral imaging (HSI) (n = 6), endoscopic/clinical imaging (n = 5), infrared thermal imaging (n = 1)45 and multimodal optical imaging (n = 1).46 In the remaining two studies, clinicopathologic, genomic and exfoliative cytological data were used to predict outcomes based on traditional statistical analysis methods.47,48

Fig. 3. Type of imaging modality/input data used.

Fig. 3

A pie chart showing the proportion of identified studies based on the type of maging modality/input data used to train AI/ML algorithms.

Histology whole-slide imaging

In nine studies, histology WSI was used to develop algorithms for evaluation of OSCC (n = 2),49,50 OPMD (n = 4),5154 laryngeal SCC (n = 1),32 oropharyngeal SCC (n = 1)35 and multiple HNC sites (n = 1).55

These studies used a variety of different ML approaches to delineate specific histological features of interest with downstream statistical analysis to compare differences in spatial architectural patterns for differentiation between benign and malignant lesions. ML tools were developed to assess differences in detection of quantity, geometry, compactness and eccentricity of sub-epithelial connective tissue cells and to identify textural differences between normal and oral submucous fibrosis tissue (with and without dysplasia or atrophy) using approaches including Brownian motion.5153

In one study, unsupervised ML methods were used for the automated identification of tissue compartments in oropharyngeal SCC (OPSCC) tissue microarrays.35 Morphological measurements of cell and nuclei were used for the classification of epithelial and stromal tissue achieving a pixel-level F1 score of 80–81%. A further study showed that stimulated Raman scattering histology integrated with DL algorithms provided the good potential for delivering a rapid intraoperative diagnosis of laryngeal SCC with an accuracy of 90%.32 Findings showed that this method could identify tissue neoplasia at the simulated resection margins that appear grossly normal with the naked eye, highlighting the potential to enhance surgical resection and reduce disease recurrence.

A recent systematic review highlights emerging evidence to support the role of ML methods for histology images as a potentially useful diagnostic aid for the detection of OSCC and some OPMD, but identifies a lack of evidence for other head and neck precancerous or cancerous lesions.44 However, the overall quality of evidence in these studies is low, mainly due to the use of small unicentric datasets and a high risk of bias that could have overestimated model accuracy rates.

Radiological imaging

Three studies used radiomic-based feature prediction from MRI to aid assessment of various HNC lesions,43 including nasopharyngeal31 and sinonasal SCC.39 Deng et al.43 proposed an automatic segmentation method using traditional ML techniques for evaluation of HNC lesions demonstrating superior segmentation performance (area overlap measure of 0.76 + /−0.08 and accuracy of 86 ± 8%), which outperformed previous similar studies. Huang et al.31 evaluated the performance of two region-based segmentation methods for the evaluation of nasopharyngeal SCC. Ramkumar et al.39 found that MRI-based textural analysis had the potential to differentiate sinonasal SCC from inverted papilloma (accuracy 89.1%) with results comparable to manual assessment by neuroradiologists (P = 0.0004).

Three studies used CT-based textural analysis for the evaluation of HNC. Ajmi et al.37 developed an approach using spectral dual-energy CT (DECT) data from multi-energy virtual monochromatic image datasets to capture the energy-dependent changes in tissue attenuation for the classification of common benign parotid gland neoplasms (Warthin tumour and pleomorphic adenoma) with an accuracy of 92%. Whereas Ranjbar et al.36 used CT-based texture analysis to classify the HPV status of oropharyngeal SCC (accuracy 75.7%). In another study, Wu et al.29 developed an automated algorithm for the identification of nasopharyngeal SCC on PET/CT examination with 100% accuracy for detection of hypermetabolic lesions larger than 1 cm in size.

Only one study used textural features derived from an ultrasound (US) using radio-frequency echo signals and image data to enable automatic differentiation between malignant and benign parotid gland lesions (accuracy 91%) based on a supervised classification system.38 In another study, gravitational search-optimised echo-state neural networks were developed for early prediction of OSCC from intraoral X-rays with a detection accuracy of 99.2%.56

Endoscopic/clinical imaging

Four studies used clinical data from endoscopic imaging for the detection of oral,57 nasopharyngeal,30 oropharyngeal34 and laryngeal cancers.33 Amongst these, two studies employed DL methods. The first study developed algorithms for early detection of nasopharyngeal malignancies (accuracy of 88.7%)30 providing surgeons with useful biopsy guidance. The second study demonstrated the superior performance of DL compared to textural feature-based ML methods (accuracy 88.3%, sensitivity 86.6% and specificity 90%) in recognition of sub-surface microanatomical in vivo cell structures using confocal laser endomicroscopy (CLE) in OSCC patients.56

In two studies, traditional ML approaches were used for the detection of oropharyngeal and laryngeal SCC. Mascharak et al.34 used multispectral narrow-band imaging and white-light endoscopy (WLE) to quantify the lymphoepithelial tissues of the oropharynx. The results showed a promising ability to differentiate between oropharyngeal SCC and healthy mucosa based on differences in colour (accuracy 65.9% compared to 52.3% under WLE, P = 0.0108), presumably a reflection of surface angiogenic and local inflammatory changes. Whereas Moccia et al.33 used traditional ML techniques to classify laryngeal tissues (normal vs malignant) in narrow-band endoscopic images by exploiting textural information (classification recall 93–98%). These studies demonstrate a promising step towards an endoscope-based processing system to support the early-stage diagnosis of HNC lesions that may go unnoticed by the human eye.

Song et al.58 developed a smartphone-based intraoral dual-modality imaging platform to classify OPMD and malignant lesions based on autofluorescence and white-light images using a convolutional neural network (CNN). The results demonstrated an accuracy of ~86.9%, although the training sample was relatively small (66 normal samples and 64 suspicious). Other limitations, including training the CNN algorithms on tissue from different anatomical regions (for normal, dysplastic and malignant tissues), are likely to exhibit different autofluorescence characteristics due to the varying tissue structural and biochemical compositions.

Other imaging modalities

Six studies used HSI for AI/ML training. Three of those studies focussed on early detection and diagnosis of OSCC. Liu et al.59 used HSI to measure the reflectance spectra in the tongue to enable differentiation between normal and cancerous tissue, with a recognition rate of 96.5%. Roblyer et al.46 used multispectral wide-field optical imaging—which included white-light reflectance, autofluorescence, narrow-band reflectance and cross-polarised imaging modalities—to distinguish between oral cancer/precancer and non-neoplastic mucosa by evaluating image contrast. Their results showed that autofluorescence imaging at 405-nm excitation wavelength provided the greatest image contrast, and the ratio of red-to-green fluorescence intensity computed from these images provided the best classification of dysplasia/cancer versus non-neoplastic tissue (sensitivity of 100%, specificity of 85%). Although this approach accurately distinguished malignant from benign tissue, the ability to separate precancerous lesions from cancers was found to be limited. In the third study, Quang et al.60 also used multimodal optical imaging, in which autofluorescence imaging was used to identify high-risk regions within the oral cavity, followed by high-resolution microendoscopy to confirm or exclude the presence of neoplasia (defined by the authors as diagnoses of moderate dysplasia or worse). Data from 92 sites (n = 30) were used to develop algorithms for the automatic identification of OSCC in vivo. Diagnostic performance was evaluated prospectively using images from 114 sites (n = 70) and the confirmed histological diagnosis based on either a biopsy or an excised surgical specimen. Amongst the sites that were biopsied (n = 56), the classification accuracy for detection of benign and cancerous lesions was 100 and 85%, respectively. Amongst the sites that corresponded to a surgical specimen (n = 58), multimodal imaging correctly classified 100% of benign and 61% of neoplastic sites.

Jeyaraj et al.45 developed a convolutional neural network (CNN) classifier for OSCC detection using multidimensional HSI (accuracy of 91.4%). Similarly, Halicek et al.41 also developed a CNN classifier that was trained using HSI of HNSCC surgical specimens, including thyroid cancer, and normal head and neck tissue samples. This model showed the potential to produce near-real-time automatic tissue labelling for intraoperative cancer detection using HSI.

One study explored the viability of digital infrared thermal imaging (DITI) for screening and detection of OSCC.61 DITI is a non-invasive, non-ionising, radiation hazard-free modality that essentially produces a heat map of an object by capturing its radiated thermal energy. The authors developed a semi-automated screening framework for OSCC by extracting significantly discriminating textural features from facial thermograms for classification into normal, precancerous or malignant categories achieving an accuracy rate of 85.42%.

Another study developed diagnostic algorithms for HNSCC detection using ML constructed by mass spectra obtained from non-cancerous (n = 15, 114 mass spectra) and HNSCC (n = 19, 141 mass spectra) specimens by probe electrospray ionisation mass spectrometry. The clinical validity of this approach was evaluated to discriminate tumour-specific spectral patterns using intraoperative specimens of HNSCC and normal mucosa with positive and negative-ion modes showing accuracies in HNSCC diagnosis of 90.48% and 95.35%, respectively.

In another study,48 exfoliative cytology, histopathology and clinical data of normal subjects (n = 102), oral leukoplakia (OLK) patients (n = 82) and OSCC patients (n = 93) were collected for quantitative risk stratification of OLK lesions. This involved expert-guided data transformation and reconstruction for automatic data processing to reveal a risk index for OSCC prediction (sensitivity: median >0.98, specificity: median >0.99).

Type of AI/ML method

Figure 4 illustrates the proportion of AI methods used in the selected studies. Traditional ML methods were employed in 22 studies (69%) with common approaches, including Support Vector Machine, Logistic Regression, Random Forest, Decision Tree, K-Nearest Neighbour, Bayesian Classifier and Linear Discriminant Analysis. DL-based neural networks were employed in eight studies (25%), and a combination of traditional ML and DL methods were used in two studies (6%).

Fig. 4. Types of AI methods used.

Fig. 4

Proportion of studies using traditional ML methods, modern deep learning methods and a combination of both types of methods.

Figure 5 provides a breakdown of the selected studies by combining the anatomical site of lesion, imaging modality and AI methods used in the recently published literature. Traditional ML methods were most frequently used for the detection of precancerous or cancerous lesions of the oral cavity (11 studies), and specifically in studies using histology WSI (four studies). DL methods were used for detection of HNSCC lesions of the oral cavity (five studies), nasopharynx (one study), larynx (one study) and various other head and neck sites as specified in a study by Halicek et al.41

Fig. 5. Overview of recently published literature.

Fig. 5

A diagram illustrating an overview of the recently published literature based on anatomical site of lesion, imaging modality and AI methods used.

Discussion

This paper provides an insight into the recent application of AI/ML for the evaluation of HNC lesions using digital image analysis. It has shown, primarily, a wide breadth of imaging modalities that have been used to retrieve input data for algorithm training in the last decade. Whilst a detailed statistical analysis of the heterogeneous dataset samples has not been undertaken, most studies have indicated that ML can achieve high degrees of accuracy and precision that can exceed the abilities of standard statistical techniques and human judgement in making predictions about data. This supports seminal claims made by Meehl in 1954,62 and more recent meta-analyses,63,64 which propose that correctly used mechanical (i.e., algorithmic) methods make more efficient and reliable decisions and predictions about patient outcomes and treatment. However, despite findings highlighted by our paper, very few ML tools have actually been deployed into clinical practice.

Whilst a formal risk of bias analysis has not been conducted for the cited studies, the reported accuracy rates should be interpreted with caution. This is because most studies have used small unicentric datasets that may be biased towards a particular patient demographics. Multi-centric research will inevitably allow a more diverse dataset with the inclusion of patients from different geographical locations, populations and demographics that will enhance algorithm performance by incorporating biological and technical variance.

Findings demonstrate the greatest proportion of studies to have evaluated the detection of OPMD and cancerous lesions within the oral cavity (Fig. 2) with histology WSI and radiological imaging being the most frequently used modalities for algorithm training (Fig. 3). This is consistent with the increasing ubiquity of digital slide scanners in pathology laboratories and the emergence of radiomics that has broadened the scope of routine medical imaging in clinical oncology.

With the continued evolution of AI algorithms and computational power, a plethora of computational methods has been developed for fast and reproducible diagnosis and prognosis of HNC, as exemplified in this paper. The emergence of various high-resolution imaging modalities (i.e., multimodal optical, microendoscopic, hyperspectral and infrared thermal) has provided an unprecedented opportunity for quantitative feature extraction by conversion into mineable images at relatively low cost and non-invasively. Having said this, histology WSI remains the most superior imaging modality for data leverage. This is because each image provides multi-gigapixel-level information, thereby resulting in hundreds of thousands of sub-images (image patches) per WSI for analysis and algorithm training.

Early work has been largely based on the development and application of traditional ML methods (Fig. 4); however, in recent years, the use of DL for HNSCC diagnosis and prognostication has evolved. This opens the opportunity to develop state-of-the-art DL techniques that can be combined with traditional approaches to improve detection accuracy of head and neck precancerous and cancerous lesions, as well as predict the course of a precancerous or cancerous lesion learning from retrospective data. Another exciting research avenue would be the development of new data fusion algorithms that combine imaging modalities such as radiologic, histologic and molecular measurements to aid disease detection, classification and outcome prediction.

Supplementary information

Supplementary information (12.7KB, docx)
Supplementary Table 1 (70.7KB, docx)

Acknowledgements

We thank the Health Sciences Library at The University of Sheffield for assistance with the electronic database search.

Author contributions

H.M. and S.A.K. conceived and designed the study. H.M. conducted the literature search with support from a medical information specialist (University of Sheffield). H.M. and M.S. screened all articles and collected data that led to the submission. S.A.K. and N.M.R. confirmed the final selection of articles. H.M., N.M.R., S.A.K. and M.S. were involved in analysing results, preparation, editing and approval of the final paper.

Ethics approval and consent to participate

Not applicable.

Consent to publish

Not applicable.

Data availability

All data generated or analysed during this study are included in this published article and its supplementary information files.

Competing interests

The authors declare no competing interests.

Funding information

H.M. is funded by the National Institute for Health Research and received a small grant from Sheffield Hospitals Charity to support her work. N.M.R. was supported by the PathLAKE digital pathology consortium, which is funded from the Data to Early Diagnosis and Precision Medicine strand of the UK government’s Industrial Strategy Challenge Fund (award# 18181), managed and delivered by UK Research and Innovation (UKRI). N.M.R. was also supported by the UK Medical Research Council (award# MR/P015476/1) and the Alan Turing Institute.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

The online version contains supplementary material available at 10.1038/s41416-021-01386-x.

References

  • 1.Pai SI, Westra WH. Molecular pathology of head and neck cancer: implications for diagnosis, prognosis, and treatment. Annu. Rev. Pathol.: Mechanisms Dis. 2009;4:49–70. doi: 10.1146/annurev.pathol.4.110807.092158. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Muto M, Nakane M, Katada C, Sano Y, Ohtsu A, Esumi H, et al. Squamous cell carcinoma in situ at oropharyngeal and hypopharyngeal mucosal sites. Cancer: Interdisciplinary Int. J. Am. Cancer Soc. 2004;101:1375–1381. doi: 10.1002/cncr.20482. [DOI] [PubMed] [Google Scholar]
  • 3.Shaw R, Beasley N. Aetiology and risk factors for head and neck cancer: United Kingdom National Multidisciplinary Guidelines. J. Laryngol. Otol. 2016;130:S9–S12. doi: 10.1017/S0022215116000360. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.IARC Working Group on the Evaluation of Carcinogenic Risks to Humans. Alcohol consumption and ethyl carbamate. IARC Monogr. Evaluation Carcinogenic Risks Hum. 2010;96:3. [PMC free article] [PubMed] [Google Scholar]
  • 5.World Health Organization, IARC Working Group on the Evaluation of Carcinogenic Risks to Humans, International Agency for Research on Cancer. Betel-quid and Areca-nut Chewing and Some Areca-nut-derived Nitrosamines (IARC, 2004). [PMC free article] [PubMed]
  • 6.State of Mouth Cancer UK Report 2020/21. Oral Health Foundation. Available via URL: https://www.dentalhealth.org/thestateofmouthcancer (2020).
  • 7.Papillomaviruses H. IARC Monographs on the Evaluation of Carcinogenic Risks to Humans. IARC; 2011. [Google Scholar]
  • 8.Shotelersuk K, Khorprasert C, Sakdikul S, Pornthanakasem W, Voravud N, Mutirangura A. Epstein-Barr virus DNA in serum/plasma as a tumor marker for nasopharyngeal cancer. Clin. Cancer Res. 2000;6:1046–1051. [PubMed] [Google Scholar]
  • 9.Jemal A, Bray F, Center MM, Ferlay J, Ward E, Forman D. Global cancer statistics. CA: Cancer J. Clin. 2011;61:69–90. doi: 10.3322/caac.20107. [DOI] [PubMed] [Google Scholar]
  • 10.Siegel RL, Miller KD, Jemal A. Cancer statistics, 2020. CA Cancer J. Clin. 2020;70:7. doi: 10.3322/caac.21590. [DOI] [PubMed] [Google Scholar]
  • 11.Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: Cancer J. Clin. 2018;68:394–424. doi: 10.3322/caac.21492. [DOI] [PubMed] [Google Scholar]
  • 12.Cancer Research UK. Head and neck cancers incidence statistics 2015–17. Available at URL: https://www.cancerresearchuk.org/healthprofessional/cancer-statistics/statistics-by-cancer-type/head-and-neck-cancers/incidence (2017).
  • 13.Kujan O, Khattab A, Oliver RJ, Roberts SA, Thakker N, Sloan P. Why oral histopathology suffers inter-observer variability on grading oral epithelial dysplasia: an attempt to understand the sources of variation. Oral. Oncol. 2007;43:224–231. doi: 10.1016/j.oraloncology.2006.03.009. [DOI] [PubMed] [Google Scholar]
  • 14.Vogel DW, Thoeny HC. Cross-sectional imaging in cancers of the head and neck: how we review and report. Cancer Imaging. 2016;16:20. doi: 10.1186/s40644-016-0075-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Schlumpf, M. F. & Haerle, S. The current role of imaging in head and neck cancer: a clinician’s perspective. Swiss Medical Weekly144, w14015 (2014). [DOI] [PubMed]
  • 16.Mehlum CS, Larsen SR, Kiss K, Groentved AM, Kjaergaard T, Möller S, et al. Laryngeal precursor lesions: Interrater and intrarater reliability of histopathological assessment. Laryngoscope. 2018;128:2375–2379. doi: 10.1002/lary.27228. [DOI] [PubMed] [Google Scholar]
  • 17.Huang S, Yang J, Fong S, Zhao Q. Artificial intelligence in cancer diagnosis and prognosis: opportunities and challenges. Cancer Lett. 2020;471:61–71. doi: 10.1016/j.canlet.2019.12.007. [DOI] [PubMed] [Google Scholar]
  • 18.Bejnordi BE, Veta M, Van Diest PJ, Van Ginneken B, Karssemeijer N, Litjens G, et al. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. J. Am. Med. Assoc. 2017;318:2199–2210. doi: 10.1001/jama.2017.14585. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Comput. Struct. Biotechnol. J. 2015;13:8–17. doi: 10.1016/j.csbj.2014.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Graham, S. & Rajpoot, N. SAMS-NET: stain-aware multi-scale network for instance-based nuclei segmentation in histology images. In 2018 Institute of Electrical and Electronics Engineers (IEEE) 15th International Symposium on Biomedical Imaging (ISBI 2018), pp. 590–594 (IEEE, 2018).
  • 21.Bera K, Schalper KA, Rimm DL, Velcheti V, Madabhushi A. Artificial intelligence in digital pathology—new tools for diagnosis and precision oncology. Nat. Rev. Clin. Oncol. 2019;16:703–715. doi: 10.1038/s41571-019-0252-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Zormpas-Petridis K, Failmezger H, Raza SE, Roxanis I, Jamin Y, Yuan Y. Superpixel-based conditional random fields (SuperCRF): incorporating global and local context for enhanced deep learning in melanoma histopathology. Front. Oncol. 2019;9:1045. doi: 10.3389/fonc.2019.01045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Wang S, Yang DM, Rong R, Zhan X, Fujimoto J, Liu H, et al. Artificial intelligence in lung cancer pathology image analysis. Cancers. 2019;11:1673. doi: 10.3390/cancers11111673. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Sirinukunwattana K, Raza SE, Tsang YW, Snead DR, Cree IA, Rajpoot NM. Locality sensitive deep learning for detection and classification of nuclei in routine colon cancer histology images. IEEE Trans. Med. Imaging. 2016;35:1196–1206. doi: 10.1109/TMI.2016.2525803. [DOI] [PubMed] [Google Scholar]
  • 25.LeCun Y, Bengio Y, Hinton G. Deep learning. nature. 2015;521:436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
  • 26.Litjens G, Kooi T, Bejnordi BE, Setio AA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med. image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
  • 27.Shen D, Wu G, Suk HI. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017;19:221–248. doi: 10.1146/annurev-bioeng-071516-044442. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Panayides, A. S., Amini, A., Filipovic, N. D., Sharma, A., Tsaftaris, S. A., Young, A. et al. AI and medical imaging informatics: current challenges and future directions. IEEE J. Biomed. Health Inform.24, 1837–1857 (2020). [DOI] [PMC free article] [PubMed]
  • 29.Wu B, Khong PL, Chan T. Automatic detection and classification of nasopharyngeal carcinoma on PET/CT with support vector machine. Int. J. Computer Assist. Radiol. Surg. 2012;7:635–646. doi: 10.1007/s11548-011-0669-y. [DOI] [PubMed] [Google Scholar]
  • 30.Li C, Jing B, Ke L, Li B, Xia W, He C, et al. Development and validation of an endoscopic images‐based deep learning model for detection with nasopharyngeal malignancies. Cancer Commun. 2018;38:1–1. doi: 10.1186/s40880-018-0325-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Huang W, Chan KL, Zhou J. Region-based nasopharyngeal carcinoma lesion segmentation from MRI using clustering-and classification-based methods with learning. J. Digital Imaging. 2013;26:472–482. doi: 10.1007/s10278-012-9520-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Zhang L, Wu Y, Zheng B, Su L, Chen Y, Ma S, et al. Rapid histology of laryngeal squamous cell carcinoma with deep-learning based stimulated Raman scattering microscopy. Theranostics. 2019;9:2541. doi: 10.7150/thno.32655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Moccia S, De Momi E, Guarnaschelli M, Savazzi M, Laborai A, Guastini L, et al. Confident texture-based laryngeal tissue classification for early stage diagnosis support. J. Med. Imaging. 2017;4:034502. doi: 10.1117/1.JMI.4.3.034502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Mascharak S, Baird BJ, Holsinger FC. Detecting oropharyngeal carcinoma using multispectral, narrow‐band imaging and machine learning. Laryngoscope. 2018;128:2514–2520. doi: 10.1002/lary.27159. [DOI] [PubMed] [Google Scholar]
  • 35.Fouad, S., Randell, D., Galton, A., Mehanna, H. & Landini, G. Unsupervised morphological segmentation of tissue compartments in histopathological images. PLoS ONE12, e0188717 (2017). [DOI] [PMC free article] [PubMed]
  • 36.Ranjbar S, Ning S, Zwart CM, Wood CP, Weindling SM, Wu T, et al. Computed tomography-based texture analysis to determine human papillomavirus status of oropharyngeal squamous cell carcinoma. J. computer Assist. Tomogr. 2018;42:299–305. doi: 10.1097/RCT.0000000000000682. [DOI] [PubMed] [Google Scholar]
  • 37.Al Ajmi E, Forghani B, Reinhold C, Bayat M, Forghani R. Spectral multi-energy CT texture analysis with machine learning for tissue classification: an investigation using classification of benign parotid tumours as a testing paradigm. Eur. Radiol. 2018;28:2604–2611. doi: 10.1007/s00330-017-5214-0. [DOI] [PubMed] [Google Scholar]
  • 38.Siebers S, Zenk J, Bozzato A, Klintworth N, Iro H, Ermert H. Computer aided diagnosis of parotid gland lesions using ultrasonic multi-feature tissue characterization. Ultrasound Med. Biol. 2010;36:1525–1534. doi: 10.1016/j.ultrasmedbio.2010.06.009. [DOI] [PubMed] [Google Scholar]
  • 39.Ramkumar S, Ranjbar S, Ning S, Lal D, Zwart CM, Wood CP, et al. MRI-based texture analysis to differentiate sinonasal squamous cell carcinoma from inverted papilloma. Am. J. Neuroradiol. 2017;38:1019–1025. doi: 10.3174/ajnr.A5106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Fei B, Lu G, Wang X, Zhang H, Little JV, Patel MR, et al. Label-free reflectance hyperspectral imaging for tumor margin assessment: a pilot study on surgical specimens of cancer patients. J. Biomed. Opt. 2017;22:086009. doi: 10.1117/1.JBO.22.8.086009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Halicek M, Lu G, Little JV, Wang X, Patel M, Griffith CC, et al. Deep convolutional neural networks for classifying head and neck cancer using hyperspectral imaging. J. Biomed. Opt. 2017;22:060503. doi: 10.1117/1.JBO.22.6.060503. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Lu G, Little JV, Wang X, Zhang H, Patel MR, Griffith CC, et al. Detection of head and neck cancer in surgical specimens using quantitative hyperspectral imaging. Clin. Cancer Res. 2017;23:5426–5436. doi: 10.1158/1078-0432.CCR-17-0906. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Deng, W., Luo, L., Lin, X., Fang, T., Liu, D., Dan, G. et al. Head and neck cancer tumor segmentation using support vector machine in dynamic contrast-enhanced MRI. Contrast Media & Molecular Imaging2017, 8612519 (2017). [DOI] [PMC free article] [PubMed]
  • 44.Mahmood H, Shaban M, Indave BI, Santos-Silva AR, Rajpoot N, Khurram SA. Use of artificial intelligence in diagnosis of head and neck precancerous and cancerous lesions: a systematic review. Oral. Oncol. 2020;110:104885. doi: 10.1016/j.oraloncology.2020.104885. [DOI] [PubMed] [Google Scholar]
  • 45.Jeyaraj PR, Nadar ER. Computer-assisted medical image classification for early diagnosis of oral cancer employing deep learning algorithm. J. Cancer Res. Clin. Oncol. 2019;145:829–837. doi: 10.1007/s00432-018-02834-7. [DOI] [PubMed] [Google Scholar]
  • 46.Roblyer DM, Kurachi C, Stepanek V, Schwarz RA, Williams MD, El-Naggar AK, et al. Comparison of multispectral wide-field optical imaging modalities to maximize image contrast for objective discrimination of oral neoplasia. J. Biomed. Opt. 2010;15:066017. doi: 10.1117/1.3516593. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Ashizawa K, Yoshimura K, Johno H, Inoue T, Katoh R, Funayama S, et al. Construction of mass spectra database and diagnosis algorithm for head and neck squamous cell carcinoma. Oral. Oncol. 2017;75:111–119. doi: 10.1016/j.oraloncology.2017.11.008. [DOI] [PubMed] [Google Scholar]
  • 48.Liu Y, Li J, Liu X, Liu X, Khawar W, Zhang X, et al. Quantitative risk stratification of oral leukoplakia with exfoliative cytology. PLoS ONE. 2015;10:e0126760. doi: 10.1371/journal.pone.0126760. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Das DK, Bose S, Maiti AK, Mitra B, Mukherjee G, Dutta PK. Automatic identification of clinically relevant regions from oral tissue histological images for oral squamous cell carcinoma diagnosis. Tissue Cell. 2018;53:111–119. doi: 10.1016/j.tice.2018.06.004. [DOI] [PubMed] [Google Scholar]
  • 50.Lu C, Lewis JS, Dupont WD, Plummer WD, Janowczyk A, Madabhushi A. An oral cavity squamous cell carcinoma quantitative histomorphometric-based image classifier of nuclear morphology can risk stratify patients for disease-specific survival. Mod. Pathol. 2017;30:1655–1665. doi: 10.1038/modpathol.2017.98. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Baik J, Ye Q, Zhang L, Poh C, Rosin M, MacAulay C, et al. Automated classification of oral premalignant lesions using image cytometry and random forests-based algorithms. Cell. Oncol. 2014;37:193–202. doi: 10.1007/s13402-014-0172-x. [DOI] [PubMed] [Google Scholar]
  • 52.Krishnan MM, Pal M, Bomminayuni SK, Chakraborty C, Paul RR, Chatterjee J, et al. Automated classification of cells in sub-epithelial connective tissue of oral sub-mucous fibrosis—an SVM based approach. Computers Biol. Med. 2009;39:1096–1104. doi: 10.1016/j.compbiomed.2009.09.004. [DOI] [PubMed] [Google Scholar]
  • 53.Mookiah MR, Shah P, Chakraborty C, Ray AK. Brownian motion curve-based textural classification and its application in cancer diagnosis. Anal. Quant. Cytol. Histol. 2011;33:158–168. [PubMed] [Google Scholar]
  • 54.Krishnan MM, Venkatraghavan V, Acharya UR, Pal M, Paul RR, Min LC, et al. Automated oral cancer identification using histopathological images: a hybrid feature extraction paradigm. Micron. 2012;43:352–364. doi: 10.1016/j.micron.2011.09.016. [DOI] [PubMed] [Google Scholar]
  • 55.Halicek M, Shahedi M, Little JV, Chen AY, Myers LL, Sumer BD, et al. Head and neck cancer detection in digitized whole-slide histology using convolutional neural networks. Sci. Rep. 2019;9:1–1. doi: 10.1038/s41598-019-50313-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Al-Ma’aitah M, AlZubi AA. Enhanced computational model for gravitational search optimized echo state neural networks based oral cancer detection. J. Med. Syst. 2018;42:205. doi: 10.1007/s10916-018-1052-0. [DOI] [PubMed] [Google Scholar]
  • 57.Aubreville, M., Knipfer, C., Oetter, N., Jaremenko, C., Rodner, E., Denzler, J. et al. Automatic classification of cancerous tissue in laserendomicroscopy images of the oral cavity using deep learning. Sci. Rep.7, 1–10 (2017). [DOI] [PMC free article] [PubMed]
  • 58.Song B, Sunny S, Uthoff RD, Patrick S, Suresh A, Kolur T, et al. Automatic classification of dual-modalilty, smartphone-based oral dysplasia and malignancy images using deep learning. Biomed. Opt. Express. 2018;9:5318–5329. doi: 10.1364/BOE.9.005318. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Liu Z, Wang H, Li Q. Tongue tumor detection in medical hyperspectral images. Sensors. 2012;12:162–174. doi: 10.3390/s120100162. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Quang T, Tran EQ, Schwarz RA, Williams MD, Vigneswaran N, Gillenwater AM, et al. Prospective evaluation of multimodal optical imaging with automated image analysis to detect oral neoplasia in vivo. Cancer Prev. Res. 2017;10:563–570. doi: 10.1158/1940-6207.CAPR-17-0054. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Chakraborty, M., Mukhopadhyay, S., Dasgupta, A., Patsa, S., Anjum, N. & Ray, J. G. A new approach of oral cancer detection using bilateral texture features in digital infrared thermal images. In 38th Annual International Conference of the Institute of Electrical and Electronics Engineers (IEEE) Engineering in Medicine and Biology Society (EMBC). (IEEE, 2016). [DOI] [PubMed]
  • 62.Meehl, P. E. Clinical versus statistical prediction: a theoretical analysis and a review of the evidence. (University of Minnesota Press, 1954).
  • 63.Ægisdóttir S, White MJ, Spengler PM, Maugherman AS, Anderson LA, Cook RS, et al. The meta-analysis of clinical judgment project: fifty-six years of accumulated research on clinical versus statistical prediction. Counseling Psychologist. 2006;34:341–382. [Google Scholar]
  • 64.Grove WM, Zald DH, Lebow BS, Snitz BE, Nelson C. Clinical versus mechanical prediction: a meta-analysis. Psychol. Assess. 2000;12:19. [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary information (12.7KB, docx)
Supplementary Table 1 (70.7KB, docx)

Data Availability Statement

All data generated or analysed during this study are included in this published article and its supplementary information files.


Articles from British Journal of Cancer are provided here courtesy of Cancer Research UK

RESOURCES