Abstract
Objective
To examine the construct validity, reliability and responsiveness of the Otology Questionnaire Amsterdam (OQUA).
Design
Multicentre, longitudinal study in 2 separate cohorts of patients visiting an ENT surgeon via an online survey programme.
Setting
Tertiary ENT clinics.
Participants
Cohort 1 consisted of patients at their first visit at an ENT outpatient clinic with an ear complaint. Cohort 2 consisted of patients who underwent surgery, with a 3‐month follow‐up post‐surgery.
Main outcome measures
Construct validity: Hypothesis testing, internal consistency and inter‐item correlation. Reliability: Test‐retest reliability. The construct approach was used for assessing responsiveness. Hypotheses were formulated based on the association between the OQUA and Glasgow Health Status Inventory (GHSI) or Global Rating Scale (GRS).
Results
Construct validity: The correlation between the individual items in the impact domain ranged from 0.424 to 0.737. Confirmatory factor analysis showed a good fit. As expected, the OQUA impact showed strong relationships with GHSI total and general scale.
Reliability: The test‐retest reliability coefficient ranged from 0.541 to 0.838.
Responsiveness: All hypotheses were conformed. As expected, the change score of the OQUA showed good correlation between OQUA impact and GHSI and moderate correlation between the GRS and OQUA complaints.
Conclusion
The OQUA has 8 complaint domains (earache, pressure sensation, itching, tinnitus, hearing loss, ear discharge, loss of taste and dizziness) and 1 impact domain. Each domain results in one score of 0‐100. The OQUA shows good results for construct validity, (test‐retest) reliability and responsiveness, supporting the potential benefit for the patient with an ear complaint visiting the ENT surgeon. The extensive validation furthermore confirms a certified generic otology PROM with an impact and a complaints' part, to be used in different types of otologic interventions and patient groups.
Keywords: audiology, chronic otitis media, mouth, otology, sensorineural hearing loss, taste, tinnitus, vertigo
Keypoints.
A validated generic otologic questionnaire covering all eight relevant ear complaints and their impact.
An effective tool to evaluate the effectiveness of surgical and non‐surgical otological interventions.
Pre‐ and post‐intervention scores can be displayed graphically.
1. INTRODUCTION
Adult patients visiting an ENT surgeon can present with a variety of ear complaints ranging from hearing loss and tinnitus to dizziness. These complaints often occur simultaneously. A patient‐reported outcome measure (PROM) captures the patient's perspective and can be useful to evaluate this mix of complaints. Recently, Viergever et al searched for all existing Otology questionnaires available in the literature through a systematic review. No less than 144 unique questionnaires were identified. Several PROMs for measuring ear complaints were found. However, these questionnaires were either symptom specific (like the Tinnitus Handicap Index or Dizziness Handicap Index) 1 , 2 or disease specific (like the Chronic‐Otitis‐Media Q‐12 or Meniere's disease Patient‐Oriented Severity Index). 3 , 4 No generic PROM is available for measuring the severity and impact of multiple ear complaints without diagnosis in adults visiting an ENT surgeon. Furthermore, since the use of multiple questionnaires seems unpractical in practice, a generic otology PROM that addresses all relevant types of ear complaints and their impact on quality of life is desirable in the outpatient clinic.
Therefore, the Otology Questionnaire Amsterdam (OQUA) was recently developed by Bruinewoud et al 2018, 5 to allow the assessment of the presence and severity of ear complaints and their impact on the quality of life of patients visiting an ENT surgeon. The OQUA was developed according to the COSMIN development guidelines and with the involvement of an expert panel, patients' interviews and extensive field‐testing, including factor analysis. 5
With a validated questionnaire capable to measuring complaints and impact for all different kinds of ear complaints in the daily ENT practice, it could potentially be used in future to assess benefit of surgery from a patient perspective.
In the present study, we further investigated the quality of the OQUA as an instrument to be used in the ENT practice. Our study focused on the following questions: Does the OQUA measure what it is intended to measure (validation)? Is it stable in the scores it provides (reliability)? And does it show changes when a change is expected (responsiveness) before and after an intervention?
Validity was defined as the degree to which an instrument truly measures the construct. 6 If no ‘gold standard’ is available, then construct validity can be tested instead of criterion validity. 7 The construct validity consists of two parts, namely examination of structural validity (factor analysis) and hypothesis testing. The ICC for construct validity was already calculated by Bruinewoud et al 5 and showed good correlation between the individual items. For multi‐item measurement instruments based on a formative model that applies to the complaint items in our model, there are no well‐known measurement theories (a framework for model of the instrument can either be reflective or formative). 7 Therefore, the development of multi‐item instruments following a formative model is merely based on common sense of experts in the field rather than on statistical procedures. 5 , 7
Reliability was defined as ‘the degree to which the measurement is free from measurement error’. 8 Mokkink et al (2010) 9 provided an extended definition of reliability: the extent to which scores for patients who have not changed are the same for repeated measurement under several conditions, for example using different sets of items from the same multi‐item measurement instrument (internal consistency) and over time (test‐retest). Internal consistency of the OQUA was already tested by Bruinewoud et al (2018). 5
The aim of this study was to investigate construct validity, test‐retest reliability and responsiveness of the generic Otology Questionnaire Amsterdam. The OQUA was expected to show good construct validity, good test‐retest reliability and if compared to a similar questionnaire like the Glasgow Health Status Inventory (GHSI) or a Global Rating Scale (GRS) to show good responsiveness.
2. MATERIALS AND METHODS
The evaluation of OQUA's psychometric properties was performed following ‘a practical guide to biostatistics and epidemiology for measurements in medicine’ of de Vet et al 2010 6 and the COSMIN checklist. 10 The translated final version of the OQUA can be found in the appendix as supplement A; the original language is Dutch. All analysis were done using IBM spss Statistics version 22.
2.1. Ethical considerations
The study protocol, including the development of the OQUA, was assessed by the Medical Ethics Review Committee of Amsterdam UMC, VU University Medical Centre, which led to the decision that the Medical Research Involving Human Subjects Act did not apply to this study. Written informed consent was obtained prior to study participation. All data from submitted questionnaires was completely anonymised. Data collection of cohort one was carried out between October 2016 and December 2016 and for cohort two between February 2017 and August 2018.
2.2. Participants
For the assessment of construct validity and test‐retest reliability, a consecutive cohort of 194 patients visiting the ENT surgeons were recruited at Amsterdam UMC, location VUmc, Amsterdam, and at the Amstelland Hospital, Amstelveen, the Netherlands. For the assessment of responsiveness, a second prospective cohort of 50 consecutive patients undergoing ear surgery was recruited one week before their scheduled surgery. Patients were eligible to participate if they were 18 years and over, had an ear complaint, were not known to have a learning disability or cognitive impairment and had a good written understanding of the Dutch language.
2.3. Procedures and measurements
All patients were asked to complete the questionnaires online. Patients in the first cohort were asked to complete the OQUA and GHSI (see below). After 6‐14 days, patients were again asked to complete both questionnaires, assuming that this time interval was sufficient to minimise recall bias, yet short enough for their ear complaints and impact to remain unchanged. Patients in the second cohort were asked to complete the OQUA and GSHI approximately 1 week before their surgery (T0) and again after 3 months after surgery, including two GRSs (T1).
The OQUA consists of eight types of ear complaints domains and one impact domain in a 34‐item questionnaire. Every complaint domain consists of a minimal of two types of items: one item for severity of the complaint measured with a Visual Analogue Scale (VAS; range 0‐100) and one or more item(s) about the frequency of a specific aspect of the complaint measured with a five‐point Likert response scale, ranging from ‘almost never’ (0p) to ‘almost always’ (5p). For both domains, a higher score indicates more complaints or higher impact.
The GHSI contains 18 items and measures the effect of health problem on the quality of life of a person. 11 The response to each question is based on a five‐point Likert scale ranging from high health status through to low health status. A total score and 3 different subscales can be calculated (general subscale, social support score and physical health score). All scores range from 0‐100. A higher score means less impact and better health status. 11
Two ‘GRS questions’ were administered at 3‐month follow‐up. Patients were asked about the change in ear complaints before and after surgery and the change of impact of the complaints on their quality of live before and after surgery. Response options were on a five‐point Likert scale ranging from prominent increase of complaints to great improvement.
2.4. Statistical analyses
We produced descriptive statistics for the scores of the measurements. No missing data were expected in both cohorts since the online questionnaire could only be submitted when completed.
2.5. Construct validity
Part of construct validity is inter‐item correlation (ICC). In this study, construct validity was furthermore assessed by confirmatory factor analysis and hypothesis testing. Fit parameters were used to determine whether the data fitted the hypothesised factor structure. To evaluate model fit, the comparative fit index (CFI), Tucker‐Lewis Index (TLI) and the root mean square error of approximation (RMSEA) were calculated. 12 Guidelines, proposed by Hu and Bentler, 12 suggest that models with CFI and TLI close to 0.95 or higher and RMSEA close to 0.06 or lower are representative of good fitting models.
For hypotheses testing, correlation hypotheses were computed comparing the sumscore of the impact items of the OQUA with the sumscore GHSI and its subscales. 13 Spearman's rho correlations were used for assessing the hypothesised relations because scores were non‐normally distributed. Correlation was considered as low <.30; moderate .30‐.59; and high ≥.60. 14
2.6. Test‐retest reliability
Patients with active ear infections or who had complaints which were expected to change within 1‐2 weeks after their visit at the ENT surgeon were regarded as unstable patients and were excluded from reliability analyses. To investigate the test‐retest reliability of the OQUA, the quadratic weighted kappa was calculated for items measured with ordinal scales, and the intra‐class correlation coefficient (ICC) was calculated for items measured with continuous scales using two‐way ANOVA random effect models for agreement. An ICC value of 0.70, in a sample of 50 patients, was recommended as a minimum standard for reliability. 15 The kappa scores were interpreted through the Landis and Koch classification system; this system divides the score into 5 different classes, and the higher the score means better the reliability. 16
2.7. Responsiveness
Responsiveness was based on testing predefined hypotheses regarding the relationship between the OQUA, GSHI and GRS questions. We hypothesised good correlation between the change in OQUA impact domain and GSHI and less, but significant, correlation between change in OQUA complaints and GRS because multiple complaints had to be compared with one GRS. Pearson's correlation coefficients (r) were used to determine the associations between the OQUA and the GHSI and GRS.
2.8. Scoring system
To interpret the OQUA results, a scoring model had to be formulated. Because of the absence of a comparable scoring system of comparable questionnaire, one had to be developed. The scoring system was developed by an expert panel and authors of this article.
3. RESULTS
3.1. Patient characteristics
In Table 1, patient characteristics of the 2 consecutive cohorts are presented. In the first cohort, 194 patients participated. With regard to the follow‐up measurement, 124 patients gave reply on our request to fill in the follow‐up questionnaire (64%). No differences in demographics between the entire group of 194 patients and the subgroup of 124 were calculated. In the second cohort, 50 patients participated, but due to missing data, 3 patients had to be excluded. In total, 47 patients were eligible for responsiveness analyses.
TABLE 1.
Demographic characteristics. T0 cohort (n = 194) underwent hypothesis testing, internal consistency and inter‐item correlation. T1 cohort (n = 124) underwent test‐retest and responsiveness (n = 47)
| Characteristics | Validity | Percentage (%) | Reliability [test‐retest] | Percentage (%) | Responsiveness | Percentage (%) |
|---|---|---|---|---|---|---|
| T0 (N = 194) | T1 (N = 124) | T2 (N = 47) | ||||
| Age, y | ||||||
| Mean ± SD | 53 ± 16 | — | 55 ± 15 | — | 52 ± 15 | — |
| Range (min.‐max.) | 18‐82 | — | 18‐82 | — | 18‐78 | — |
| Sex | ||||||
| Female | 94 | 48.5 | 57 | 46.0 | 26 | 55.3 |
| Male | 100 | 51.5 | 67 | 54.0 | 21 | 44.7 |
| Hospital | ||||||
| VU Medical Center | 180 | 92.8 | 114 | 91.9 | 44 | 93.6 |
| Amstelland Hospital | 14 | 7.2 | 10 | 8.1 | 3 | 6.4 |
| Type of surgery performed | ||||||
| Stapedotomy | N/A | — | N/A | — | 9 | 19.2 |
| Cochleair implant | N/A | — | N/A | — | 7 | 14.9 |
| Ossiculair chain reconstruction | N/A | — | N/A | — | 9 | 19.2 |
| Cholesteatoma surgery | N/A | — | N/A | — | 12 | 25.5 |
| Myringoplasty | N/A | — | N/A | — | 5 | 10.6 |
| Other | N/A | — | N/A | — | 5 | 10.6 |
| Diagnosis | ||||||
| Diagnosis was known by the patient | 143 | 73.3 | 91 | 73.4 | N/A | — |
| Diagnosis was unknown for the patient | 20 | 10.3 | 14 | 11.3 | N/A | — |
| No diagnosis was made (yet) | 31 | 16 | 19 | 15.3 | N/A | — |
| Distribution of diagnosis a | ||||||
| Cholesteatoma | 14 | 9.8 | 10 | 11 | N/A | — |
| Middle ear infection | 20 | 13.9 | 14 | 15.4 | N/A | — |
| Otosclerosis | 16 | 11.2 | 10 | 11 | N/A | — |
| Open cavity | 19 | 13.2 | 12 | 13.2 | N/A | — |
| Hearing impaired | 69 | 48.2 | 46 | 50.5 | N/A | N/A |
| Tinnitus | 45 | 31.4 | 31 | 34 | N/A | N/A |
| Ear canal infection | 35 | 24.4 | 25 | 2.7 | N/A | N/A |
| Excessive ear wax (cerumen) | 15 | 10.4 | 11 | 12.1 | N/A | N/A |
| Benign positional vertigo (BPPV) | 9 | 6.3 | 7 | 7.6 | N/A | N/A |
| Meniere's disease | 2 | 1.4 | 1 | 1.1 | N/A | N/A |
| Neurinoma | 4 | 2.8 | 3 | 3.2 | N/A | N/A |
| Ear canal stenosis | 6 | 4.2 | 4 | 4.4 | N/A | N/A |
| Other | 29 | 20.2 | 17 | 18.6 | N/A | N/A |
A patient can have multiple diagnosis.
Abbreviations: AA: Amino acid; NT: Nucleotide.
3.2. Construct validity
3.2.1. Inter‐item correlation
The inter‐item correlation (ICC) matrix for construct was already performed by Bruinewoud et al 2018 5 and showed good correlation between the individual items. No items had a too high correlation with another item (>.9; ie most likely identical items) nor a too low correlation (<.2; ie no correlation; Table 2).
TABLE 2.
Inter‐item correlation matrix showing good correlation between the individual items
| Impact_1 | Impact_2 | Impact_3 | Impact_4 | Impact_5 | Impact_6 | Impact_7 | Impact_8 | Impact_9 | |
|---|---|---|---|---|---|---|---|---|---|
| Impact_1 | 1.000 | .647 | .547 | .542 | .628 | .481 | .424 | .608 | .551 |
| Impact_2 | .647 | 1.000 | .499 | .641 | .513 | .431 | .491 | .533 | .633 |
| Impact_3 | .547 | .499 | 1.000 | .540 | .574 | .498 | .426 | .559 | .463 |
| Impact_4 | .542 | .641 | .540 | 1.000 | .468 | .467 | .494 | .492 | .600 |
| Impact_5 | .628 | .513 | .574 | .468 | 1.000 | .613 | .610 | .737 | .548 |
| Impact_6 | .481 | .431 | .498 | .467 | .613 | 1.000 | .700 | .695 | .492 |
| Impact_7 | .424 | .491 | .426 | .494 | .610 | .700 | 1.000 | .625 | .519 |
| Impact_8 | .608 | .533 | .559 | .492 | .737 | .695 | .625 | 1.000 | .635 |
| Impact_9 | .551 | .633 | .463 | .600 | .548 | .492 | .519 | .635 | 1.000 |
>.9 means a too high correlation (ie most likely identical items); <.2 means a too low correlation (ie no correlation).
3.2.2. Confirmatory factor analysis (CFA)
The CFA was performed using collapsed data; that is, answer categories containing less than 5% of the answers filled in were combined with the adjacent category. In items 1, 2, 3, 6 and 7, the answer categories 4 and 5 were combined (4; ‘Agree’, 5; ‘Strongly agree’). In Item 4, answer categories 3, 4 and 5 were combined (3; ‘Neutral’, 4; ‘Agree’, 5; ‘Strongly agree’). There were no residual correlations above 0.2, which indicates that there were no local dependencies.
3.2.3. Hypothesis testing
All hypotheses were confirmed. As expected, moderate to high correlation coefficients were found between the impact items of the OQUA and the total score of the GHSI (0.672) and the general subscale of the GHSI (0.721). In addition, in line with expectations, weak to no correlation between the OQUA and the social (0.005) and physical subscale (0.177) of the GHSI was confirmed,
3.3. Reliability
3.3.1. Internal consistency
Cronbach's alpha for the impact items of the OQUA was .916, indicating strong evidence for a good internal consistency.
3.3.2. Test‐retest reliability
The mean interval between T0 and T1 was 8.69 days (range, 6‐14 days). The quadratic weighted kappa's with associated 95% confidence interval bounds in the parentheses and standard error of measurements of the 26 items, 9 scored a very good (ak > 0.80) test‐retest reliability. Sixteen items scored good (ak = 0.6‐0.8), and only one item (hearing loss Q1; ‘Can you hear someone approaching from behind?’) scored moderate (ak < 0.6). Four of the items had a very good test‐retest reliability (ICC > 0.8), and the other five had a good test‐retest reliability (ICC = 0.6‐0.8).
3.3.3. Responsiveness
Change scores on both the OQUA impact items and GHSI total score were normally distributed. Responsiveness showed good correlation between OQUA impact and GHSI using Pearson's coefficient (−.711). This result matches our hypothesis. Correlation between the GRS and the complaints of the OQUA was calculated using the Spearman ρ and showed (.623).
Tables 3 and 4 show the responsiveness results in a subgroup analysis of cholesteatoma and otosclerosis patients. No other subgroup analysis was made due to small sample size. In the both subgroups, the preoperative response of the OQUA was spread over multiple complaints. The overall impact preoperative response was higher than postoperative response, but the change in complaints varies from patient to patient.
Table 3.
Responsiveness subgroup cholesteatoma: Pre‐ and post‐intervention score impact and complaints

Numbers in green and bold show improvement, in red and italic shows deterioration and it remains black if unaltered.
Table 4.
Responsiveness subgroup cholesteatoma: Pre‐ and post‐intervention score impact and complaints

Numbers in green and bold show improvement, in red and italic shows deterioration and it remains black if unaltered.
3.3.4. Sequence
A considerable number of patients made a comment about the sequence of the questions. In the first version of the OQUA, 5 the sequence of the items was more randomly distributed to avoid duplication of answer categories to items that appeared to be alike. Keeping these questions close together could have interfered with the explorative factor analysis. An example of an item which appeared to be alike is ‘I have ear discharge’ compared to ‘Liquid comes out of my ear’. After finishing the validation and item reduction, we have changed the sequence of the items in order to be more logical and bringing the items about the same domain together. In Figure 1, the latest version of the OQUA is presented.
FIGURE 1.

OQUA questionnaire (English translation, Original language Dutch). The questionnaire contains 34 items of which 9 are about impact and 25 about the 8 most prominent ear complaints (earache, pressure sensation, itching, tinnitus, hearing loss, ear discharge, loss of taste, dizziness). Each complaint has one question in VAS scale to score the severity of the complaint. 18 questions with a 5‐likert scale answer category. Scoring system is explained in Figure 2.
3.3.5. Scoring system
In order to implement the OQUA in daily practice, a scoring system was developed (Figure 2). The expert panel decided that every ear complaint in the complaint domain should be equally weighted and also that the frequency (measured with VAS) and severity (measured with 5‐point Likert scales) in the ear complaint domain should be equally weighted. Additionally, every ear complaint domain (eg tinnitus, hearing loss) should have one score and impact scale should have a separate score. It was decided that a total complaint score covering all domains is not desirable and should not be calculated as it hides differences that may occur in the scoring of the separate domain. For example, the tinnitus score could decrease, while at the same time, the hearing score could increase. This would have resulted in an unchanged total score. Whether it is practically correct to mathematically balance the items and complaints, as done now, is an issue that needs further investigation in future. In Figure 3, a graphical representation of how the OQUA result could look in the (electronic) patient file is shown.
FIGURE 2.

Scoring system of the OQUA
FIGURE 3.

Example of graphical representation of the OQUA
4. DISCUSSION
PROMs are essential for the assessment of subjective complaints and disability and are increasingly utilised. Still, in a general ENT practice, the use of otologic questionnaires is limited and it varies strongly between countries and clinics to which extent disease‐ or complaint‐specific questionnaires are used. The OQUA is a newly developed generic otologic questionnaire, designed to measure the severity of ear complaints, the impact of these complaints on daily life and to assess the effectiveness of an intervention from a patients' perspective. It could be useful to avoid all kinds of questionnaires in a single practise and to compare outcomes in a larger group.
In this study, we analysed the validity of the OQUA. Good validity, reliability and responsiveness were confirmed. The validated version of the OQUA consists of 34 items and was slightly changed in its original sequence for practical reasons. It now has a scoring manual and is ready for broader use. This study shows that the OQUA is set for implementation as standard care in ENT practise in an increasing number of countries.
4.1. Strength and weakness of the study
The design of the OQUA is based on the COSMIN guidelines. 10 All required steps for the development of new instruments were taken, and the OQUA was extensively field‐tested with sufficient number of patients. The study population represented an heterogeneous group by including patients of different ages, sexes, complaints and pathologies. Furthermore, the cohort in which the responsiveness of the OQUA was assessed had an additional heterogeneous distribution of surgical procedures.
Although the results confirm our hypotheses and show satisfactory results for construct validity, test‐retest reliability and responsiveness, some limitations of the study should be mentioned. Unfortunately, there is no consensus in literature about the group size to evaluate validity. Fifty patients is advised by de Vet et al. 7 Accordingly, 50 patients were included in the cohort for examining the OQUA's responsiveness, but due to missing data, three patients had to be excluded. However, the remaining sample is thought to be a good representation of a general ENT practice and the number big enough to examine responsiveness. On the other hand, one could argue that this amount is critical for confirmation of responsiveness in each domain.
The GRS used to examine OQUA's complaints scores did not distinguish in the pre‐determined separate domains (eg tinnitus, hearing loss). Rather, the GRS asked for a change in complaints overall. Using this method, complaints might have cancelled each other out in the correlation with GRS or downsized the effect. For example, at baseline (pre‐surgery), hearing loss is the most prominent complaint of patient x which improves in after surgery. In addition, the same patient might suffer from vertigo after surgery which was not present pre‐surgery. The positive and negative result may have downsized the correlation in the responsiveness statistics. Furthermore, the patient might relate the GRS to his/hers most prominent complaint and not to the broader spectrum of complaints. Another limitation could be a partial floor effect; when patients fill‐out rather low scores at the preoperative moment, there will be little to no room for improvement 3 months after surgery.
The decision to calculate separate domain scores and not a total complaint score covering all domains was made based on the fact that the complaints are all separate domains in a formative model and can be seen as different (sub)constructs. Meaning that the items of the different domains not necessarily correlate with each other as can been seen in the development paper of the OQUA by Bruinewoud et al 2018. 5 By using separate scores, a more accurate insight in the patients' scala of complaints is given. This could, on the other hand, also be seen as a disadvantage, as it can be challenging to compare the results of the complaints with other cross‐sectional studies in literature. Many questionnaires often use a single score for the total of complaints or a score for the complaints and impact as a whole. Up till now, studies often used a complaint‐ or disease‐specific questionnaire. Complaint‐specific questionnaires can be compared with the OQUA‐related domain, but will lack information on other ear complaints. Disease‐specific questionnaires are difficult to compare with the OQUA as the questions in a disease‐specific questionnaire will be tailored towards the specific group of patients with their specific complaints. These studies might be designed to evaluated response to treatment, but most likely will only focus on the patients' ear complaints specific for that disease and discarding the possible other ear complaints which can occur after ear surgery (eg tinnitus or taste problems). Disease‐specific questionnaires should preferably be combined with a separate QoL questionnaire to assess the relevance and impact of these complaints on the quality of life. Comparing cross‐sectional ear‐related quality‐of‐life questionnaires with the OQUA will be possible as the OQUA has a separate scoring system on impact.
4.2. Current position of the OQUA
In the systematic review that was mentioned in the introduction, one generic otologic questionnaire, the COQOL, 17 was identified. The COQOL was developed with the aim to asses quality of life and impact of ear complaints of patients visiting an ENT surgeon. The OQUA distinguishes itself from to the COQOL by having two separate scales/instead of one, measuring not only the impact on quality of life, but also the severity of all major ear complaints as separate domains. In line with their objective to measure impact/quality of life, the COQOL validated its responsiveness compared with the SF36 questionnaire. The questionnaire does not include items about the severity of ear complaints. Despite this difference, the COQOL is mapped for economic evaluation, 18 which is still a future perspective of the OQUA. The OQUA is currently implemented in our electronic health record using EPIC version AUG 2019. After implementation, this construct file can be used across the country for all hospitals using Epic as well. Furthermore, the OQUA will also be build‐in in other electronic health records (EHR) used across the country, for example Hix (Chipsoft).
4.3. Future research
The OQUA is designed to be easily applicable in daily practice, being able to be used on a computer, tablet/phone and on paper. With the scoring manual (Figure 2), a graphical representation can be developed according to the wishes of the ENT surgeon and in line with the capacity of their facility (electronic patient file, ICT department or institution). Future studies might focus on enablers of and barriers to its implementation. By adapting the graphical representation to the preferences of the users and having the OQUA being filled out days before consultation, the implementation is more likely to be successful in the end. Furthermore, translation in foreign languages and cross‐culturally validation is of additional value. Chinese validation and Swedish validation are currently under investigation. Optimisation of graphical representation might be needed in the near future.
The OQUA should give us a better insight in the patient complaints and impact in the large amount of different diagnoses in otology. A more evidence‐based approach with the help of this PROM and broad evaluation of the current treatments might change our future treatment and hopefully our results for the best. Longitudinal analysis, before and after otologic interventions, using the OQUA and at different time intervals in large groups should give us these insights.
5. CONCLUSION
A validated generic otologic questionnaire covering all relevant ear complaints, and their impact, is presented. The results of this study show good construct validity, test‐retest reliability and responsiveness, The OQUA is ready for broad implementation for all patients visiting the ENT surgeon with an ear complaint. Its design furthermore makes the OQUA an effective tool to evaluate the effectiveness of surgical and non‐surgical interventions by the ENT surgeon.
CONFLICT OF INTEREST
None to declare.
Kraak JT, van Dam TF, van Leeuwen LM, Kramer SE, Merkus P. The Otology Questionnaire Amsterdam: A generic patient‐reported outcome measure about the severity and impact of ear complaints. Validation, reliability and responsiveness. Clin Otolaryngol. 2020;45:506–516. 10.1111/coa.13545
DATA AVAILABILITY STATEMENT
The data that support the findings of this study are available from the corresponding author, upon reasonable request.
REFERENCES
- 1. Jacobson GP, Newman CW. The development of the Dizziness Handicap Inventory. Arch Otolaryngol Head Neck Surg. 1990;116(4):424‐427. [DOI] [PubMed] [Google Scholar]
- 2. Newman CW, Jacobson GP, Spitzer JB. Development of the Tinnitus Handicap Inventory. Arch Otolaryngol Head Neck Surg. 1996;122(2):143‐148. [DOI] [PubMed] [Google Scholar]
- 3. Philips JS, Haggard M, Yung M. A new health‐related quality of life measure for active chronic otitis media (COMQ‐12): development and initial validation. Otol Neurotol. 2014;35:454‐458. [DOI] [PubMed] [Google Scholar]
- 4. Gates GA, Veral AM. Validation of the ménières disease patient‐oriented symptom‐Severity Index. Arch Otolaryngol Head Neck Surg. 2005;131:863‐867. [DOI] [PubMed] [Google Scholar]
- 5. Bruinewoud EM, Kraak JT, van Leeuwen LM, Kramer SE, Merkus P. The Otology Questionnaire Amsterdam: a generic patient reported outcome measure about the severity and impact of ear complaints. A cross‐sectional study on the development of this questionnaire. Clin Otolaryngol. 2018;43(1):240‐248. [DOI] [PubMed] [Google Scholar]
- 6. Vet HCWd, Terwee CB, Mokkink LB. Knol DL. Measurements in Medicine – A Practical Guide. Cambridge, UK: Cambridge University Press; 2011. [Google Scholar]
- 7. de Vet HCW, Terwee CB, Mokkink LB, Knol DL. Measurement in Medicine. Cambridge, UK: Cambridge University Press; 2011. [Google Scholar]
- 8. Mokkink LB, Terwee CB, Patrick DL, et al. The COSMIN study reached international consensus on taxonomy, terminology, and definitions of measurement properties for health‐related patient‐reported outcomes. J Clin Epidemiol. 2010;63(7):737‐745. [DOI] [PubMed] [Google Scholar]
- 9. Mokkink LBTC, Patrik DL, et al. International consensus on taxonomy, terminology, and definitions of measurement properties for health‐related patient‐reported outcomes: results of the COSMIN study. J Clin Epidemiol. 2010;63:737‐745. [DOI] [PubMed] [Google Scholar]
- 10. Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, et al.COSMIN checklist manual 2012.
- 11. Robinson K, Gatehouse S, Browning GG. Measuring patient benefit from otorhinolaryngological surgery and therapy. Ann Otol Rhinol Laryngol. 1996;105(6):415‐422. [DOI] [PubMed] [Google Scholar]
- 12. Hu L, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct Eq Model. 1999;6(1):1‐55. [Google Scholar]
- 13. Robinson K, Gatehouse S, Browning GG. The Glasgow Health Status Questionnaires Manual. Nottingham, UK: MRC Institute of Hearing Research, Glasgow Royal Infirmary; 1998. [Google Scholar]
- 14. Jacob C. Statistical power analysis for the behavioral sciences. (2nd ed). Department of Psychology, University New York, New York, NY: Lawrence Erlbaum Associates Publishers; 1988. 567 p. [Google Scholar]
- 15. Terwee CB, Bot SD, de Boer MR, et al. Quality criteria were proposed for measurement properties of health status questionnaires. J Clin Epidemiol. 2007;60(1):34‐42. [DOI] [PubMed] [Google Scholar]
- 16. Landis JR, Koch GC. The measurement of observer agreement for categorical data. Biometrics. 1977;33:159‐174. [PubMed] [Google Scholar]
- 17. Martin TPC, Moualed D, Paul A, et al. The Cambridge Otology Quality of Life questionnaire: an otologyspecific patient‐reported outcome measure. A paper describing the instrument design and report of preliminart reliability and validity. Clin Otolaryngol. 2015;40:13‐139. [DOI] [PubMed] [Google Scholar]
- 18. Frew EJ, Harrison M, Rossello Roig M, Martin TP. Providing an extended use of an otological‐specific outcome instrument to derive cost‐effectiveness estimates of treatment. Clin Otolaryngol. 2015;40(6):593‐599. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data that support the findings of this study are available from the corresponding author, upon reasonable request.
