Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Jul 1.
Published in final edited form as: Ear Hear. 2019 Jul-Aug;40(4):1016–1024. doi: 10.1097/AUD.0000000000000684

Development of the Cochlear Implant Quality of Life (CIQOL) Item Bank

Theodore R McRackan 1, Brittany N Hand 2; Cochlear Implant Quality of Life Consortium, Craig A Velozo 2, Judy R Dubno 1
PMCID: PMC6749825  NIHMSID: NIHMS1510577  PMID: 30531259

Abstract

Objective:

Functional outcomes following cochlear implantation have traditionally been focused on word and sentence recognition, which, although important, do not capture the varied communication and other experiences of adult cochlear implant (CI) users. Although the inadequacies of speech recognition to quantify CI user benefits are widely acknowledged, rarely have adult CI user outcomes been comprehensively assessed beyond these conventional measures. An important limitation in addressing this knowledge gap is that patient reported outcome measures (PROMs) have not been developed and validated in adult CI patients using rigorous scientific methods. The purpose of the present study is to build on our previous work and create an item bank that can be used to develop new PROMs that assess CIQOL in the adult CI population.

Design:

An online questionnaire was made available to 500 adult CI users who represented the adult CI population and were recruited through a consortium of 20 CI centers in the United States. The questionnaire included the 101 question CIQOL item pool and additional questions related to demographics, hearing and CI history, and speech recognition scores. In accordance with the Patient-Reported Outcomes Measurement Information System (PROMIS), responses were psychometrically analyzed using confirmatory factor analysis (CFA) and item response theory (IRT).

Results:

Of the 500 questionnaires sent, 371 (74.2%) subjects completed the questionnaire. Subjects represented the full range of age, durations of CI use, speech recognition abilities, and listening modalities of the adult CI population; subjects were implanted with each of the three CI manufacturers’ devices. The initial item pool consisted of the following domain constructs: communication, emotional, entertainment, environment, independence, listening effort, and social. Through psychometric analysis, after removing locally dependent and misfitting items, all of the domains were found to have sound psychometric properties, with the exception of the independence domain. This resulted in a final CIQOL item bank of 81 items in 6 domains with good psychometric properties.

Conclusion:

Our findings reveal that hypothesis-driven quantitative analyses result in a psychometrically sound CIQOL item bank, organized into unique domains that are comprised of independent items, that measure the full ability range of the adult CI population. The final item bank will now be used to develop new instruments that evaluate and differentiate adult CIQOL across the patient ability spectrum.

Introduction

Substantial evidence exists that cochlear implantation has a dramatic impact on a patient’s life (McRackan et al., 2018a; McRackan et al., 2018b; Olze et al., 2011; Vermeire et al., 2005). However, the standard outcome measures used to assess this impact have primarily focused on word and sentence recognition under conditions that are not representative of typical communication environments and, therefore, correlate poorly with patients’ self-reported communication abilities (Capretta et al., 2016; McRackan et al., 2018a; McRackan et al., 2018b; Moberly et al., 2017). How cochlear implant (CI) users listen, communicate and interact with their environment is far more complex than revealed by commonly used speech recognition tasks, even tasks that include background noise (McRackan et al., 2018a; McRackan et al., 2018b). For example, most CI users rely on both auditory and visual cues for communication (Stevenson et al., 2017), and converse with both ears in the sound field with multiple communication partners in complex listening environments, factors that are not captured with our current test measures. In addition, current outcome measures do not assess the impact of CIs on the social, emotional, and functional aspects of CI users’ lives. These gaps provide support for the development and use of a CI-specific quality of life (QOL) instrument that can comprehensively assess patients’ experiences beyond speech recognition abilities.

Results of several studies and meta-analyses have revealed the significant impact of hearing loss on QOL (Chia et al., 2007) and the improvement that results after cochlear implantation (McRackan et al., 2018a; McRackan et al., 2018b). QOL is typically composed of several constructs or domains that can differ across QOL instruments. For example, following implantation, adult CI recipients may have improved social function (Chung et al., 2012; Hinderink et al., 2000; Olze et al., 2011; Vermeire et al., 2005), better emotional well-being or mental health (Kobosko et al., 2015; Looi et al., 2011; Olze et al., 2011; Vermeire et al., 2005), and decreased listening effort, in addition to improved communication. (Hughes et al., 2018). However, the degree of domain-specific QOL improvement varies greatly depending on the instrument used to measure these outcomes (McRackan et al., 2018a; McRackan et al., 2018b) with some showing no improvement in social (Arnoldner et al., 2014; Damen et al., 2007; Klop et al., 2008) or emotional function (Arnoldner et al., 2014; Damen et al., 2007; Klop et al., 2008).

Consistent with these results, evaluation of health-related quality of life (QOL) through patient-reported outcomes measures (PROMs) has been increasingly emphasized by the Food and Drug Administration (Patrick et al., 2007), Centers for Medicare and Medicaid Services, and specifically in cochlear implantation by the National Institute on Deafness and Other Communication Disorders of the National Institutes of Health (NIH). However, no QOL instruments have been developed for the adult CI population using rigorous, psychometrically sound methodologies. Rather, the most common instruments used with CI recipients are hearing-specific instruments that have not been validated for use by these patients. These include the Hearing Handicap Inventory for Adults/Elderly (HHIA/E) (Newman et al., 1990; Ventry et al., 1982), Speech, Spatial and Quality of Hearing Scale (SSQ) (Gatehouse et al., 2004), and the Abbreviated Profile of Hearing Aid Benefit (APHAB) (Cox et al., 1995), which were developed and validated primarily for individuals who have mild to moderate hearing loss and those who use hearing aids. As such, researchers and clinicians cannot be confident that these instruments accurately and reliably capture the constructs they purport to measure for adults with CI. Moreover, adults with CI may face unique barriers to QOL that may not be captured by instruments developed for and validated with other individuals.

In addition to hearing-specific PROMs, CI-specific PROMs have been used, especially the Nijmegen Cochlear Implant Questionnaire (NCIQ) (Hinderink et al., 2000). The NCIQ includes six QOL domains (basic sound perception, advanced sound perception, speech production, self-esteem, activity, and social interactions), each consisting of ten items that were developed based on expert opinion. Initial validation and testing of the NCIQ was completed using a sample of 91 participants (including 46 controls) from a single clinical site (Hinderink et al., 2000). Although expert consensus was reached on the domains and items included in the NCIQ, the domains do not include certain QOL domains that CI users perceive to be important (Hughes et al., 2018; McRackan et al., 2017). This speaks to the value of directly soliciting input from CI users to ensure that a QOL instrument adequately captures concepts that are important to members of the target population. Moreover, since the NCIQ was established, more rigorous methods of psychometric testing during PROM development have become standard (Cella et al., 2007; Hays et al., 2007; Pilkonis et al., 2011). As such, there is a clear need to develop and validate CI-specific PROMs using modern psychometric techniques and recruitment of a large, representative sample from multiple clinical sites.

The NIH established the Patient-Reported Outcomes Measurement Information System (PROMIS) in 2004 to develop, evaluate, and disseminate PROMs that focus on patient-centered outcomes. Since that time, PROMIS has established rigorous guidelines for how PROMs should be developed and validated (Figure 1). This process includes five steps: (1) comprehensive literature search of existing measures (Klem et al., 2009; McRackan et al., 2018a; McRackan et al., 2018b), (2) focus groups with thematic analysis of discussed topics to create a question or item pool (DeWalt et al., 2007; McRackan et al., 2017), (3) cognitive interviewing for feedback on language and item clarity (DeWalt et al., 2007), (4) confirmation of domain factor structure and item analysis using item response theory (IRT) to develop the item bank (Hays et al., 2007), and (5) validity testing of final QOL instruments (Pilkonis et al., 2014).

Figure 1:

Figure 1:

Flow diagram of steps involved in the development of a QOL instrument using the PROMIS guidelines and our progress to date. The current study establishes the CIQOL item bank, which will be used to develop subsequent instruments. (CAT: computerized adaptive testing)

IRT is the modern standard for evaluation of items for inclusion in a PROM and offers several advantages over classical test theory, which was previously the standard for PROM development and used in the legacy hearing- and CI-specific QOL instruments (discussed later). Perhaps most importantly, instruments developed with IRT are considered to have psychometric properties that are sample- and test-independent (Prieto et al., 2003). Classical test theory is predicated on observed and true scores, which are sample-dependent as subjects will have higher true scores on easy tests and lower true scores on more difficult tests. In contrast, IRT focuses on the measurement of an underlying latent trait, commonly referred to as person ability or person measure, which remains independent of test difficulty.

A second advantage of IRT is the focus on item-level, rather than test-level, psychometrics. IRT analyses provide data on each individual item to determine its characteristics and utility for inclusion in subsequent instruments. Through IRT, researchers can evaluate each item in the pool for ceiling and floor effects, identify fit to the model, match individual item difficulty level to person ability level, and ensure that the items cover the ability range of the population of interest. This analysis leads to the development of an item bank that measures and differentiates individuals across the range of ability levels. This item bank then serves as the source for items to be used for subsequent PROMs including short form, profile, and computerized adaptive testing (CAT) instruments. With the psychometric properties of each item established, selection of items for each instrument is based on highest discrimination across the ability range and best match between item difficulty and subject ability. The result is an optimized instrument with increased capacity to differentiate individuals across a greater range of the latent trait—termed precision (Rose et al., 2008).

Three core assumptions about the item bank must be met (Reeve et al., 2007): (1) items only contribute to one domain of QOL (unidimensionality), (2) responses to each item are unrelated to responses to other items (local independence), and (3) items fit the IRT measurement model (item fit). Confirmatory factor analysis (CFA) is used to confirm unidimensionality and local independence. Items are eliminated from the item pool if they do not substantially contribute to the unidimensional QOL domain captured by the other items, or if responses to the item are dependent upon responses to other items in the pool. In addition, indicators of item fit to the IRT model, such as infit and outfit, are examined to ensure that the included items adequately measure the construct of interest for individuals at ability levels close to and far from the item difficulty.

Building on past work, the current manuscript represents the transition of our research from qualitative methodologies to quantitative analysis to evaluate our item pool using IRT (McRackan et al., 2017). The goal is to create an item bank that will be the source for items to be used in instruments that provide comprehensive and patient-centered evaluations of QOL in the adult CI population.

Materials and Methods

Subjects

In order to enroll a large and diverse sample with respect to age at implantation, sex, CI listening modalities, and communication abilities, we established a Cochlear Implant Quality of Life Development Consortium, which includes 20 CI centers that represent all regions of the United States. Recruitment flyers were distributed to CI recipients electronically and on paper through these centers. Interested patients then emailed our research team to be enrolled. To be enrolled, patients must have been 18–89 years of age, a CI user for at least one year, and not have received a CI for single-sided deafness. This upper age limit was selected as individuals over the age of 89 are considered a special population by our Institutional Review Board, requiring in-person consent, which was not possible given the online nature of the study’s data collection. Data collection was performed using REDCap (Research Electronic Data Capture)—a secure, web-based data collection platform.

Sample size needs were determined based on CFA, which was the most sample-size dependent portion of the analysis. Sample sizes of 300 are considered conservative for CFA under a variety of sample conditions based on Monte Carlo simulations (MacCallum et al., 1999). We assumed a 60% response rate and therefore sent questionnaires to the first 500 subjects who contacted the research team.

Of the 500 patients emailed, 408 (81.6%) returned questionnaires. Subjects with missing data in item pool responses (n=37, 7.4%) were excluded from analyses resulting in 371 (74.2%) subjects with complete data. The 371 subjects who completed the questionnaire included more females than males (Table 1). Most were married but did not have children <18 years of age living in the household. The majority lived in suburban environments, and essentially equal numbers lived in urban and rural locales. The vast majority had some education beyond a high school diploma and either worked full time or were retired. Subjects were fairly evenly split among the household income categories except in the lowest ($0-$20,000). All regions of the United States were represented with the South Atlantic region having the highest number of subjects (25.3%). Individuals from our institution represented only 2.9% of those who completed the questionnaire. Subjects represented the full range of age at implantation, duration of CI use, speech recognition abilities, and listening modalities of the adult CI population and used all three CI manufacturers’ devices (Tables 2 and 3).

Table 1:

Subject demographics

N%
Sex Male 149 (40.2%)
Female 222 (59.8%)
Marital Status Single, never married 54 (14.6%)
Married/Domestic partnership 251 (67.7%)
Widowed 24 (6.5%)
Separated/Divorced 42 (11.3%)
Have Children <18 in the Home Yes 56 (15.1%)
No 315 (84.9%)
Environment Where Subject lives Urban 81 (21.8%)
Suburban 214 (57.7%)
Rural 76 (20.5%)
Race Asian 3 (0.8%)
Black or African American 3 (0.8%)
White 351 (94.6%)
More than one race 4 (1.1%)
Not reported 10 (2.7%)
Ethnicity Hispanic or Latino 13 (3.5%)
Not Hispanic or Latino 300 (80.9%)
Not Reported 58 (15.6%)
Combined Household Income $0-$20,000 26 (7.0%)
$20,001-$50,000 63 (16.9%)
$50,001-$80,000 87 (23.4%)
$80,001-$110,000 66 (17.7%)
>$100,000 93 (25.0%)
Unknown/Not reported 36 (9.7%)
Highest Level of Education No schooling completed 0 (0%)
Nursery school to 8th grade 1 (0.2%)
Some high school, no diploma 2 (0.5%)
High school graduate or equivalent 27 (7.2%)
Some college 55 (14.8%)
Trade/Tech/Vocational training 17 (4.5%)
Associate degree 37 (9.9%)
Bachelor’s degree 112 (30.1%)
Master’s degree 75 (20.2%)
Professional degree 18 (4.9%)
Doctorate degree 27 (7.2%)
Employment status Employed, working ≥40 hours per week 120 (32.3%)
Employed, working <40 hours per week 40 (10.7%)
Not employed, Looking for work 8 (2.1%)
Not employed, not looking for work 17 (4.5%)
Retired 166 (44.7%)
Disabled, not able to work 20 (5.3%)
Region Northeast New England 17 (4.5%)
Northeast Mid-Atlantic 30 (8.0%)
Midwest East North Central 57 (15.3%)
Midwest West North Central 33 (8.8%)
South Atlantic 94 (25.3%)
South East South Central 18 (4.8%)
South West South Central 30 (8.0%)
West Mountain 37 (9.9%)
West Pacific 52 (14.0%)
Not Reported 3 (0.8%)

Table 2:

Subject hearing and CI history

Mean (SD) Range
Age 59.5 (14.9) 19-88
Duration of hearing loss prior to CI (years) 27.1 (18.4) 0-80
Duration of CI use (years) 7.6 (6.5) 1.0-33.0
CNC Word scores (%; n=173) 69.6 (24.4) 0-100
HINT Sentence scores in quiet (%, n=78) 76.1 (30.2) 0-100
AzBio Sentences in quiet (%; n= 185) 81.2 (23.0) 0-100
AzBio Sentence scores in noise at +10 dB SNR (%; n=121) 64.3 (27.5) 0-100

Speech recognition scores were obtained by the subject from their audiologist. N represents the number of subjects who were able to obtain that score. CI: cochlear implant; CNC: (consonant-nucleus-consonant) word recognition test performed in quiet; HINT: (Hearing in Noise Test) sentence recognition test performed in quiet; AzBio: sentence recognition test performed in quiet or in noise at +10 dB signal-to-noise ratio (SNR).

Table 3:

Subject CI device information

N (%)
CI company Advanced Bionics 43 (11.5%)
Cochlear 216 (58.2%)
MED-EL 112 (30.1%)
Listening Modality Unilateral CI with no contralateral HA 87 (23.4%)
Unilateral CI with hearing aid 96 (25.8%)
Bilateral CI 188 (50.6%)
Combined electro-acoustic hearing (Hybrid) No 358 (96.4%)
Yes 12 (3.2%)
No response 1 (0.2%)

Data Collection

The questionnaire consisted of three sections: (1) subject demographics, (2) hearing and CI history (including speech recognition scores), and (3) the CIQOL item pool. Subject demographics are displayed in Table 1. US Census Bureau definitions were used to define geographic region (Bureau, 2010). Subjects were asked to self-identify the developed area where they lived as urban, suburban, or rural. The hearing and CI history collected are included in Table 2 and Table 3. If subjects had bilateral CIs, duration of CI use was based on when the subject had their first CI activated. Subjects received their most recent best aided speech recognition scores from their audiologist and entered them into the questionnaire. CNC words, HINT sentences in quiet, and AzBio sentences in quiet and in noise at a +10 dB signal-to-noise ratio (SNR) were selected as these are part of the minimum standard test battery (2011). CNC word test is an open set 50 monosyllabic word recognition task performed in quiet (Peterson and Lehiste, 1962). HINT and AzBio are open set sentence recognition tasks consisting of 10 and 20 sentences per list, respectively (Nilsson et al., 1994; Spahr et al., 2012). At least one of these speech recognition scores were available for 236 subjects (63.6%). Subjects were not excluded from analyses if they were unable to obtain their scores.

The development of the initial item pool for an adult CIQOL instrument has been previously described (McRackan et al., 2017) and follows the first and second steps in the PROMIS guidelines described earlier. Briefly, a systematic literature search was used to develop a protocol for three adult CI recipient focus groups. Participants in the three focus groups (n=23) were representative of the adult CI population in terms of demographics, communication abilities, and listening modalities and were stratified based on speech recognition ability (McRackan et al., 2017). The development, execution, and analysis of the focus group protocol was based on grounded theory (Ralph N, 2015) and the consolidated criteria for reporting qualitative research (COREQ-32) was followed (Tong et al., 2007). In accordance with COREQ-32, sample size was determined based on data saturation. The 101-item pool was created based on the central and minor themes identified from transcript coding of our CI patient focus groups (McRackan et al., 2017). The majority of items were derived from direct quotes from focus group participants, whereas others were synthesized from similar comments made by multiple focus group participants. The items were reviewed for content validity and clarity by three fellowship-trained neurotologists who routinely perform CI surgeries, two adult CI audiologists, a PhD public health researcher with expertise in community engagement research, a PhD hearing research scientist with expertise in adult hearing loss, and a PhD psychometrician. Afterward, cognitive interviews were performed with 20 additional adults with CIs in order to ensure the clarity of the 101-item pool (McRackan et al., 2017). Item response options used one of the scales recommended by PROMIS: Never, Rarely, Sometimes, Often, Always.

Based on these results, the initial item pool was separated into 7 hypothesized domain constructs, including communication (receptive and expressive communication ability in different situations), emotional (impact of hearing ability on emotional well-being), entertainment (enjoyment and clarity of TV, radio, music, etc.), environmental (ability to distinguish and localize environmental sounds), independence (ability to function without assistance from others), listening effort (degree of effort and resulting fatigue associated with listening), and social (ability to interact in groups and to attend and enjoy social functions). Figure 2 displays the number of items per construct.

Figure 2:

Figure 2:

Flow chart showing the results of psychometric analysis of the item pool to develop the final item bank. The X indicates that the independence domain was not included in the final item bank.

Data Analysis

Descriptive statistics were used to describe the subject demographics. CFA was used to determine the degree to which items in each domain represented a single unidimensional construct. Item residual correlations were used to determine if the responses to each item were unrelated to responses to other items (local independence). The assumption of unidimensionality for each domain was analyzed with an ordered-category CFA with diagonal weighted least squares estimation using the package “lavaan” in the statistical software R (Rosseel et al., 2017). It is best practice in CFA to examine multiple types of fit indicators, including those that are reflective of absolute fit (standardized root mean square residual <0.08), those that have parsimony corrections (root mean square error of approximation <0.08), and comparative fit indicators (comparative fit index >0.95; Tucker-Lewis index >0.95). Acceptable CFA model fit was defined a-priori by standardized root mean square residual, root mean square error of approximation, comparative fit index and Tucker-Lewis index (Brown, 2015). In addition, standardized item factor loadings were examined. A minimum factor loading of 0.32 was chosen as the level of significance, as this equates to approximately 10% overlapping variance with other items in that factor (Tabachnick et al., 2013). Item residual correlations were examined for local dependence, with correlations >0.2 indicating dependence (Pilkonis et al., 2011).

Once a unidimensional set of items for each domain was identified, a one-parameter logistic IRT analysis was conducted. A rating scale model with joint maximum likelihood estimation was conducted using WINSTEPS, version 3.90.0 (Linacre, 2016b). Results of the IRT analysis were examined using a multi-step approach. First, the appropriateness of the rating scale was evaluated using the following criteria (Linacre, 2002): (1) at least 10 observations of each category, collapsed across all items; (2) monotonicity of rating scale categories (that is, 0–4) as evidenced by an increase in average category difficulty with increasing category value; and (3) outfit mean-square is <2.0. Second, the fit of the items and persons to the IRT model was evaluated by examining infit and outfit mean squares and standardized z- values (Linacre, 2002). Mean square values >1.70, as well as standardized z-values greater than 2.0 were considered indicative of misfit to the IRT model (Wright et al., 1994). Third, reliability indicators were examined including: (1) person reliability, which represents the reproducibility of person ordering and was interpreted such that values ≥0.5 were considered adequate, ≥0.80 were considered good, and ≥ 0.90 were considered high,(Linacre, 2016a) and (2) the separation index was used to calculate the number of statistically distinct ability strata in the sample (Wright et al., 2002). The number of person strata is calculated according to the formula (4G+1)3, where G is the person separation index and is an indicator of the number of statistically distinct person measures with centers three calibration errors apart. Test targeting, test coverage, and item hierarchy were examined visually using person-item maps.

Results

Confirmatory Factor Analysis

A summary of CFA results used to evaluate unidimensionality and item local independence is provided in Table 4. Fit indices reflected adequate-to-good model fit for the communication, emotional, and social domains. Root mean square error of approximation indicated poor model fit for the entertainment, environment, independence, and listening effort domains. However, as these domains had fewer degrees of freedom due to smaller item pools, we focused on other indicators of model fit for these domains as evidence suggests that root mean square error of approximation may not be a reliable indicator of fit in these cases (Kenny et al., 2015). Remaining fit indices reflected good model fit for the entertainment, environment, and listening effort domains. However, standardized root mean square revealed poor model fit for the independence domain.

Table 4:

Indicators of unidimensionality and item local independence.

Communication (n=32) Emotional (n=15) Entertainment (n=8) Environment (n=8) Independence (n=11) Listening Effort (n=8) Social (n=19)
RMSEA 0.09 0.06 0.17 0.14 0.20 0.15 0.07
SRMR 0.06 0.05 0.08 0.08 0.12 0.07 0.05
CFI 0.99 0.99 0.99 0.98 0.99 0.98 0.99
TLI 0.99 0.99 0.98 0.96 0.99 0.98 0.99
Locally dependent items 2 0 0 3 0 0 3

n indicates the number of items in each domain. Better fit is denoted by lower values for RMSEA and SRMR, but higher values for CFI and TLI. Fit indices in italics indicate poor model fit. RMSEA: root mean square error of approximation; SRMR: standardized root mean square residual; CFI: comparative fit index; TLI: Tucker-Lewis index.

All items had standardized factor loadings of ≥0.32 on their respective domains. Examination of residual correlation matrices revealed no local dependence for the emotional, entertainment, independence, and listening effort domains. The communication domain had two items that were locally dependent upon one another (residual correlation >0.2). These items were “I have to ask people to look at me when they speak” and “I have to ask a lot of questions about what is being said in a conversation.” The environment and social domains each had three items that demonstrated local dependence. For the environmental domain, the item “Noises from household appliances are bothersome” demonstrated dependence upon the items “Noises in my car are bothersome” and “I am able to hear cars approaching in traffic.” In the social domain, the item “I am able to communicate with my family and friends” demonstrated dependence with the items “I avoid socializing with friends, relatives, or neighbors due to my hearing loss” and “I avoid social situations due to my hearing loss.” For each of these items, the one item was locally dependent with two other items, but the two other items were not dependent upon each other. Therefore, the one item that displayed local dependence from each domain was excluded from subsequent analyses.

Item Response Theory Analysis

A summary of the IRT results are provided in Table 5. The rating scales of all domains met our three a-priori criteria, with the exception of the independence domain. The rating scale for the independence domain did not demonstrate monotonicity or outfit mean square <2.0, indicating poor performance of the rating scale. All items of the emotional, entertainment, and listening effort domains fit the IRT model, whereas the communication, independence, social, and environment domains each had at least one misfitting item. The misfitting items were excluded from the final item bank.

Table 5:

Summary of IRT results.

Communication Emotional Entertainment Environment Independence Listening Effort Social
Number of items 31 15 8 7 11 8 18
Misfitting items 3 0 0 1 2 0 2
Misfitting persons (%) 46 (12.4%)* 34 (9.2%) 26 (7.0%) 23 (6.2%) 22 (5.9%) 30 (8.1%) 39 (10.5%)
Subjects reaching ceiling (%) 0 (0%) 6 (1.6%) 5 (1.3%) 3 (0.8%) 134 (36.1%) 0 (0%) 5 (1.3%)
Subjects reaching floor (%) 0 (0%) 0 (0%) 3 (0.8%) 0 (0%) 1 (0.3%) 0 (0%) 0 (0%)
Mean Person ability (logits) 0.73 1.64 0.82 1.32 2.39 −0.17 1.62
Person Separation 4.51 3.79 2.71 2.16 0.63 2.79 3.36
Person Reliability 0.95 0.93 0.88 0.82 0.29 0.89 0.92
Number of person Strata 6.35 5.39 3.94 3.21 1.17 4.05 4.81
Cronbach’s alpha 0.96 0.95 0.91 0.86 1.00 0.90 0.95
*

Denotes that the value presented represents the number of misfitting persons after the misfitting items were removed.

IRT: item response theory

The number of subjects who misfit the model ranged from n=22 (5.9%; independence) to n=74 (19.9%; communication) across the domains. After removing the three misfitting items from the communication domain, the number of misfitting subjects decreased to 46 (12.4%). The mean person ability was greater than two logits from the mean item difficulty for the independence domain, indicating a poor match between item difficulty and person ability. Separation indices ranged from 0.63 (independence domain) to 4.51 (communication domain) with a larger number representing greater capacity to separate individuals. Minimal ceiling and floor effects were observed for all domains except the independence domain, which showed a 36.1% ceiling effect. All domains demonstrated strong person reliability (>0.80) with the exception of the independence domain (0.29). The number of person strata, which represents the number of statistically distinct ability levels that the items can differentiate, ranged from 1.17 (independence domain) to 6.35 (communication domain).

In summary, the independence domain was not included in the final item bank because it lacked unidimensionality, monotonicity, had poor person reliability and low separation index. After removal of the locally dependent and misfitting items, all other domains and remaining items were included in the final item bank. In total, 9 other items were removed from other domains leaving 81 items in the final item bank and 6 domains (Figure 2). The full CIQOL item bank is provided as supplemental digital content (online Appendix 1).

Discussion

Although many hearing- and CI-specific PROMs have been developed, the current research differs in that it followed stringent and widely accepted guidelines and applied rigorous psychometric methods. This combined use of qualitative and quantitative approaches provides hypothesis-driven methods to develop PROMs. Based on the results of the CI patient focus groups (McRackan et al., 2017), we hypothesized domain constructs that encompass QOL for CI recipients. In addition, given that actual CI patients (rather than expert panels of providers) contributed to item development, we hypothesized that the item pool would appropriately measure the ability range of adult CI patients, providing content validity.

Several domains of the legacy QOL instruments used in the adult CI population align with our results. For example, similar to the CIQOL item bank, the NCIQ and HHIA/E include domains related to social/social interaction and emotional/psychological function. Other domains are unique to the CIQOL item bank, including entertainment, environment, and listening effort. Thus, using focus groups rather than expert panels to create the item bank may uncover topics that have been previously unknown or ignored. Although the SSQ is typically separated into speech, spatial, and qualities of hearing domains, there is some evidence for an additional listening effort domain, which aligns with our findings. However, it is important to note that the listening effort domain did not meet all criteria for a unidimensional construct in a post-hoc analysis of the SSQ and this domain, along with the entire SSQ, has not been validated in the CI population (Akeroyd et al., 2014).

The relationship between the CIQOL communication domain and the NCIQ is more complex. CFA results from the current study clearly demonstrate that that the communication domain represents a unidimensional construct. In contrast, the NCIQ separates communication into three domains: basic sound perception, advanced sound perception, and speech production, but a comparable CFA has not been performed. In addition, IRT analyses provide the difficulty level for each item, which can be used to develop subsequent PROMs. As discussed later, quantitative analysis determined whether an item is “basic” versus “advanced,” rather than relying on the assumptions of expert panels.

The importance of the methodology used to develop the CIQOL can be contrasted to existing hearing- and CI-specific PROMs —using classical test theory or by simply evaluating the test-retest reliability in samples of patients with hearing. A critical step in developing a new QOL instrument is to establish a clear QOL construct to conceptualize the values of the affected population. As noted earlier, this is accomplished using results from patient focus groups, which provide a means for patients to participate directly in the item development process, rather than relying on expert panels (Hays et al., 2007; Velozo et al., 2012). This ensures content validity of the items used in the PROMs and allows interpretation of results within a meaningful QOL framework.

The manner in which domains were developed and evaluated is another advantage of the current methods. Here, thematic analysis of the focus group transcripts was used to create hypothesized constructs that compose QOL. CFA was then applied to ensure that the items in each domain contribute to a single construct (unidimensionality). This quantitative approach to confirm that each domain is measuring a unique construct has rarely been applied in our field. Rather, domains have been traditionally selected based on the research team’s or expert panel’s opinions and accepted without thorough analysis or input from the affected population (Hinderink et al., 2000), as described earlier for the NCIQ. The application of CFA in this manner allows researchers to have greater certainty of what is actually being measured within each domain.

The importance of this methodology is highlighted in the results of the current study where the independence domain did not meet criteria for inclusion in the final item bank. Focus group participants identified independence as an important QOL theme as it relates to their CI. However, our analysis showed that the items in this domain were important, but they represented more than one distinct QOL construct (multidimensional), did not show monotonicity, poorly stratified subjects with regard to ability, and had a ceiling effect for 36.1% of subjects. In contrast, the emotion domain demonstrated the next highest ceiling effect, where 1.6% of subjects had maximum scores.

In addition, previous hearing- and CI-specific PROMs have relied on Cronbach’s α as a marker of internal consistency and reliability (Hinderink et al., 2000; Newman et al., 1990), although it is known to over-estimate both (Sijtsma, 2009). Here, the importance of more in-depth analysis is seen as Cronbach’s α of the independence domain was 1.0, but person reliability was well below acceptable standards (0.29). One may argue that these data suggest that items in this domain should be changed to be made more difficult to provide a better fit with the CI population. However, these items were developed from patient focus group responses on how their implants have impacted their lives. To add more difficult items to this domain as a means to improve psychometric qualities would sacrifice content validity, resulting in an instrument that is not relevant the population of interest.

Rather than relying on traditional methods (such as classical test theory and Cronbach’s α), the application of CFA and IRT in this analysis provides a means to test these hypotheses at the item and group levels to create an item bank that represents and stratifies adult CI patients with regard to QOL. This analysis resulted in six domains with 81 psychometrically sound items in the item bank. Our findings show that each of these domains represents a single construct, which has not been determined for previous hearing- and CI-specific PROMs. This aids in interpretation of results because it is understood that each domain represents a singular latent trait. In addition, we know that the item bank has the capacity to measure the full range of functional abilities within the adult CI population, which has also not been established in legacy hearing- and CI-specific PROMs, such as the HHIA/E, SSQ, APHAB, and NCIQ. With the psychometric properties of the item bank, domain constructs, and individual items known, these data can then be used to develop a suite of new QOL instruments.

Future Directions

The analysis resulted in a psychometrically sound CIQOL item bank with domains and items removed that did not meet accepted standards. We will use the data from this psychometric analysis to select the optimal items for new short-form and profile CIQOL instruments and to develop a CAT. Short-form PROMs provide a global evaluation of a CI user’s QOL and are important for routine use in the busy clinical setting where clinicians and patients may not have sufficient time to complete longer instruments. In addition, short-form instruments are ideal for inclusion in research protocols to minimize the effect of questionnaire fatigue when multiple PROMs are used (Porter et al., 2004). In contrast, profile instruments are longer but expand upon the short-form evaluation and provide additional domain-specific QOL data. CAT, the most advanced and efficient method to administer PROMs, incorporate real-time IRT to select subsequent items based on a patient’s responses to previous items. As such, items presented are dynamically selected for an individual patient based on their ability. Relative to short-forms, CAT minimizes floor and ceiling effects and allows greater differentiation among individual patients relative to static instruments. Moreover, because items are individually selected based on ability, differentiating patient responses can be accomplished using a minimal number of items, thus reducing patient and practitioner burden (Fries et al., 2014). The information obtained through the psychometric evaluation of the item pool to create the CIQOL item bank will guide the selection of items for short-form and profile instruments and development of the CIQOL CAT.

Study Limitations

The study’s limitations are inherent to the online design. Overall, the study population tended to be relatively well-educated and have high socioeconomic status. Nevertheless, benefits of enrolling a large population from multiple CI centers (representing diverse geographic locations), which is required for CFA and IRT analyses, outweighed this limitation. In addition, the items were developed using focus groups that were more racially diverse (McRackan et al., 2017). The current study’s sample does not match the ethnic and racial demographics of the general public. However, no data are available regarding ethnic and racial demographics of the US adult CI population. Even large studies of CI utilization (Agabigum et al., 2018; Sorkin, 2013) and CI outcomes (Angelo et al., 2016; Chung et al., 2012; Gifford et al., 2008; Holden et al., 2013; Reeder et al., 2014; Wanna et al., 2014) routinely fail to report race and ethnicity data. The lack of racial/ethnic data for the adult CI population makes it difficult to determine the effects of this study’s demographics on our results. Although lack of racial/ethnic diversity was a potential limitation, a diverse sample of subjects was enrolled with respect to household income, living environments, employment status, age at implantation, duration of hearing loss, duration of CI use, device types, and CI listening modality.

An additional limitation was speech recognition measures, which were not available for all subjects (unavailable for 36.4%). In addition, because subjects were recruited from a large number of institutions, the specific conditions in which speech recognition tasks were performed could not be controlled. Nevertheless, the results reported revealed that subjects’ speech recognition outcomes were consistent with published data and represented full range of speech recognition abilities (Gifford et al., 2008). Given the known weak correlation between speech recognition ability and self-reported QOL in the CI population (Capretta et al., 2016; McRackan et al., 2018a; McRackan et al., 2018b; Moberly et al., 2017), recruiting an adequate sample size for robust psychometric analyses was more important than obtaining speech recognition data for all subjects. However, determining how speech recognition ability, patient demographics, and hearing/CI history impact CIQOL is important and will be addressed in future studies.

Conclusions

This study builds on our prior work in the development of adult CIQOL PROMs using rigorous established guidelines. We have completed the systematic review, patient focus groups, cognitive interviews, and psychometric analyses to create a CIQOL item bank, toward the overall goal of developing a suite of CIQOL instruments. The value of using these methodologies is seen through the elimination of a domain and items that were not psychometrically valid, ensuring that the items included in the bank covered the ability range of the adult CI population. We also determined each item’s difficulty level, which will be used to optimally select items for other CIQOL instruments. Our future work will develop and validate a suite of short-form, profile, and CAT-based CIQOL instruments and compare results using these new instruments to legacy CIQOL instruments and functional outcome measures. Following additional research, these CIQOL instruments are anticipated to have significant impact on patient outcomes by revealing the communication, emotional, and social benefits of CIs in adults, which will provide tools to determine CI candidacy, set appropriate expectations, identify domain-specific therapies, and determine how device technologies, listening modalities, and novel processing strategies impact CIQOL outcomes.

Supplementary Material

Supplemental Data File (.doc, .tif, pdf, etc.)

Acknowledgements:

The Cochlear Implant Quality of Life Development Consortium consists of the following institutions (and individuals): University of Cincinnati (Ravi N. Samy, MD), University of Colorado (Samuel P. Gubbels, MD), Columbia University (Justin S. Golub, MD, MSCR), House Ear Clinic (Eric P. Wilkinson, MD; Dawna Mills, AuD), Johns Hopkins University (John P. Carey, MD), Kaiser Permanente-Los Angeles (Nopawan Vorasubin, MD), Kaiser Permanente-San Diego (Vickie Brunk, AuD), Mayo Clinic Rochester (Matthew L. Carlson, MD; Colin L. Driscoll, MD; Douglas P. Sladen, PhD), Medical University of South Carolina (Elizabeth L. Camposeo, AuD; Meredith A. Holcomb AuD; Paul R. Lambert, MD; Ted A. Meyer, MD PhD; Cameron Thomas, BS), Ohio State University (Aaron C. Moberly, MD), Stanford University (Nikolas H. Blevins, MD; Jannine B. Larky, MA), University of Maryland (Ronna P Herzano, MD, PhD), University of Miami (Michael E. Hoffer, MD; Sandra M. Prentiss, PhD), University of Pennsylvania (Jason Brant, MD), University of Texas Southwestern (Jacob B. Hunter, MD; Brandon Isaacson, MD; J. Walter Kutz, MD), University of Utah (Richard K. Gurgel, MD), Virginia Mason Medical Center (Daniel M. Zeitler, MD), Washington University-Saint Louis (Craig A. Buchman, MD; Jill B. Firszt, PhD); Vanderbilt University (Rene H. Gifford, PhD; David S. Haynes, MD; Robert F. Labadie, MD PhD). TRM had full access to all the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.

Conflicts of interest and sources of funding:

This research was made possible by funding from a K12 award through the South Carolina Clinical & Translational Research (SCTR) Institute, with an academic home at the Medical University of South Carolina, NIH/NCATS Grant Number UL1TR001450, a grant from the American Cochlear Implant Alliance, and a grant from the Doris Duke Charitable Foundation

References:

Centers for Medicare and Medicaid Services: Quality Strategy.

PROMIS: Instrument Development and Validation Scientific Standards.

2011. MSTB: The new minimum speech test battery.

  1. Agabigum B, Mir A, Arianpour K, Svider PF, Walsh EM, Hong RS 2018. Evolving Trends in Cochlear Implantation: A Critical Look at the Older Population. Otol Neurotol 39, e660–e664. [DOI] [PubMed] [Google Scholar]
  2. Akeroyd MA, Guy FH, Harrison DL, Suller SL 2014. A factor analysis of the SSQ (Speech, Spatial, and Qualities of Hearing Scale). Int J Audiol 53, 101–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Angelo TC, Moret AL, Costa OA, Nascimento LT, Alvarenga K.e.F. 2016. Quality of life in adult cochlear implant users. Codas 28, 106–12. [DOI] [PubMed] [Google Scholar]
  4. Arnoldner C, Lin VY, Bresler R, Kaider A, Kuthubutheen J, Shipp D, Chen JM 2014. Quality of life in cochlear implantees: comparing utility values obtained through the Medical Outcome Study Short-Form Survey-6D and the Health Utility Index Mark 3. Laryngoscope 124, 2586–90. [DOI] [PubMed] [Google Scholar]
  5. Brown TA 2015. Confirmatory factor analysis for applied research. Second edition. ed. The Guilford Press, New York ; London. [Google Scholar]
  6. Bureau, U.S.C. 2010. Geographic Terms and Concepts - Census Divisions and Census Region [Online] https://www.census.gov/geo/reference/gtc/gtc_census_divreg.html (verified Nov. 14).
  7. Capretta NR, Moberly AC 2016. Does quality of life depend on speech recognition performance for adult cochlear implant users? Laryngoscope 126, 699–706. [DOI] [PubMed] [Google Scholar]
  8. Cella D, Yount S, Rothrock N, Gershon R, Cook K, Reeve B, Ader D, Fries JF, Bruce B, Rose M, Group PC 2007. The Patient-Reported Outcomes Measurement Information System (PROMIS): progress of an NIH Roadmap cooperative group during its first two years. Med Care 45, S3–S11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Chia EM, Wang JJ, Rochtchina E, Cumming RR, Newall P, Mitchell P 2007. Hearing impairment and health-related quality of life: the Blue Mountains Hearing Study. Ear Hear 28, 187–95. [DOI] [PubMed] [Google Scholar]
  10. Chung J, Chueng K, Shipp D, Friesen L, Chen JM, Nedzelski JM, Lin VY 2012. Unilateral multi-channel cochlear implantation results in significant improvement in quality of life. Otol Neurotol 33, 566–71. [DOI] [PubMed] [Google Scholar]
  11. Cox RM, Alexander GC 1995. The abbreviated profile of hearing aid benefit. Ear Hear 16, 176–86. [DOI] [PubMed] [Google Scholar]
  12. Damen GW, Beynon AJ, Krabbe PF, Mulder JJ, Mylanus EA 2007. Cochlear implantation and quality of life in postlingually deaf adults: long-term follow-up. Otolaryngol Head Neck Surg 136, 597–604. [DOI] [PubMed] [Google Scholar]
  13. DeWalt DA, Rothrock N, Yount S, Stone AA, Group PC 2007. Evaluation of item candidates: the PROMIS qualitative item review. Med Care 45, S12–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Fries JF, Witter J, Rose M, Cella D, Khanna D, Morgan-DeWitt E 2014. Item response theory, computerized adaptive testing, and PROMIS: assessment of physical function. J Rheumatol 41, 153–8. [DOI] [PubMed] [Google Scholar]
  15. Gatehouse S, Noble W 2004. The Speech, Spatial and Qualities of Hearing Scale (SSQ). Int J Audiol 43, 85–99. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Gifford RH, Shallop JK, Peterson AM 2008. Speech recognition materials and ceiling effects: considerations for cochlear implant programs. Audiol Neurootol 13, 193–205. [DOI] [PubMed] [Google Scholar]
  17. Hays RD, Liu H, Spritzer K, Cella D 2007. Item response theory analyses of physical functioning items in the medical outcomes study. Med Care 45, S32–8. [DOI] [PubMed] [Google Scholar]
  18. Hinderink JB, Krabbe PF, Van Den Broek P 2000. Development and application of a health-related quality-of-life instrument for adults with cochlear implants: the Nijmegen cochlear implant questionnaire. Otolaryngol Head Neck Surg 123, 756–65. [DOI] [PubMed] [Google Scholar]
  19. Holden LK, Finley CC, Firszt JB, Holden TA, Brenner C, Potts LG, Gotter BD, Vanderhoof SS, Mispagel K, Heydebrand G, Skinner MW 2013. Factors affecting open-set word recognition in adults with cochlear implants. Ear Hear 34, 342–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Hughes SE, Hutchings HA, Rapport FL, McMahon CM, Boisvert I 2018. Social Connectedness and Perceived Listening Effort in Adult Cochlear Implant Users: A Grounded Theory to Establish Content Validity for a New Patient-Reported Outcome Measure. Ear Hear 39, 922–934. [DOI] [PubMed] [Google Scholar]
  21. Kenny DA, Kaniskan B, DB M 2015. The Performance of RMSEA in Models With Small Degrees of Freedom. Sociological Methods & Research 44, 486–507. [Google Scholar]
  22. Klem M, Saghafi E, Abromitis R, Stover A, Dew MA, Pilkonis P 2009. Building PROMIS item banks: librarians as co-investigators. Qual Life Res 18, 881–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Klop WM, Boermans PP, Ferrier MB, van den Hout WB, Stiggelbout AM, Frijns JH 2008. Clinical relevance of quality of life outcome in cochlear implantation in postlingually deafened adults. Otol Neurotol 29, 615–21. [DOI] [PubMed] [Google Scholar]
  24. Kobosko J, Jedrzejczak WW, Pilka E, Pankowska A, Skarzynski H 2015. Satisfaction With Cochlear Implants in Postlingually Deaf Adults and Its Nonaudiological Predictors: Psychological Distress, Coping Strategies, and Self-Esteem. Ear Hear 36, 605–18. [DOI] [PubMed] [Google Scholar]
  25. Linacre J 2016a. Rasch Measurement Computer Program: User’s Guide [Online] http://www.winsteps.com/winman/principalcomponents.htm . (verified 11/22/17).
  26. Linacre J 2016b. Rasch Measurement Computer Program [Online] http://www.winsteps.com/ (verified 11/14/17).
  27. Linacre JM 2002. Optimizing rating scale category effectiveness. J Appl Meas 3, 85–106. [PubMed] [Google Scholar]
  28. Looi V, Mackenzie M, Bird P, Lawrenson R 2011. Quality-of-life outcomes for adult cochlear implant recipients in New Zealand. N Z Med J 124, 21–34. [PubMed] [Google Scholar]
  29. MacCallum RC, Widaman KF, Zhang S, Hong S 1999. Sample size in factor analysis. Psychological Methods 4, 84–99. [Google Scholar]
  30. McRackan TR, Bauschard M, Hatch JL, Franko-Tobin E, Droghini HR, Nguyen SA, Dubno JR 2018a. Meta-analysis of quality-of-life improvement after cochlear implantation and associations with speech recognition abilities. Laryngoscope 128, 982–990. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. McRackan TR, Bauschard M, Hatch JL, Franko-Tobin E, Droghini HR, Velozo CA, Nguyen SA, Dubno JR 2018b. Meta-analysis of Cochlear Implantation Outcomes Evaluated With General Health-related Patient-reported Outcome Measures. Otol Neurotol 39, 29–36. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. McRackan TR, Velozo CA, Holcomb MA, Camposeo EL, Hatch JL, Meyer TA, Lambert PR, Melvin CL, Dubno JR 2017. Use of Adult Patient Focus Groups to Develop the Initial Item Bank for a Cochlear Implant Quality-of-Life Instrument. JAMA Otolaryngol Head Neck Surg 143, 975–982. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Moberly AC, Harris MS, Boyce L, Vasil K, Wucinich T, Pisoni DB, Baxter J, Ray C, Shafiro V 2017. Relating quality of life to outcomes and predictors in adult cochlear implant users: Are we measuring the right things? Laryngoscope. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Newman CW, Weinstein BE, Jacobson GP, Hug GA 1990. The Hearing Handicap Inventory for Adults: psychometric adequacy and audiometric correlates. Ear Hear 11, 430–3. [DOI] [PubMed] [Google Scholar]
  35. Nilsson M, Soli SD, Sullivan JA 1994. Development of the Hearing in Noise Test for the measurement of speech reception thresholds in quiet and in noise. J Acoust Soc Am 95, 1085–99. [DOI] [PubMed] [Google Scholar]
  36. Olze H, Szczepek AJ, Haupt H, Forster U, Zirke N, Grabel S, Mazurek B 2011. Cochlear implantation has a positive influence on quality of life, tinnitus, and psychological comorbidity. Laryngoscope 121, 2220–7. [DOI] [PubMed] [Google Scholar]
  37. Patrick DL, Burke LB, Powers JH, Scott JA, Rock EP, Dawisha S, O’Neill R, Kennedy DL 2007. Patient-reported outcomes to support medical product labeling claims: FDA perspective. Value Health 10 Suppl 2, S125–37. [DOI] [PubMed] [Google Scholar]
  38. Peterson GE, Lehiste I 1962. Revised CNC lists for auditory tests. J Speech Hear Disord 27, 62–70. [DOI] [PubMed] [Google Scholar]
  39. Pilkonis PA, Yu L, Dodds NE, Johnston KL, Maihoefer CC, Lawrence SM 2014. Validation of the depression item bank from the Patient-Reported Outcomes Measurement Information System (PROMIS) in a three-month observational study. J Psychiatr Res 56, 112–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Pilkonis PA, Choi SW, Reise SP, Stover AM, Riley WT, Cella D, Group PC 2011. Item banks for measuring emotional distress from the Patient-Reported Outcomes Measurement Information System (PROMIS®): depression, anxiety, and anger. Assessment 18, 263–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Porter SR, Whitcomb ME, Weitzer WH 2004. Multiple surveys of students and survey fatigue. New Directions for Institutional Research 2004, 63–73. [Google Scholar]
  42. Prieto L, Alonso J, Lamarca R 2003. Classical Test Theory versus Rasch analysis for quality of life questionnaire reduction. Health Qual Life Outcomes 1, 27. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Ralph N BM, Chapman Y 2015. The methodological dynamism of grounded theory. International Journal of Qualitative Methods. 14 4. [Google Scholar]
  44. Reeder RM, Firszt JB, Holden LK, Strube MJ 2014. A longitudinal study in adults with sequential bilateral cochlear implants: time course for individual ear and bilateral performance. J Speech Lang Hear Res 57, 1108–26. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Reeve BB, Hays RD, Bjorner JB, Cook KF, Crane PK, Teresi JA, Thissen D, Revicki DA, Weiss DJ, Hambleton RK, Liu H, Gershon R, Reise SP, Lai JS, Cella D, Group PC 2007. Psychometric evaluation and calibration of health-related quality of life item banks: plans for the Patient-Reported Outcomes Measurement Information System (PROMIS). Med Care 45, S22–31. [DOI] [PubMed] [Google Scholar]
  46. Rose M, Bjorner JB, Becker J, Fries JF, Ware JE 2008. Evaluation of a preliminary physical function item bank supported the expected advantages of the Patient-Reported Outcomes Measurement Information System (PROMIS). J Clin Epidemiol 61, 17–33. [DOI] [PubMed] [Google Scholar]
  47. Rosseel Y, Byrnes J, Vanbrabant L, Savalei V, Merkle E, Hallquist M, Rhemtulla M 2017. Package ‘lavaan’ [Online] https://cran.r-project.org/web/packages/lavaan/lavaan.pdf (verified 11/12/17).
  48. Sijtsma K 2009. On the Use, the Misuse, and the Very Limited Usefulness of Cronbach’s Alpha. Psychometrika 74, 107–120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Sorkin DL 2013. Cochlear implantation in the world’s largest medical device market: utilization and awareness of cochlear implants in the United States. Cochlear Implants Int 14 Suppl 1, S4–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Spahr AJ, Dorman MF, Litvak LM, Van Wie S, Gifford RH, Loizou PC, Loiselle LM, Oakes T, Cook S 2012. Development and validation of the AzBio sentence lists. Ear Hear 33, 112–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Stevenson RA, Sheffield SW, Butera IM, Gifford RH, Wallace MT 2017. Multisensory Integration in Cochlear Implant Recipients. Ear Hear 38, 521–538. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Tabachnick BG, Fidell LS 2013. Using multivariate statistics. 6th ed Pearson Education, Boston. [Google Scholar]
  53. Tong A, Sainsbury P, Craig J 2007. Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. Int J Qual Health Care 19, 349–57. [DOI] [PubMed] [Google Scholar]
  54. Velozo CA, Seel RT, Magasi S, Heinemann AW, Romero S 2012. Improving measurement methods in rehabilitation: core concepts and recommendations for scale development. Arch Phys Med Rehabil 93, S154–63. [DOI] [PubMed] [Google Scholar]
  55. Ventry IM, Weinstein BE 1982. The hearing handicap inventory for the elderly: a new tool. Ear Hear 3, 128–34. [DOI] [PubMed] [Google Scholar]
  56. Vermeire K, Brokx JP, Wuyts FL, Cochet E, Hofkens A, Van de Heyning PH 2005. Quality-of-life benefit from cochlear implantation in the elderly. Otol Neurotol 26, 188–95. [DOI] [PubMed] [Google Scholar]
  57. Wanna GB, Noble JH, Carlson ML, Gifford RH, Dietrich MS, Haynes DS, Dawant BM, Labadie RF 2014. Impact of electrode design and surgical approach on scalar location and cochlear implant outcomes. Laryngoscope 124 Suppl 6, S1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Wright B, GN M 2002. Number of Person or Item Strata. Rasch Measurement Transactions 16, 888. [Google Scholar]
  59. Wright B, Linacre J, Gustafson J, Martin-Löf P 1994. Reasonable mean-square fit values. Rasch Measurement Transactions 8, 370. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Data File (.doc, .tif, pdf, etc.)

RESOURCES