Preprint
Article

This version is not peer-reviewed.

What to measure? Development of a Core Outcome Set to assess remote technologies for cochlear implant users.

A peer-reviewed article of this preprint also exists.

Submitted:

03 September 2025

Posted:

03 September 2025

You are already at the latest version

Abstract

Background/Objectives: Uptake of remote cochlear implant (CI) services is feasible in clinical studies, but implementation into regular clinical practice is limited. Effective implementation requires demonstration of at least equivalent outcomes to in-person care. Use of outcome measures that are relevant, and sensitive to both modes of service facilitates evidence-based provision of CI services. Following on from our study that developed a core outcome domain set (CODS), this study aimed to 1) review current awareness and use of outcome measures used clinically, in-person or remotely, and 2) provide recommendations for a pragmatic core outcome set (COS) to assess remote technologies for CI users. Methods: Expert Australian/New Zealand clinical CI professionals (n=20) completed an online survey regarding use of, and familiarity with, pre-identified outcome measures mapping to the previously identified CODS. Respondents rated the outcomes’ usefulness, ease of use, trustworthiness, and recommendation for future use. Stakeholder workshops (clinician, n=3, CI users n=4) finalised recommendations. Results: Four of the six most regularly used and familiar measures were speech perception tests: BKB-A sentences, CNC words, CUNY sentences, and AB words. The long- and short-form Speech, Spatial, and Qualities of Hearing Scales (SSQ/SSQ-12) were the top-ranked patient reported outcome measures (PROMs). These outcome measures were also perceived as the most trustworthy, easy to use, and likely to be used if recommended. Conclusions: A pragmatic COS, relevant to both remote and in-person delivery of CI services, including recommendations for measurement of service, clinician-measured and patient-reported outcomes and how these might be developed in future is recommended.

Keywords: 
;  ;  ;  ;  ;  

1. Introduction

Opportunities to access cochlear implant (CI) care via remote technologies, rather than the traditional method of accessing care through in-person appointments at specialised clinics, have rapidly expanded in recent years [1,2,3,4,5,6]. Various studies have shown that both synchronous and asynchronous options for remote CI services, such as intraoperative CI telemetry, implant programming, electrode-specific measures, and post-operative assessment of speech recognition, management, and review are possible and feasible [5,6,7,8,9]. The use of telehealth for CI service provision has the potential to significantly improve efficiency, effectiveness, and equity of care for CI users, in a personalised manner. However, it is vital that its implementation is well-considered [1,10,11].
Common barriers to integration of remote care into hearing services for either CIs or hearing aids include lack of adequate funding frameworks; poor integration with current clinical practices; mistrust of the accuracy and quality of the remote care service; measures and outcomes; and audiologists’ confidence in their clients’ ability to utilise the remote technology [1,2,12,13,14]. Nevertheless, most studies show that use of a hybrid system in which both remote and in-person care is provided is the preferred method of service delivery both by clients and audiologists [9,13,14,15].
In order to implement remote care effectively into clinical practice, it is essential to demonstrate that remote service provision provides equivalent, if not superior, care compared to the current standard of in-person clinical care [16]. This is necessary for regulatory purposes, as well as to ensure that CI service providers and users have sufficient trust and confidence in the outcomes of the remote service to consider using it.
Within audiology, a vast number of outcome measures exist for measuring the effectiveness of CIs and hearing aids [17,18,19]. Danermark et al. [20] suggested a concise set of outcome measures for assessment of hearing in general, and Allen et al. [21] identified a core outcome domain set for hearing rehabilitation, primarily for hearing aids. These sets of measures and domains, however, do not address some of the specific auditory issues associated with severe-profound hearing loss, or the technical issues associated with the use of CIs. More specific to CIs, Andries et al. [22] recommended a CI-specific outcome assessment protocol for adult CI users which an expert group of CI professionals selected based on the WHO international classification of Functioning, Disability and Health (ICF) framework. The ability to assess outcomes for CIs, such as speech perception, is particularly problematic in remote care given the difficulties determining presentation levels and establishing standardised test environments (e.g. a sound-treated booth), compared to in-person clinical measures. Currently, there is no set of measures for CI specific outcomes when also used in combination with remote technology.
The use of relevant and sensitive outcome measures to evaluate CI services delivered via remote technologies is vital to facilitate the provision of evidence-based health care services, allowing stakeholders to make informed decisions about how to best care for their patients. The current approach to audiological outcome measures is essentially non-standardised [23] both for in-person and remote services, making it difficult to compare and integrate results across different studies and services, for example in systematic reviews with meta-analyses.
To address these issues, over the last decade there has been an increase in the development and use of core outcome sets (COSs) [24]. A COS is an agreed standardised set of outcomes that should be measured and reported as a minimum dataset for a specific condition [25], ideally with input from end-users including patients, clinicians, industry and other key stakeholders. Outcome measures are identified as part of pre-specified outcome domains. A core outcome domain set (CODS) has been defined for hearing aids with separate, and significant, input from both patient and hearing care professional stakeholder groups [26] based on best practice guidelines [27], however there is nothing similar for CIs. Traditionally, CI outcomes have focused on the domains of speech perception and CI uptake, although there is growing evidence that self-reported measures offer a more functional real-world outcome, tapping into different mechanisms of benefit to speech perception outcomes [28,29]. More recently, there has been a growing number of self-reported measures specific to cochlear implants [30,31], although the extent of how they are used in clinical practice is unclear.
With the increase in use of remote technologies there is a need to consider the meaningful domains that are specific to these technologies, which may also be relevant to in-person services. For example, empowerment has recently emerged as a feature of remote technologies [32,33], but it likely also applies to in-person services. Furthermore, there are other considerations specific to service delivery of remote technologies that are often identified as benefits to both patients and services, such as reduced time, convenience and costs [1,2,12,34].
An essential tenet of this research is to be able to demonstrate the equivalence of remote care services, either as a stand-alone or hybrid model of care, to the ‘gold-standard’ in-person clinical model of care. Thus, the overall objective of this research was to develop a COS to evaluate remote technologies delivered within CI services to maximise the potential benefits of this model of care. Feedback was sought from relevant parties (e.g. CI users and their families service providers, including management and clinicians, CI manufacturers, CI advocacy groups) to ensure a broad range of perspectives were considered. Outcome domains encompassing measures specific to the delivery of remote technologies in terms of both (i) patient outcomes (i.e. benefits of remote technologies for CI patients), and (ii) service delivery, were included, ensuring the outcomes can be easily integrated into clinical care. This paper reports on the final phase of a broader 3-phase study (see Figure 1) that followed the COS development roadmap described by Hall et al (2015). The study was registered on the COMET (Core Outcome Measures in effectiveness Trials) website https://www.comet-initiative.org/Studies/Details/2586.
Phase one of the study included a systematic review of outcomes measures identified in studies documenting use of remote services for the provision of CI and hearing aid care [29]. A total of 250 different outcome measures were identified, with CI studies revealing significantly more outcomes in the ear and labyrinth domains compared to hearing aid studies (43% vs 10%), and hearing aid studies revealing significantly more outcomes in the cognitive (28% vs 5%) and emotional (35% vs 10%) domains than CI studies.
Phase two involved a combination of stakeholder workshops with CI users and their significant others, CI professionals, and hearing advocates, followed by a series of three parallel e-Delphi reviews conducted separately for CI professionals and CI users across Australia, the UK and USA. This utilised a methodology described by Allen et al [35] outlining development of a CODS for adult CI outcome domains. This phase aimed to identify by consensus, the most important outcome domains based on stakeholder input [35]. The Delphi review assessed 58 domains across three supradomains: Service, Clinical (assessment-based) and Patient (self-report). The top three domains in which consensus of ≥80% was achieved within each supradomain for both groups (i.e. CI users and CI professionals), are shown in Table 1. Agreement was good for the Service supradomain, however consensus was poorer for the Clinical supra-domain and there was no between-group agreement for the Patient supra-domain. Many domains ranked highly by CI users were ranked far less important by professionals.
The aim of the final phase of the study reported in this paper, Phase 3, was to identify a COS to evaluate remote technologies delivered within CI services based on the previously defined CODS. Due to the substantial mismatch in outcome domains for both the Clinical and Patient supra-domains between CI users and CI professionals noted in Phase 2, Allen et al [35] recommended inclusion of domains ranked most-highly by CI professionals for the Clinical supra-domain, and by CI users for the Patient supra-domain in an interim, pragmatic COS. This would facilitate an easy transition into a robust, pragmatic, and clinically acceptable COS utilising clinical measures used regularly and trusted by CI programs across Australia and New Zealand.

2. Materials and Methods

Outcome measures that mapped onto the CODS were selected from those identified in Phase 1, as well as measures identified by the research team as commonly used in clinical care in Australia and New Zealand were included in Phase 3 of the study. The list of identified outcome measures was appraised according to their content validity, and their development methodology to determine which outcome measures to include in Phase 3. The final list of included measures, consisting of 43 PROMs and 10 speech perception measures (see Supplementary Tables S1 and S2) was presented to experienced CI clinicians from Australia and New Zealand in a single round online survey. Each clinician was asked to rate the outcome measures for their use of the measure, and if used, the measures’ usefulness, trustworthiness, ease of use and likelihood of future use if it were recommended to them.

2.1. Single Round Online Survey

Individuals who agreed to participate completed a single round online survey utilizing Qualtrics software. The online survey consisted of questions about use of 10 speech perception tests and 43 patient-reported outcome measures (PROMs) using a 4-point categorical scale; (never heard of, never used, occasionally used, regularly used). A short description was provided for each PROM listed. For example; “The Social Participation Restrictions Questionnaire (SParQ) is a hearing-specific, patient-reported outcome measure that was originally developed through consultation with adults with hearing loss, clinicians, and researchers. It has 19 items, each assessed on an 11-point scale. Responses are averaged to form two subscales: Social Behaviours and Social Perceptions”. A copy of the survey is available in the Supplementary Digital Content.
Eligibility criteria for participants were: recent or current CI clinicians from a range of large CI clinics and research institutes and identified by the research team as having extensive knowledge of current CI clinical practices, and/or extensive knowledge of currently available adult-focused CI outcome measures used in Australia and New Zealand. Clinicians included experienced clinicians who participated in Phase 2 [35]. They were invited to participate via an email message.
When participants indicated that they had regularly, or occasionally used a measure, they were asked to provide a rating, using a 5-point Likert scale, ranging from “strongly disagree” to “strongly agree” for each of the following statements;
  • This measure is easy to use in clinical practice (ease of use)
  • This measure gives results that are trustworthy/believable
  • This measure gives results that are useful in clinical practice
  • I would use this measure in clinical practice if it were recommended to me
Respondents were also asked about other clinical measures they used as part of their standard protocol; their usual approach to testing asymmetrical hearing losses; and the factors that they considered when choosing a speech test to ensure that no measures were missed.
Statistical analysis was conducted in Python (v3.11.0) [36]using pandas (v2.2.1) [37], numpy (v1.23.5) [38], and scipy (v1.11.4) [39] and scikit-learn (v1.4.2) [40]. Graphics were generated using matplotlib (v3.8.4) [41]. Descriptive statistics were calculated for demographic variables and survey responses.

2.2. Online Final Recommendation Workshops

Two Final Recommendation workshops, approximately 90 minutes each in duration, were held online through Microsoft Teams videoconferencing software with CI users and CI professionals to finalise key recommendations for the interim pragmatic COS. The outline presentation of the study was provided, one for professionals, and a lay person version for CI users, summarising the key results from the CODS and the current study. A semi-structured interview guide (see final workshop agenda in supplementary material) was developed. Questions included:
What domains should be included in future iterations of the COS?
(CI professionals only) Which outcome measures or subdomains should be recommended as a minimum standard?
How should we prioritise outcome measures within each subdomain?
Participants consisted of two groups: 1) Adult CI users, and 2) CI professionals aged ≥18 years of age with sufficient self-reported English proficiency to participate in the workshop. Adult CI users were required to have at least 6 months experience using a CI. CI professionals were required to have at least 12 months experience providing CI services to adult CI users. Individuals with self-reported disability, other than hearing loss, that precluded full participation in the workshop were excluded.

3. Results

3.1. Single Round Online Survey

Twenty CI clinicians from Australia (n=18) and New Zealand (n=2) participated, 17 of whom responded to both the PROM and the speech perception test familiarity survey questions. Participants’ clinical and research experience is detailed in Table 2. Home and work postcodes for Australian participants were mapped to the Index of Relative Social Advantage and Disadvantage (IRSAD) decile with all Australian participants living in the top 30% of postcodes, and working in the top 50% of postcodes, suggesting that participants skewed toward relative social advantage. For participants from New Zealand, postcodes were mapped using the New Zealand Index of Deprivation for 2023 [42], with one participant living in the top 30% of postcodes bur working in an inner-city location in the bottom 30% of postcodes, and the other living in the bottom 30% of postcodes but working in the top 50%.

3.1.1. Familiarity Ratings (Speech Perception and PROM Measures)

A summary of familiarity ratings is shown in Table 3. Participants were most familiar with speech perception measures. Four speech perception measures (BKB-A, CUNY sentences, CNC words, and AB words) fell within the top five most-used outcome measures and were regularly used by 50% of participants. The DIN/DTT test, BKB-SIN and QuickSIN were occasionally used by > 50% of participants. The remaining three speech perception tests (HINT, Austin and AzBio) had either been “never heard of “ or “never been used” by 59%, 53% and 76% of participants respectively.
Familiarity with PROMs was lower than for speech perception measures. The SSQ (90% of participants) and the SSQ-12 (75% of participants) were the most used PROMs. Participants used the SSQ-12 (70%) slightly more regularly than the SSQ (65%). The Glasgow Hearing Aid Benefit Profile (GHABP), Hearing Handicap for the Elderley (HHIE), Abbreviated Profile for Hearing Aid Benefit (APHAB) were occasionally or regularly used by at least half the participants. No participant regularly used the short or revised version of the HHIE. The Nijmegen Cochlear Implant Questionnaire [55] (NCIQ), and the Cochlear Implant Quality of Life Questionnaire (CIQoL Profile and CIQoL-Global) [64], which have been recommended by the Adult Hearing Standards of Care; Living Guidelines [91] were not commonly used. The NCIQ was only used by 5% of participants regularly, and by 45% occasionally. The CIQoL was used only occasionally by 35% (Global version, 35 items) and 20% (Profile version, 10 items) of participants.

3.1.2. Ease of Use, Trustworthiness, Usefulness and Likely Recommendation to Use Ratings

Ratings were provided for ten speech perception measures (Figure 2), and 21 PROMs (Figure 3) which had been used by ≥3 participants. The CIQoL Profile ratings were also included although they had only been used by 2 participants. Measures with which participants were more familiar were, in general, considered easier to use (τB = 0.383, p < 0.001), more trustworthy (τB = 0.323, p < 0.001), and as providing more useful results (τB = 0.300, p < 0.001). Participants also reported that they would be more likely to use them in practice, if they were recommended (τB = 0.406, p < 0.001) (Figure 4). Ratings were generally high, with very few respondents disagreeing with any of the statements.
The correlation between ratings of ease of use, trustworthiness, clinical usefulness and willingness to use outcome measures was assessed using univariate and bivariate kernal density plots, and correlations between all rating scales were high (see Supplementary Figure S1).

3.1.3. Free Text Responses

Participants suggested several additional PROMS not listed in the survey (Supplementary Table S2) including the Australian Quality of Life Scale (AQoL; n=4), the Strengths and Difficulties Questionnaire (n=3) and the Listening Effort Questionnaire (LEQ-CI; n=3).). The most recommended physiological test was Neural Response/Auditory Response Telemetry (n=3). The Ling Sounds speech sounds identification test was also suggested (n=3).
The most important factors to consider when choosing a speech perception test (see Supplementary Table S3) were: Tests available in the primary language (n=11), and accent (n=7) of the CI user, Cognitive appropriateness (n=4), Speed of delivery (n=2) and measure length (n=2) (see Supplementary Table S3).

3.2. Final Recommendation Workshops

Transcriptions of the final workshops were reviewed and summarised by authors CS and MF, then reviewed by all other authors. Key findings from the workshops are shown in Table 4.
Clinicians felt that future outcome measures should include domains such as cognition, listening effort, listening fatigue, empowerment, social connectedness, relationships, and fatigue. There was a general consensus between both groups that more holistic measures of CI outcomes are needed to provide a more comprehensive understanding of the communication difficulties of CI users and their real-life impact. A combination of different types of measures and understanding their interactions was considered crucial for advancing the field and improving clinical outcomes. However, with the emergence of newly developed measures, interpretation of results may be challenging due to lack of clinician familiarity. Thus, when considering the introduction of new outcome measures, it is vital to consider training as part of implementation, to raise awareness, familiarity and ensure trust in the data obtained with the measure.
CI users felt interaction with the clinician in some aspect of the remote service was important to feel engaged in the process. They recommended that remote services be well-considered and designed to ensure a seamless process for all aspects of the service, from enrolment, validation of enrolment, and login to completion of the remote checks and appointments, payment etc. Whilst benefit was seen in the ability to adjust CI settings remotely, CI users reported it was essential that there was a built-in fail safe or reboot option at the CI user’s end if the service failed midway for some reason.

4. Discussion

Whilst remote care has been shown to be a feasible option for CI service provision, uptake and sustained use of such services has remained low. Research has shown that key barriers to it’s use include concerns about the accuracy and reliability of remote technology methods, limited confidence in the ability of client’s to access and use communication technologies, and the inability to demonstrate that remote services can provide equivalence of service compared to ‘gold standard’ in-person care [12]. In order to compare remote and in-person services effectively, as well as address the concerns about accessibility, usability of services, and accuracy of results, it is essential that the same set of outcomes measures that are sensitive and meaningful to both CI users and CI professionals, are compared across clinics, across modes of service provision, and clinical trials/studies.
While the COS recommended by Andries et al. [22] included CI-specific outcome measures, the measures and domains selected for inclusion were not selected with the input of CI users but rather by a core group of CI experts. While well-known, commonly used instruments and assessment methods were identified, several of the PROMs selected were not designed to current recommended best practice, e.g. using consumer input, considering the risks of bias, or following evidence-based criteria for good psychometric measurement properties [92]. Nor were they CI-specific. Finally, the measures recommended did not specifically consider implementation within a remote care service, and the issues associated with this. The COS [22] included a large number of measures including: PROMs (Work Rehabilitation Questionnaire [WORQ; 59 items], Abbreviated Profile of Hearing Aid Benefit [APHAB; 24 items], Audio Processor Satisfaction Questionnaire [APSQ; 15 items], Speech Spatial and Qualities of Hearing Questionnaire [SSQ-12; 12 items], Hearing Implant Sound Quality Index 19 [HISQUI19; 19 items], and Audiometric measures; Aided Pure tone audiometry, Speech perception [Monosyllabic words in quiet, Sentences in noise], and Sound localisation.
Several of the domains identified as most important by CI users and CI clinicians in the earlier stages of our study [35], were not included in the domains in the above-mentioned COS. Of particular note, device integrity and status, device use and hearing-related quality of life. Understandably remote service domains of reliability, accessibility and ease of use, were not included. In Phase 2, a complete lack of consensus between CI users and CI professionals was observed in the most important domains for self-report “Patient” measures and there was only limited consensus for objective “Clinical” measures. Thus outcomes recommended in the COS by Andreis et al’s [22] may not be important to CI users given the lack of CI user input. Further, our final recommendation workshop revealed a general consensus between CI users and CI clinicians for a minimalist approach to the number of core measures. One must consider the potential for overburdening CI users and clinicians with multiple outcome measures, particularly PROMs with large numbers of items, some of which ask similar questions. A large number of time-consuming outcome measures may result in poor completion compliance, and inefficient or limited clinical use of completed measures. Measures completed or received immediately prior to, or during an appointment, particularly PROMs, may be difficult and time-consuming for the clinician to analyse appropriately during the appointment.
Although the original aim of this study was to develop a COS for CI users utilising remote technology in Australia and New Zealand, this has proved difficult for several reasons:
1. Lack of consensus between CI users and CI professionals on the most important domains for the patient supra-domain [35] mean that implementation of a concise COS is problematic if one is to measure the most important domains within each supra-domain.
2. Lack of well-designed and/or well-validated outcome measures for some of the domains rated as most important to assess. Rigorous development and assessment of novel outcome measures is therefore required.
3. Current clinical practice trends in Australia and New Zealand, observed in our online survey of CI clinicians, indicate that CI services rely predominantly on a relatively small pool of specific speech perception measures as the primary measure of CI outcomes. Clinicians are far less familiar with most PROMs that align with the CODS.
4. Several of the outcome measures identified and explored in the outcomes survey have proprietary test materials, technical requirements, and licencing costs associated with them.
These issues present several problems when considering wide-spread implementation of a COS for remote technology into well-established CI clinics. Large CI clinics often perform retrospective analysis of CI outcomes over time as an important indicator of the success of both individual CI users and CI clinics as a whole [93]. Thus, implementation of a brand-new set of outcome measures must consider the impact on the ability to compare outcomes over time. It may be necessary to align the outcomes of new measures with pre-existing measures for a period of time, in order to retain the ability to compare outcomes over time. Significant support, resources and training around measures that are new to the field, or simply just new to the clinic will be required.
Clinicians must have confidence and trust in the recommended outcome measures, both in the methods of data collection and interpretation of results, as evidence by the high levels of correlation between the ease of use, trustworthiness, usefulness, and recommendation ratings provided by participants. Familiarity with an outcome measure engenders an understanding of effective methods of use and interpretation, and in turn outcome measures that are inherently easier to apply and interpret are more likely to become part of existing clinical practice [94,95,96]. Measures that provide trustworthy results are also more likely to be considered useful in clinical practice. Understanding which of these four factors of clinician experience, if any, are primarily responsible for positive clinician experience and uptake is essential to support implantation efforts. This is particularly salient given our findings in relation to some more commonly used surveys such as the CIQoL and the NCIQ, both of which were recommended outcome measures in the recently drafted Adult Hearing Standards of Care; Living Guidelines [91]. Our study revealed poor ratings for ease of use, likelihood to use if recommended and, to a lesser extent usefulness for the NCIQ which would indicate that it is unlikely that this measure would be readily adopted into CI clinical practice in Australia or New Zealand. Furthermore, the NCIQ contains 60 items and so fails to meet the recommendations of our CI users about shorter, more concise measures. The CIQoL, whilst receiving relatively good ratings for all four categories, had only been used by a maximum of four clinicians, thus training for its implementation is required.
In light of these considerations, it appears most appropriate to recommend a pragmatic, interim COS for remote technologies for CI users in order to facilitate uptake into current clinical practice with the recognition that CI outcomes are constantly evolving [91], and as such, so are the important outcome domains, and outcome measures with which to assess them are also evolving. Furthermore, it was decided to limit measures to those commonly used in English-speaking countries in the first instance, as Australia and New Zealand were the focus of this study, to further facilitate compliance with use of the COS as commonly used outcome measures differ substantially across countries. There was a strong focus in the workshops on the need to ensure that implementation of any new set of outcome measures did not overburden either CI users or CI clinicians. Whilst there was a push to utilise more meaningful, “real-life” outcome measures, this was not to be at the expense of additional time and effort for key stakeholders. In fact, the preference was for a reduction in time allocated towards assessment of outcomes. Similar findings have been noted in other allied health fields [97]. However, a reduction in the length of speech test lists, or the number of questions in surveys should not be at the expense of a reduction in their psychometric properties, such as test-retest reliability and validity. Any recommended outcome measures must have, and retain, good psychometric properties to ensure their usefulness.

Recommended Interim, Pragmatic COS

4.1.1. Service Outcomes

A single Likert-item for each of reliability, usability and acceptability, with an option for free text.
Wording needs to be further defined but a regularly mentioned example was a 5-response option based on agreement (e.g. the remote technology was reliable: strongly agree to strongly disagree)

4.1.2. Clinically-Measured Outcomes

CNC (or similar CVC) and optional digit triplet testing (DTT), testing device integrity/system check, device use, and adverse events
Recommended included outcome measures are:
An adaptive speech test that is presented in noise but could also be completed in quiet depending upon the CI user’s speech perception ability. Of the four most regularly used tests identified by CI professionals (BKB-A, CUNY, CNC and AB), the CNC test was included as this is a current requirement for CI candidacy determination.
The DTT is a speech in noise test that is often delivered remotely, thus having the appropriate underpinning architecture for delivery via remote technology systems. Other advantages of this test include that it is often delivered adaptively, the digit material is easily translatable into other languages with easily understandable stimuli, it can be delivered without the need for calibration equipment, and it is widely used across the world.
Device integrity and status, device use, and adverse events were the other three most highly rated clinical tests in addition to speech perception testing. Both groups felt it was vital to ensure that both internal and external components of the CI were functioning appropriately (device integrity) to ensure that any outcome measures additional to these measures are not impacted by device malfunction. Device use, via datalogging, has been included in the COS to ensure that limited CI outcomes are not the result of limited CI use. We acknowledge, however, that there are known discrepancies between reported use and logged data, and that this discrepancy could be due to either technology errors or the user’s decision not to report limited use.
Any potential worry about device usage should be discussed in a supportive and caring manner with the CI user.
Measures excluded include:
The BKB-A test due to fixed level presentation of sentences, the fact that it was originally developed for a low (kindergarten age) literacy level, so it is somewhat child-like, resultant ceiling effects, and its typical presentation mode in quiet.
The CUNY sentence test due to ceiling effects, and the relatively high language knowledge/literacy level required, in addition to the potential influence of auditory memory on outcomes.
The AB word test due to the limited number of lists available, which could lead to practice effects, and because the Australian version of the test materials is no longer available for purchase from the National Acoustic Laboratories.
Speech sound identification/discrimination assessment (e.g. the LING test), whilst rated highly by CI users, was perceived by CI professionals as a more diagnostic measure to indicate specific hearing difficulties, rather than an overall measure of CI outcome, thus was not included in the current interim COS.

4.1.3. Patient Reported Outcome Measures

Given the discrepancy in domain importance between CI users and CI professionals in this supra-domain, preference was given to CI users based on feedback provided in Phase 2, in which it was suggested that CI users’ everyday life experiences should be prioritised.
Recommended included outcome measures are:
The SSQ or short-form SSQ-12 as an interim PROM because it is the most regularly used, easy to use, trustworthy, and most likely to be used if recommended whilst other PROMs gain acceptance and are used. CI professionals also felt that in the context of remote services, given the potential impacts associated with variability in the home test environment at each test point, there was a necessity to have a reported measure of hearing ability to confirm the behavioural test measure. The SSQ-12 has been recommended in the ANZ adaption of the Adult Standards of Hearing Care; Living Guidelines ANZ adaption [98]. However, it should be noted that it does not address the most important domains, and although it is commonly used, this alone is not an appropriate criterion for its long-term inclusion in future COSs.
Hearing-related quality of life, satisfaction, and wellbeing are domains for consideration. Based on these domains, the CIQOL (Cochlear Implant Quality of Life) [31] would be a suitable PROM, augmented by a satisfaction measure. The CIQOL is a well-validated measures developed with stakeholder input, using modern psychometric Item Response Theory analysis. The CIQoL Profile (35 items) has sub-domains hearing, communication, social relationships, emotional well-being, independence and daily life, device satisfaction and use, cognitive and mental engagement and perception of self and identify. It is not a unidimensional measure (i.e. quality of life) but the authors suggest the broader sub-domains reflect quality of life. Alternatively, the Living with Cochlear Implants (LivCI) [99,100,101] , a recently developed CI-specific, 22-item PROM which includes four sub-domains addressing psychosocial and wellbeing, participation (i.e. HRQoL), aesthetics and visibility (a primary driver of satisfaction), and stigma could be considered. Like the CIQoL, the LivCI has been developed according to COSMIN best practice principles, including extensive stakeholder (e.g. CI professionals and CI users), content evaluation, contemporary Rasch analysis to ensure high-quality items that are independent, alongside Classical Test Theory analyses. Either of these measures could be a potential candidate to replace the SSQ in future. Both measures are recommended for use in the ANZ adaption of the Adult Standards of Hearing Care; Living Guidelines ANZ adaption [98].
In the absence of an appropriate PROM for CI user satisfaction of devices, it was suggested that, as for Service measures, a Likert single-item measure could be used in the interim.

4.1.4. Future Emerging Domains

Other domains that are emerging as important for remote technologies within audiology [21], but not widely considered in the CI field, such as empowerment, listening effort, auditory fatigue, should also be considered for a future COS. There are a number well-developed CI- or hearing-specific PROMs which address such domains. Additionally, an assessment of digital literacy, whilst not an outcome measure per se, prior to CI users using remote technologies should also be considered [102]. It is strongly recommended that ongoing monitoring of the clinical practices and opinions of CI clinicians is carried out to ensure advancement of clinical practice. A part of this process would be to update the interim COS recommended here over time, as well as to guide development of policy and ongoing implementation, training and de-implementation within clinical practice [103].

5. Conclusions

Development of a core outcome set (COS) to assess remote technologies used by CI users is vital given the increase in the use of remote technologies for CI care. It is important that such a COS is relevant across both remote and in-clinic services to enable comparison and seamless integration of the two modes of service. It must also incorporate meaningful, useful outcome measures for CI users, their families and CI clinicians alike using well-designed, trusted measures, that can be incorporated into clinical practice without unnecessarily over-burdening staff, CI users and their families. We present a pragmatic, interim COS for use in hybrid clinical practice, noting that ongoing monitoring of meaningful future outcomes and clinical practices may result in adaptions to the recommended COS in the future.

Supplementary Materials

The following supporting information can be downloaded at the website of this paper posted on Preprints.org, Figure S1: Bivariate kernel density plots of the correlations between outcome measure ratings.; Table S1: PROM description and development information; Table S2: Additional measures using in clinical practice in Australia and New Zealand; Table S3: Factors identified as important to consider when choosing a speech test.

Author Contributions

C.S.: Project conception and design, administrative support, collection and assembly of data, data interpretation, manuscript writing and review, Funding acquisition; D.A.: Project design, administrative support, collection and assembly of data, Data analysis and interpretation, Manuscript review; E.L.: Project design, administrative support, Data interpretation, Manuscript review; I.B.: Project design, administrative support, Data interpretation, Manuscript review; M.F.: Project conception and design, data interpretation, substantial manuscript review, Funding acquisition, Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by an Investigator Initiated Grant from Cochlear Ltd.

Institutional Review Board Statement

Ethical approval was obtained from. The study was conducted in accordance with the Declaration of Helsinki, and approved by the Human Ethics Office of Research at the University of Western Australia (Re: 2022/ET001007, date of approval 09/01/2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the participant(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy and ethical restrictions enforced on research relating to human participants by the University of Western Australia.

Acknowledgments

We would like to thank Helen Cullington and Rene Gifford for their valuable insights, and support of this study.

Conflicts of Interest

The employing organization of MF received funding from Cochlear Ltd for this study. The remaining authors declare no conflicts of interest.” The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

CI Cochlear Implant
CODS Core Outcome Domain Set
COS Core Outcome Set
PROM Patient Reported Outcome Measure
WHO World Health Organisation
ICF International classification of Functioning, Disability and Health framework
COMET Core Outcome Measures in Effectiveness Trials

References

  1. Ferguson, M.A., et al., Remote Technologies to Enhance Service Delivery for Adults: Clinical Research Perspectives. Semin Hear, 2023. 44(3): p. 328-350.
  2. Kim, J., et al., A Review of Contemporary Teleaudiology: Literature Review, Technology, and Considerations for Practicing. J Audiol Otol, 2021. 25(1): p. 1-7.
  3. Luryi, A.L., et al., Cochlear Implant Mapping Through Telemedicine-A Feasibility Study. Otol Neurotol, 2020. 41(3): p. e330-e333.
  4. Maruthurkkara, S., et al., Remote check test battery for cochlear implant recipients: proof of concept study. Int J Audiol, 2022. 61(6): p. 443-452.
  5. Schepers, K., et al., Remote programming of cochlear implants in users of all ages. Acta Otolaryngol, 2019. 139(3): p. 251-257.
  6. Cullington, H., et al., Feasibility of personalised remote long-term follow-up of people with cochlear implants: a randomised controlled trial. BMJ Open, 2018. 8(4): p. e019640.
  7. Maruthurkkara, S., S. Case, and R. Rottier, Evaluation of Remote Check: A Clinical Tool for Asynchronous Monitoring and Triage of Cochlear Implant Recipients. Ear Hear, 2022. 43(2): p. 495-506.
  8. Philips, B., et al., Empowering Senior Cochlear Implant Users at Home via a Tablet Computer Application. Am J Audiol, 2018. 27(3S): p. 417-430.
  9. Carner, M., et al., Personal experience with the remote check telehealth in cochlear implant users: from COVID-19 emergency to routine service. Eur Arch Otorhinolaryngol, 2023. 280(12): p. 5293-5298.
  10. Nassiri, A.M., et al., Implementation Strategy for Highly-Coordinated Cochlear Implant Care With Remote Programming: The Complete Cochlear Implant Care Model. Otol Neurotol, 2022. 43(8): p. e916-e923.
  11. Nittari, G., et al., Telemedicine in the COVID-19 Era: A Narrative Review Based on Current Evidence. Int J Environ Res Public Health, 2022. 19(9).
  12. Chong-White, N., et al., Exploring teleaudiology adoption, perceptions and challenges among audiologists before and during the COVID-19 pandemic. BMC Digital Health, 2023. 1(24).
  13. Lilies, A., et al., Independent Evaluation of CHOICE 2021, P. Darnton, Editor. 2021, The Health Foundation: Wessex Academic Health Science Network, Innovation Centre, 2 Venture Road, Southampton Science Park, SO16 7NP. p. 1-53.
  14. Sucher, C., et al., Patient preferences for Remote cochlear implant management: A discrete choice experiment. PLoS One, 2025. 20(6): p. e0320421.
  15. Barreira-Nielsen, C.S.C. and L.S. Campos, Implementation of the hybrid teleaudiology model: acceptance, feasibility and satisfaction in a cochlear implant program. Audiol., Commun. Res, 2022. 27.
  16. Department of Health, V., Virtual Care Operational Framework, V. Department of Health, Editor. 2023.
  17. Granberg, S., et al., The ICF Core Sets for hearing loss-researcher perspective. Part I: Systematic review of outcome measures identified in audiological research. International Journal of Audiology, 2014. 53(2): p. 65-76.
  18. Akeroyd, M.A., et al., A comprehensive survey of hearing questionnaires: how many are there, what do they measure, and how have they been validated? Trials, 2015. 16(1): p. 1-1.
  19. Neal, K., et al., Listening-based communication ability in adults with hearing loss: A scoping review of existing measures. Frontiers in Psychology, 2022. 13: p. 786347.
  20. Danermark, B., et al., The creation of a comprehensive and a brief core set for hearing loss using the international classification of functioning, disability and health. Am J Audiol, 2013. 22(2): p. 323-8.
  21. Allen, D., L. Hickson, and M. Ferguson, Defining a Patient-Centred Core Outcome Domain Set for the Assessment of Hearing Rehabilitation With Clients and Professionals. Front Neurosci, 2022. 16: p. 787607.
  22. Andries, E., et al., Implementation of the international classification of functioning, disability and health model in cochlear implant recipients: a multi-center prospective follow-up cohort study. . Front Audiol Otol, 2023. 1: p. 1257504.
  23. Boisvert, I., et al., Editorial: Outcome Measures to Assess the Benefit of Interventions for Adults With Hearing Loss: From Research to Clinical Application. Front Neurosci, 2022. 16: p. 955189.
  24. Clarke, M. and P.R. Williamson, Core outcome sets and systematic reviews. Systematic reviews, 2016. 5(1): p. 1-4.
  25. COMET. Core Outcome Measures in Effectiveness Trials. 2022; Available from: https://www.comet-initiative.org/.
  26. Allen, D., L. Hickson, and M. Ferguson, Defining a patient-centred core outcome domain set for the assessment of hearing rehabilitation with clients and professionals. Frontiers in Neuroscience, 2022. 16.
  27. Hall, D.A., et al., Toward a global consensus on outcome measures for clinical trials in tinnitus: report from the first international meeting of the COMiT Initiative, November 14, 2014, Amsterdam, The Netherlands. Trends in Hearing, 2015. 19: p. 2331216515580272.
  28. Dietz, A., et al., The effectiveness of cochlear implantation on performance-based and patient-reported outcome measures in Finnish recipients. Frontiers in Neuroscience, 2022.
  29. Laird, E., et al., Systematic review of patient and service outcome measures of remote digital technologies for cochlear implant and hearing aid users. Frontiers in Audiology and Otology, 2024. 2.
  30. Hughes, S.E., et al., Rasch analysis of the listening effort Questionnaire—Cochlear implant. Ear and Hearing, 2021. 42(6): p. 1699-1711.
  31. McRackan, T.R., et al., Cochlear Implant Quality of Life (CIQOL): development of a profile instrument (CIQOL-35 Profile) and a global measure (CIQOL-10 Global). Journal of Speech, Language, and Hearing Research, 2019. 62(9): p. 3554-3563.
  32. Maidment, D., et al., Evaluating a theoretically informed and co-created mHealth educational intervention for first-time hearing aid users: a qualitative interview study. Journal of Medical Internet Research, 2020. 8(8): p. e17193.
  33. Gomez, R., et al., Smartphone-Connected Hearing Aids Enable and Empower Self-Management of Hearing Loss: A Qualitative Interview Study Underpinned by the Behavior Change Wheel. Ear and Hearing, 2022. 43(3): p. 921-932.
  34. Sucher, C., et al., Patient preferences for remote cochlear implant management: A discrete choice experiment. PLoS One, in review.
  35. Allen, D., et al., Developing a Core Outcome Domain Set for Remote Cochlear Implant Management. Submitted.
  36. Foundation., P.S., The Python Language Reference.
  37. McKinney, W. Data Structures for Statistical Computing in Python. in Proceedings of the 9th Python in Science Conference. 2010.
  38. Harris, C.R. and et al., Array Programming with NumPy. Nature, 2020. 585(7825): p. 357-362.
  39. Virtanen, P. and et al., SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nature Methods, 2020. 17(3): p. 261-272.
  40. F., P. and et al., Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research, 2011. 12: p. 2825-2830.
  41. Team., T.M.D., Matplotlib: Visualisation with Python, Zenodo, Editor. 2025.
  42. Atkinson, J., et al., NZDep2023 Index of Socioeconomic Deprivation: Research Report. 2024: Wellington.
  43. Gatehouse, S. and W. Noble, The Speech, Spatial and Qualities of Hearing Scale (SSQ). Int J Audiol, 2004. 43(2): p. 85-99.
  44. Bench, J. and J. Doyle, The BKB/A (Banford-Kowal-Bench/Australian version) sentence lists for hearing-impaired children. . 1979, La Trobe University: Victoria, Australia.
  45. Boothroyd A, H.L., & Hnath T., A sentence test of speech perception: reliability, set equivalence, and short term learning. CUNY Academic Works, 1985. 1985.
  46. Peterson, G.E. and I. Lehiste, Revised CNC lists for auditory tests. J Speech Hear Disord, 1962. 27: p. 62-70.
  47. Boothroyd, A., Developments in Speech Audiometry. Sound, 1968. 2: p. 3-10.
  48. Noble, W., et al., A short form of the Speech, Spatial and Qualities of Hearing scale suitable for clinical use: the SSQ12. Int J Audiol, 2013. 52(6): p. 409-12.
  49. Ventry, I.M. and B.E. Weinstein, The hearing handicap inventory for the elderly: a new tool. Ear Hear, 1982. 3(3): p. 128-34.
  50. Killion, M.C., et al., Development of a quick speech-in-noise test for measuring signal-to-noise ratio loss in normal-hearing and hearing-impaired listeners. J Acoust Soc Am, 2004. 116(4 Pt 1): p. 2395-405.
  51. Etymotic Research, Etymotic BKB-SIN Speech-in-Noise Test User Manual. 2005. p. 1-27.
  52. Smits, C., S. Theo Goverts, and J.M. Festen, The digits-in-noise test: assessing auditory speech recognition abilities in noise. J Acoust Soc Am, 2013. 133(3): p. 1693-706.
  53. Gatehouse, S., A self-report outcome measure for the evaluation of hearing aid fittings and services. Health Bull (Edinb), 1999. 57(6): p. 424-36.
  54. Cox, R.M. and G.C. Alexander, The abbreviated profile of hearing aid benefit. Ear Hear, 1995. 16(2): p. 176-86.
  55. Hinderink, J.B., P.F. Krabbe, and P. Van Den Broek, Development and application of a health-related quality-of-life instrument for adults with cochlear implants: the Nijmegen cochlear implant questionnaire. Otolaryngol Head Neck Surg, 2000. 123(6): p. 756-65.
  56. King, N., et al., A new comprehensive cochlear implant questionnaire for measuring quality of life after sequential bilateral cochlear implantation. Otol Neurotol, 2014. 35(3): p. 407-13.
  57. Spitzer, R.L., et al., A brief measure for assessing generalized anxiety disorder: the GAD-7. Arch Intern Med, 2006. 166(10): p. 1092-7.
  58. Nilsson, M., S.D. Soli, and J.A. Sullivan, Development of the Hearing in Noise Test for the measurement of speech reception thresholds in quiet and in noise. J Acoust Soc Am, 1994. 95(2): p. 1085-99.
  59. Cassarly, C., et al., The Revised Hearing Handicap Inventory and Screening Tool Based on Psychometric Reevaluation of the Hearing Handicap Inventories for the Elderly and Adults. Ear Hear, 2020. 41(1): p. 95-105.
  60. Dawson, P.W., A.A. Hersbach, and B.A. Swanson, An adaptive Australian Sentence Test in Noise (AuSTIN). Ear Hear, 2013. 34(5): p. 592-600.
  61. Spahr, A.J. and M.F. Dorman, Performance of subjects fit with the Advanced Bionics CII and Nucleus 3G cochlear implant devices. Arch Otolaryngol Head Neck Surg, 2004. 130(5): p. 624-8.
  62. Cox, R., et al., Optimal outcome measures, research priorities, and international cooperation. Ear Hear, 2000. 21(4 Suppl): p. 106S-115S.
  63. Dillon, H., G. Birtles, and R. Lovegrove, Measuring the Outcomes of a National Rehabilitation Program: Normative Data for the Client Oriented Scale of Improvement (COSI) and the Hearing Aid User’s Questionnaire (HAUQ). J Am Acad Audiol 1999, 1999. 10(02): p. 67-79.
  64. McRackan, T.R., et al., Cochlear Implant Quality of Life (CIQOL): Development of a Profile Instrument (CIQOL-35 Profile) and a Global Measure (CIQOL-10 Global). J Speech Lang Hear Res, 2019. 62(9): p. 3554-3563.
  65. Amann, E. and I. Anderson, Development and validation of a questionnaire for hearing implant users to self-assess their auditory abilities in everyday communication situations: the Hearing Implant Sound Quality Index (HISQUI19). Acta Otolaryngol, 2014. 134(9): p. 915-23.
  66. The ida Institute. ida Institute Motivational Tools: The Line. [cited 2025 17/03/2024].
  67. Hawthorne, G. and A. Hogan, Measuring disability-specific patient benefit in cochlear implant programs: developing a short form of the Glasgow Health Status Inventory, the Hearing Participation Scale. Int J Audiol, 2002. 41(8): p. 535-44.
  68. McBride, W.S., et al., Methods for screening for hearing loss in older adults. Am J Med Sci, 1994. 307(1): p. 40-2.
  69. Yesavage, J.A., et al., Development and validation of a geriatric depression screening scale: a preliminary report. J Psychiatr Res, 1982. 17(1): p. 37-49.
  70. Luetje, C.M., et al., Phase III clinical trial results with the Vibrant Soundbridge implantable middle ear hearing device: a prospective controlled multicenter study. Otolaryngol Head Neck Surg, 2002. 126(2): p. 97-107.
  71. Beck, A.T., et al., An inventory for measuring depression. Arch Gen Psychiatry, 1961. 4: p. 561-71.
  72. Antony, M.M., et al., Psychometric properties of the 42-item and 21-item versions of the Depression Anxiety Stress Scales in clinical groups and a community sample. Psychological Assessment, 1998. 10(2): p. 176–181.
  73. Zigmond, A.S. and R.P. Snaith, The hospital anxiety and depression scale. Acta Psychiatr Scand, 1983. 67(6): p. 361-70.
  74. Kompis, M., et al., Factors influencing the decision for Baha in unilateral deafness: the Bern benefit in single-sided deafness questionnaire. Adv Otorhinolaryngol, 2011. 71: p. 103-111.
  75. Topp, C.W., et al., The WHO-5 Well-Being Index: a systematic review of the literature. Psychother Psychosom, 2015. 84(3): p. 167-76.
  76. Cox, R.M. and G.C. Alexander, Expectations about hearing aids and their relationship to fitting outcome. Journal of the American Academy of Audiology, 2000. 11: p. 368-382.
  77. Billinger-Finke, M., et al., Development and validation of the audio processor satisfaction questionnaire (APSQ) for hearing implant users. Int J Audiol, 2020. 59(5): p. 392-397.
  78. De Jong Gierveld, J. and F. Kamphuls, The Development of a Rasch-Type Loneliness Scale. Applied Psychological Measurement, 1985. 9(3): p. 289-299.
  79. De Jong Gierveld, J. and T. Van Tilburg, A 6-Item Scale forOverall, Emotional, and Social Loneliness. Confirmatory Tests on Survey Data. Research on Aging, 2006. 28: p. 582-598.
  80. Terluin, B., et al., The Four-Dimensional Symptom Questionnaire (4DSQ): a validation study of a multidimensional self-report questionnaire to assess distress, depression, anxiety and somatization. BMC Psychiatry, 2006. 6: p. 34.
  81. Diener, E., et al., The Satisfaction with Life Scale. Journal of Personality Assessment, 1985. 49: p. 71-75.
  82. Russell, D., L.A. Peplau, and C.E. Cutrona, The revised UCLA Loneliness Scale: Concurrent and discriminant validity evidence. Journal of Personality and Social Psychology, 1980. 39: p. 472-480.
  83. Davies, A.R. and J.E. Ware Jr, GHAA’s Consumer Satisfaction Survey and User’s Manual. 2nd ed, ed. G.H.A.o. America. 1991, Washington DC.
  84. McConnaughy, E.A., J.O. Prochaska, and W.F. Velicer, Stages of change in psychotherapy: Measurement and sample profiles. Psychotherapy: Theory, Research & Practice, 1983. 20(3): p. 368–375.
  85. Levenstein, S., et al., Development of the Perceived Stress Questionnaire: a new tool for psychosomatic research. J Psychosom Res, 1993. 37(1): p. 19-32.
  86. Heffernan, E., N.S. Coulson, and M.A. Ferguson, Development of the Social Participation Restrictions Questionnaire (SPaRQ) through consultation with adults with hearing loss, researchers, and clinicians: a content evaluation study. Int J Audiol, 2018. 57(10): p. 791-799.
  87. Sansoni, J., et al., Technical Manual and Instructions for the Revised Incontinence and Patient Satisfaction Tools, ed. C.f.H.S. Development. 2011, Wollongong, NSW: University of Wollongong.
  88. Cox, R.M. and G.C. Alexander, Measuring Satisfaction with Amplification in Daily Life: the SADL scale. Ear Hear, 1999. 20(4): p. 306-20.
  89. Reichheld, F.F., The one number you need to grow, in Harvard Business Review. 2003, Harvard Business School Publishing: Massachusetts, US.
  90. Heffernan, E., A. Habib, and M. Ferguson, Evaluation of the psychometric properties of the social isolation measure (SIM) in adults with hearing loss. Int J Audiol, 2019. 58(1): p. 45-52.
  91. CI Task Force. Adult Hearing Standards of Care; Living Guidelines. 2022 [cited 2025 24/03/2025]; Available from: https://adulthearing.com/standards-of-care/.
  92. Mokkink, L.B., E.B.M. Elsman, and C.B. Terwee, COSMIN guideline for systematic reviews of patient-reported outcome measures version 2.0. Qual Life Res, 2024. 33(11): p. 2929-2939.
  93. Boisvert, I., et al., Cochlear implantation outcomes in adults: A scoping review. PLoS One, 2020. 15(5).
  94. Duncan, E.A. and J. Murray, The barriers and facilitators to routine outcome measurement by allied health professionals in practice: a systematic review. BMC Health Serv Res, 2012. 12: p. 96.
  95. Hatfield, D.R. and B.M. Ogles, Why some clinicians use outcome measures and others do not. Adm Policy Ment Health, 2007. 34(3): p. 283-91.
  96. O'Connor, B., et al., Understanding allied health practitioners' use of evidence-based assessments for children with cerebral palsy: a mixed methods study. Disabil Rehabil, 2019. 41(1): p. 53-65.
  97. Aiyegbusi, O.L., et al., Recommendations to address respondent burden associated with patient-reported outcome assessment. Nat Med, 2024. 30(3): p. 650-659.
  98. Collaborative., A.H.H. ANZ Hearing Health Collaborative (ANZ HHC): Living Guidelines for Cochlear Implant (CI) Referral, CI Evaluation and Candidacy, and CI Outcome Evaluation in Adults 2025 [cited 2025 10/08/2025]; Available from: https://hhc.anz.adulthearing.com/hhc-2/projects/#follow.
  99. Sucher, C., et al., Development of the LivCI: a patient-reported outcome measure of personal factors that affect quality of life, use and acceptance of cochlear implants. , in World Congress of Audiology. 2024: Paris.
  100. Laird, E., C. Sucher, and M. Ferguson, Development of a self-report measure of Living with Cochlear Implants (LivCI): A content evaluation. International Journal of Audiology, submitted.
  101. Hughes, S., et al., Living with Cochlear Implants (LivCI): Development and validation of a new patient-reported outcome measure (PROM) of personal factors associated with living with cochlear implants (LivCI). submitted.
  102. Ferguson, M., T. Sahota, and C. Sucher, “Don’t assume I’m too old!”: Assessment of Digital Literacy in a Clinical Sample of Adults with Hearing Loss. Submitted.
  103. Prusaczyk, B., T. Swindle, and G. Curran, Defining and conceptualizing outcomes for de-implementation: key distinctions from implementation outcomes. Implement Sci Commun, 2020. 1: p. 43.
Figure 1. The three phases of the study.
Figure 1. The three phases of the study.
Preprints 175114 g001
Figure 2. Speech perception measures. Ratings for Ease of Use (E), Trustworthiness (T), Usefulness (U), Likelihood of future recommendation for use (R). Number of participants who had used each measure, either occasionally or regularly, and thus provided ratings, are shown in parentheses after the named outcome measure. For abbreviations, see Table 2.
Figure 2. Speech perception measures. Ratings for Ease of Use (E), Trustworthiness (T), Usefulness (U), Likelihood of future recommendation for use (R). Number of participants who had used each measure, either occasionally or regularly, and thus provided ratings, are shown in parentheses after the named outcome measure. For abbreviations, see Table 2.
Preprints 175114 g002
Figure 3. PROM Ratings for Ease of Use (E), Trustworthiness (T), Usefulness (U), Likelihood of future recommendation for use (R). Number of participants who had used each measure, either occasionally or regularly, and thus provided ratings, are shown in parentheses after the named outcome measure.
Figure 3. PROM Ratings for Ease of Use (E), Trustworthiness (T), Usefulness (U), Likelihood of future recommendation for use (R). Number of participants who had used each measure, either occasionally or regularly, and thus provided ratings, are shown in parentheses after the named outcome measure.
Preprints 175114 g003
Table 1. Core outcome domains identified in Phase 2 [35]. *Equal rating of 2nd for both domains by CI users. HL: hearing loss; CI: cochlear implant.
Table 1. Core outcome domains identified in Phase 2 [35]. *Equal rating of 2nd for both domains by CI users. HL: hearing loss; CI: cochlear implant.
SUPRADOMAIN
SERVICE CLINICAL PATIENT
Domain
Priority
CI
Users
CI Professionals CI
Users
CI Professionals CI
Users
CI Professionals
1st Reliability
of remote technology
Usability
of remote technology
Speech recognition
in noise
Device
integrity & status
Participation restriction due to HL Expectations of hearing health outcomes
2nd Usability
of the remote technology
Accessibility
of the remote service (for CI user)
Speech recognition
in quiet
Speech discrimination Hearing Related Quality of Life

AND*

Satisfaction with CI
Motivation & Readiness to Act
on hearing difficulties
3rd Accessibility
of the remote service (for CI user)
Reliability
of remote technology
Speech discrimination Device Use Mental Health & Wellbeing Acceptability & Tolerability
of the CI
(for CI user)
Table 2. Clinical and Research Experience of CI clinicians who participated in the survey.
Table 2. Clinical and Research Experience of CI clinicians who participated in the survey.
Number of Participants (%) Median (years) Range
(years)
Duration of Clinical Audiology Experience 20 (100%) 20.0 7-41
Duration of CI specific clinical Audiology Experience 20 (100%) 19.0 4-40
Experience in Audiology-focused research 12 (60%) 7.5 0-40
Experience in CI-specific research 14 (70%) 12 0-40
Table 3. Familiarity ratings for PROMs and clinical measures, ordered by median response. PROMs are shaded in light grey, speech perception tests are unshaded. Numbers under ratings represent the number of participants providing each rating for each measure.
Table 3. Familiarity ratings for PROMs and clinical measures, ordered by median response. PROMs are shaded in light grey, speech perception tests are unshaded. Numbers under ratings represent the number of participants providing each rating for each measure.
Clinical
Measure/PROM
Never Heard Never Used Occasionally Used Regularly Used Median Response
Speech and Spatial Qualities Scale (SSQ) [43] 0 2 5 13 Regularly Used
Bamford-Kowal-Bench Sentence Test, Australian Version (BKB/A) [44] 0 1 6 10 Regularly Used
City University of New York Sentence Test (CUNY©) [45] 0 2 2 13 Regularly Used
Consonant-Nucleus-Consonant Words (CNC Words) - [46] 0 1 0 16 Regularly Used
Arthur Boothroyd Words (AB Words) [47] 0 1 1 15 Regularly Used
Short Form Speech and Spatial Qualities Scale (SSQ-12) [48] 4 1 1 14 Regularly Used
Hearing Handicap Inventory for the Elderly (HHIE) [49] 2 6 10 2 Occasionally Used
Quick Speech In Noise Test (QuickSIN™) [50] 2 4 9 2 Occasionally Used
Bamford-Kowal-Bench Sentences In Noise Test (BKB-SIN™) [51] 1 5 4 7 Occasionally Used
Digits-In-Noise/Digit Triplet Test (DIN/DTT) [52] 2 3 7 5 Occasionally Used
Glasgow Hearing Aid Benefit Profile (GHABP) [53] 0 9 8 3 Occasionally Used
Abbreviated Profile of Hearing Aid Benefit (APHAB) [54] 2 7 4 7 Occasionally Used
Nijmegen Cochlear Implant Questionnaire (NCIQ) [55] 4 6 9 1 Never Used
Comprehensive Cochlear Implant Questionnaire (CCIQ) [56] 8 5 7 0 Never Used
General Anxiety Disorder-7 (GAD-7) [57] 7 13 0 0 Never Used
Hearing In Noise Test (HINT) [58] 2 8 6 1 Never Used
Revised Hearing Handicap for the Elderly (RHHI) [59] 7 12 1 0 Never Used
Revised Hearing Handicap for the Elderly - Screening (RHHI-S) [59] 8 11 1 0 Never Used
Austin Sentence Test (Austin) [60] 3 6 4 4 Never Used
AzBio Sentence Test (AzBio) [61] 1 12 3 1 Never Used
International Outcomes Inventory - Cochlear Implants (IOI-CI) [62] 2 11 4 3 Never Used
Hearing Aid Users Questionnaire (HAUQ) [63] 9 8 2 1 Never Used
Cochlear Implant Quality of Life Questionnaire - Global (CIQoL-Global) [64] 9 4 7 0 Never Used
Hearing Implant Sound Quality Index (HISQUI19) [65] 9 7 3 1 Never Used
IDA Tool - The Line (The Line) [66] 8 8 2 2 Never Used
Hearing Participation Scale (HPS) [67] 8 11 1 0 Never Used
Hearing Handicap Inventory for the Elderly - Screening (HHIE-S) [68] 4 9 5 2 Never Used
Geriatric Depression Scale - Long (GDS-L) [69] 9 11 0 0 Never Used
Cochlear Implant Quality of Life Questionnaire - Profile (CIQoL-Profile) [64] 9 7 4 0 Never Used
Hearing Device Satisfaction Scale (HDSS) [70] 8 11 1 0 Never Used
Beck's Depression Index (BDI) [71] 9 10 1 0 Never Used
Depression Anxiety Stress Scale (21 Item) (DASS-21) [72] 8 10 2 0 Never Used
Depression Anxiety Stress Scale (42 Item) (DASS-42) [72] 7 11 2 0 Never Used
Hospital Anxiety and Depression Scale (HADS) [73] 9 5 5 1 Never Used
Bern Benefit in Single-Sided Deafness (BBSS) [74] 10 9 1 0 Never Heard
WHO Well-being Index (WHO-S) [75] 10 10 0 0 Never Heard
Expected Consequences of Hearing Aid Ownership (ECHO) [76] 13 7 0 0 Never Heard
Audio Processor Satisfaction Questionnaire (APSQ) [77] 12 7 1 0 Never Heard
De Jong Gierveld Loneliness scale (11 Item) (DJGLS-11) [78] 15 5 0 0 Never Heard
De Jong Gierveld Loneliness scale (6 Item) (DJGLS-6) [79] 15 5 0 0 Never Heard
The Four Dimensional Symptom Questionnaire (4DSQ) [80] 16 4 0 0 Never Heard
Satisfaction With Life Scale (SWLS) [81] 16 4 0 0 Never Heard
UCLA Loneliness Index (Revised) (UCLA) [82] 16 4 0 0 Never Heard
Visit-Specific Satisfaction Questionnaire (VSQ-9) [83] 19 1 0 0 Never Heard
University of Rhode Island Change Assessment adapted for hearing loss (URICA-HL) [84] 13 7 0 0 Never Heard
Perceived Stress Questionnaire (PSQ) [85] 12 8 0 0 Never Heard
Social Participation Restrictions Questionnaire (SPaRQ) [86] 12 8 0 0 Never Heard
Short Assessment of Patient Satisfaction (SAPS) [87] 14 5 0 1 Never Heard
Satisfaction with Amplification in Daily Life (SADL) [88] 11 9 0 0 Never Heard
Net Promoter Score (NPS) [89] 12 6 2 0 Never Heard
Social Isolation Measure (SIM) [90] 14 6 0 0 Never Heard
Table 4. Key findings from the CI user and CI professional final recommendation workshops. *CALD = Culturally and Linguistically Diverse.
Table 4. Key findings from the CI user and CI professional final recommendation workshops. *CALD = Culturally and Linguistically Diverse.
CI Professionals CI Users
General Assessment of all 3 supradomains is important Assessment of all 3 supradomains is important
Asynchronous remote assessments must consider the amount of time required for the user to complete
-
Completion can be burdensome on the user
Service
Supra-domain
Preference for simplicity of measurement
-
1-item measures per domain
-
Likert/binary outcome scale
-
Use of automated methods of data collection to reduce burden of collection and completion
Preference for simplicity of measurement
-
1-item measures per domain
-
Likert/star rating scale
-
Option to expand on answer
-
Outcomes should be assessed immediately after service use to ensure responses are contextual
No more than 3-4 times a year
Technology for remote services should be accessible for everyone
-
Consider CALD*
-
Important to assess CI users’ ability to use the remote service
The need to do so is likely to reduce in future with increased familiarity with technology
Technology for remote services should be accessible for everyone
-
May depend on end user connectivity
-
CI users should be able to complete remote care sessions independently if required.
-
Digital literacy is an important consideration
Clinical
Supra-domain
System checks are essential
-
outcomes are dependent on working hardware
System checks are essential
-
should be routine
-
Could be performed 1/month without need for patient feedback
may catch issues quicker than client self-report
Consider sending a status report to the client
Preference for a minimalist approach focusing on a few key measures
-
Recommendation to assess device use and speech perception in noise
CI outcomes dependant on CI use
Preference for a minimalist approach focusing on a few key measures
-
Adaptive tests are often quicker
-
Device use
important to measure but ultimately up to the CI user to determine how much they wear their device.
Preference for datalogging rather than self-report
Speech perception tests
-
should be suitable for a range of hearing abilities and be “real-world” applicable
speech in noise
adaptive speech tests
need to keep abreast of tests in development as they may be more appropriate (e.g. ECO-SIN test)
-
Important to differentiate between diagnostic measures (e.g. confusion matrices to identify which speech sounds are not perceived) which may only be required for some individuals at certain times, and functional outcome measures that assess overall ability to follow speech in quiet and noisy conditions and are often used to track general progress of both individuals and CI groups as a whole.
Speech perception tests
-
Should be “real-world” applicable
Historical testing
-
Recognition that some outcome measures (e.g. speech in quiet) persist for historical reasons
CI users and CI clinicians to compare outcomes over time
Use in retrospective outcomes research
Historical testing
-
important to be able to compare current and previously measured results to view progress over time
Remote test environment
-
must be considered when implementing speech test measures remotely
-
replication of the same test environment may not be possible
If this doesn’t matter, should be communicated to the CI user
Patient
Supra-domain
PROMs (patient reported outcome measures)
-
must be practical to implement and use in clinic. Need to consider;
length and ease of administration
use of a mix of broad and specific PROMs for future COS
PROMs (patient reported outcome measures)
-
need to be short and quick to complete
Maximum 20-25 items
No mandatory free text items but there should be free text options
Shouldn’t include multiple items asking similar things
Must use simple language
-
CI users must be made aware that PROMs need to be completed prior to the appointment
Mental Health and Well-being
-
can both affect and be affected by hearing loss
-
assessment of this area is vital for provision of holistic care and support
-
may be difficult to distinguish hearing loss related mental health issues from those caused by other life stressors
hearing-related mental health tools are vital
Subjective hearing disability
-
Concerns PROMs measuring subjective hearing disability are not sensitive enough to pick up deterioration in performance
-
Perception that a speech test that reflects “real-life” situations may be more accurate
Satisfaction with CI
-
Assessment of satisfaction is crucial because it reflects overall quality of life as well as effectiveness of the CI
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated