Cochlear Implants and Listening Effort: the Interaction of Cognitive and Sensory Constraints
2 other identifiers
interventional
460
1 country
2
Brief Summary
This study examines how cochlear implant users understand and comprehend speech in realistic communication situations. Through six experiments measuring listening effort via pupillometry and discourse comprehension, we will investigate how linguistic context, cognitive demands, and processing time affect speech understanding in CI users, and in normal-hearing controls) to identify factors underlying communication resilience versus vulnerability and develop improved, ecologically valid assessment and rehabilitation strategies.
Trial Health
Trial Health Score
Automated assessment based on enrollment pace, timeline, and geographic reach
participants targeted
Target at P75+ for not_applicable
Started Jan 2025
Longer than P75 for not_applicable
2 active sites
Health score is calculated from publicly available data and should be used for screening purposes only.
Trial Relationships
Click on a node to explore related trials.
Study Timeline
Key milestones and dates
Study Start
First participant enrolled
January 2, 2025
CompletedFirst Submitted
Initial submission to the registry
December 9, 2025
CompletedFirst Posted
Study publicly available on registry
December 12, 2025
CompletedPrimary Completion
Last participant's last visit for primary outcome
January 2, 2030
ExpectedStudy Completion
Last participant's last visit for all outcomes
January 2, 2030
December 12, 2025
December 1, 2025
5 years
December 9, 2025
December 9, 2025
Conditions
Keywords
Outcome Measures
Primary Outcomes (12)
Percent correct Consonant-Nucleus-Consonant (CNC) words across experiments 1-3
Participants' ability to recognize individual spoken words will be assessed using the Consonant-Nucleus-Consonant (CNC) word test. Outcome is based on the percent of key words correctly repeated. Higher score indicate better speech understanding.
End experiments 1-3 (up to 9 hours)
Percent correct AzBio sentences across experiments 1-3
The AzBio sentence test consists of 15 lists of 20 sentences each. AzBio sentences are spoken by different talkers in a conversational style with limited contextual cues that the listener can use to predict or 'fill in' unintelligible words. Outcome is based on the percent of AzBio sentences correctly repeated. Higher score indicate better speech understanding.
End experiments 1-3 (up to 9 hours)
Error rates for word identification tasks across experiments 1-3
End experiments 1-3 (up to 9 hours)
Percentage of propositions recalled from narrative passages
Subjects will hear 24 150- word passages, with one half of the passages being narrative (tell a story with a setting, theme, plot, etc.) and the other half being expository (e.g. instructions from a health care provider). After each passage the subject will be asked to recall as much of the passage content as possible. Higher score indicate better speech understanding.
End of experiment 4 (up to 3 hours)
Ratio of main idea recall to detail recall
Subjects will hear recorded narratives and will be asked to recall them as best as they can. The narratives will include main ideas (defined as propositions whose arguments are directly related to the overall meaning of the passage), mid-level propositions (those that take main propositions as their arguments), and details (propositions that take mid-level or other minor propositions as their arguments. Ratio of main idea recall to detail recall will be assessed.
End of experiment 4 (up to 3 hours)
Percentage of true/false statements identified
Subjects will hear sentences with syntactic forms known to be graded in comprehension difficulty: active-conjoined sentences (The author insulted the critic // and the critic hired a lawyer), subject-relative sentences (The author that insulted the critic // hired a lawyer), and object-relative sentences (The author that the critic insulted // hired a lawyer). The subject will hear an equal number of each sentence type, where half are presented with self-pacing, and half are presented without interruption. After each sentence the subject will be presented that they must indicate is true or false about the sentence they just heard.
End of experiment 6 (up to 3 hours)
Percent correct increase from unstructured word lists to anomalous sentences (syntactic gain)
Participants will hear 30 meaningful, eight-word sentences to begin. The words from this list of sentences will then be used to create the unstructured word lists and syntactic strings, such that each word would be heard equally as often across the three stimulus conditions. Therefore, each subject will hear a total of 30 meaningful sentences, 30 anomalous strings, and 30 unstructured lists, presented in a counter-balanced design. Recall accuracy for the three conditions will be compared.
End of experiment 1 (up to 3 hours)
Percent correct increase from anomalous sentences to meaningful sentences (semantic gain).
Participants will hear 30 meaningful, eight-word sentences to begin. The words from this list of sentences will then be used to create the unstructured word lists and syntactic strings, such that each word would be heard equally as often across the three stimulus conditions. Therefore, each subject will hear a total of 30 meaningful sentences, 30 anomalous strings, and 30 unstructured lists, presented in a counter-balanced design. Recall accuracy for the three conditions will be compared.
End of experiment 1 (up to 3 hours)
Number semantically-driven misrecognitions in lure conditions
Subjects will hear word pairs, with the second (target) word of the pair presented in multi-talker babble. There will be three conditions: Neutral prime (the first word is unrelated to the target word, e.g. Jaw-PASS), Semantic prime (the two words are semantic associates, e.g. Row - BOAT), and Semantic lure (the target word is a semantic associate of a similar word e.g. Row - GOAT, where GOAT is a lure for BOAT), thus putting the semantic context in conflict with successful perception. Number semantically-driven misrecognitions in lure conditions will be assessed.
End of experiment 2 (up to 3 hours)
Frequency of semantically-driven misrecognitions in lure conditions
Subjects will hear word pairs, with the second (target) word of the pair presented in multi-talker babble. There will be three conditions: Neutral prime (the first word is unrelated to the target word, e.g. Jaw-PASS), Semantic prime (the two words are semantic associates, e.g. Row - BOAT), and Semantic lure (the target word is a semantic associate of a similar word e.g. Row - GOAT, where GOAT is a lure for BOAT), thus putting the semantic context in conflict with successful perception. Frequency of semantically-driven misrecognitions in lure conditions will be assessed.
End of experiment 2 (up to 3 hours)
Difference in recall accuracy between single-sentence and two-sentence conditions
Subjects will undergo four test conditions (two will require remembering only one sentence and the other two will require remembering two sentences). For each condition there will be two combinations of semantic association between the first and second sentence (half will have high predictability, and the other half will have low predictability). Difference in recall accuracy between single-sentence and two-sentence conditions will be assessed.
End of experiment 3 (up to 3 hours)
Difference in comprehension improvement (percentage of proposition's correctly recalled) with self-paced vs. continuous presentation
Subjects will hear 24 150- word passages, with one half of the passages being narrative (tell a story with a setting, theme, plot, etc.) and the other half being expository (e.g. instructions from a health care provider). After each passage the subject will be asked to recall as much of the passage content as possible. Subjects will be presented with equal numbers of "self-paced" passages (stopping at major clauses and sentence boundaries, subject decides when to continue) and passages presented without interruption.
End of experiment 5 (up to 3 hours)
Study Arms (2)
Cochlear Implant Users
EXPERIMENTALPostlingually deaf adults (age 18-80) with at least one year of CI experience. Participants will complete behavioral speech perception and comprehension tasks with pupillometry measurement.
Normal-Hearing Controls (Vocoder Simulation)
ACTIVE COMPARATORNormal-hearing adults (age 18-80) listening to degraded speech via 4- and 8-channel vocoders. Participants will complete the same behavioral tasks as CI users but with (or without) acoustically degraded speech simulation.
Interventions
* Recall of meaningful sentences, anomalous word strings, and unstructured word lists * Measurement of syntactic and semantic gain * Pupillometry during auditory and visual presentation
* Two-choice word recognition task with semantic priming/luring in multi-talker babble * Three Signal-to-Noise Ratio (SNR) levels (heavy, medium, light noise) * Confidence ratings for responses * Pupillometry measurement
* Speech recognition and recall of single sentences vs. paired sentences * Manipulation of inter-sentence semantic predictability (high vs. low) * Four test conditions: 1-sentence, 2-sentences, 2-sentences+pre-prompt, 2-sentences+post-prompt * Pupillometry during task
* Recall of 27 narrative passages (67-97 words each) * Propositional analysis scoring (main ideas, mid-level ideas, details) * Measurement of semantic hierarchy effect * Pupillometry during listening
* 24 discourse passages (150 words each): 12 narrative, 12 expository * Continuous presentation vs. self-paced presentation (stops at clause/sentence boundaries) * Measurement of pause times and comprehension recall * Pupillometry during task
* Self-Paced Sentence Comprehension * Sentences with varying syntactic complexity (active-conjoined, subject-relative, object-relative) * Continuous vs. self-paced (with pause at major clause boundary) presentation * True/false comprehension verification statements * Pupillometry measurement
Eligibility Criteria
You may qualify if:
- Subjects will be otherwise healthy normal-hearing and cochlear implant (CI) adult listeners (between 18 and 80 years old).
You may not qualify if:
- Individuals below 18 years of age.
- Individuals with evidence of neurologic, vascular or psychiatric disease or dementia, and taking medications that might interfere with task performance.
- Individuals with a history of language disorders (besides those associated with hearing loss for the CI users). Individuals who are non-native speakers of American English.
Contact the study team to confirm eligibility.
Sponsors & Collaborators
Study Sites (2)
Brandeis University
Waltham, Massachusetts, 02453, United States
NYU Langone Health
New York, New York, 10016, United States
Study Officials
- PRINCIPAL INVESTIGATOR
Mario A. Svirsky, PhD
NYU Langone Health
Central Study Contacts
Study Design
- Study Type
- interventional
- Phase
- not applicable
- Allocation
- NON RANDOMIZED
- Masking
- NONE
- Purpose
- OTHER
- Intervention Model
- PARALLEL
- Sponsor Type
- OTHER
- Responsible Party
- SPONSOR
Study Record Dates
First Submitted
December 9, 2025
First Posted
December 12, 2025
Study Start
January 2, 2025
Primary Completion (Estimated)
January 2, 2030
Study Completion (Estimated)
January 2, 2030
Last Updated
December 12, 2025
Record last verified: 2025-12
Data Sharing
- IPD Sharing
- Will share
- Shared Documents
- STUDY PROTOCOL, SAP
- Time Frame
- Data will be made available as soon as possible or at the time of associated publications. All data to be shared will be shared by the close of the award. Data will be made available, at minimum, for seven years.
- Access Criteria
- The investigator who proposed to use the data will be granted access upon reasonable request. Data will be available for at least 7 years at a third party website (https://databrary.org/)
The de-identified participant data from the final research dataset will be shared upon reasonable request. Data will be made available as soon as possible or at the time of associated publications. All data to be shared will be shared by the close of the award. Data will be made available, at minimum, for seven years at a third party website (https://databrary.org/)