| Paper ID | SPE-30.7 | 
  
    | Paper Title | 
     A CAUSAL DEEP LEARNING FRAMEWORK FOR CLASSIFYING PHONEMES IN COCHLEAR IMPLANTS | 
  
	| Authors | 
    Kevin Chu, Leslie Collins, Boyla Mainsah, Duke University, United States | 
  | Session | SPE-30: Speech Processing 2: General Topics | 
  | Location | Gather.Town | 
  | Session Time: | Wednesday, 09 June, 16:30 - 17:15 | 
  | Presentation Time: | Wednesday, 09 June, 16:30 - 17:15 | 
  | Presentation | 
     Poster
     | 
	 | Topic | 
     Speech Processing: [SPE-GASR] General Topics in Speech Recognition | 
  
	
    | IEEE Xplore Open Preview | 
     Click here to view in IEEE Xplore | 
  
  
	
    | Virtual Presentation | 
     Click here to watch in the Virtual Conference | 
  
  
  
    | Abstract | 
     Speech intelligibility in cochlear implant (CI) users degrades considerably in listening environments with reverberation and noise. Previous research in automatic speech recognition (ASR) has shown that phoneme-based speech enhancement algorithms improve ASR system performance in reverberant environments as compared to a global model. However, phoneme-specific speech processing has not yet been implemented in CIs. In this paper, we propose a causal deep learning framework for classifying phonemes using features extracted at the time-frequency resolution of a CI processor. We trained and tested long short-term memory networks to classify phonemes and manner of articulation in anechoic and reverberant conditions. The results showed that CI-inspired features provide slightly higher levels of performance than traditional ASR features. To the best of our knowledge, this study is the first to provide a classification framework with the potential to categorize phonetic units in real-time in a CI. |