2021 IEEE International Conference on Acoustics, Speech and Signal Processing

6-11 June 2021 • Toronto, Ontario, Canada

Extracting Knowledge from Information

2021 IEEE International Conference on Acoustics, Speech and Signal Processing

6-11 June 2021 • Toronto, Ontario, Canada

Extracting Knowledge from Information
Login Paper Search My Schedule Paper Index Help

My ICASSP 2021 Schedule

Note: Your custom schedule will not be saved unless you create a new account or login to an existing account.
  1. Create a login based on your email (takes less than one minute)
  2. Perform 'Paper Search'
  3. Select papers that you desire to save in your personalized schedule
  4. Click on 'My Schedule' to see the current list of selected papers
  5. Click on 'Printable Version' to create a separate window suitable for printing (the header and menu will appear, but will not actually print)

Paper Detail

Paper IDSPE-41.3
Paper Title SELF-ATTENTIVE VAD: CONTEXT-AWARE DETECTION OF VOICE FROM NOISE
Authors Yong Rae Jo, Voithru, South Korea; Young Ki Moon, Voithru, Inha University, South Korea; Won Ik Cho, Seoul National University, South Korea; Geun Sik Jo, Inha University, South Korea
SessionSPE-41: Voice Activity and Disfluency Detection
LocationGather.Town
Session Time:Thursday, 10 June, 15:30 - 16:15
Presentation Time:Thursday, 10 June, 15:30 - 16:15
Presentation Poster
Topic Speech Processing: [SPE-VAD] Voice Activity Detection and End-pointing
IEEE Xplore Open Preview  Click here to view in IEEE Xplore
Abstract Recent voice activity detection (VAD) schemes have aimed at leveraging the decent neural architectures, but few were successful with applying the attention network due to its high reliance on the encoder-decoder framework. This has often let the built systems have high dependency on the recurrent neural networks which are costly and sometimes less context-sensitive considering the scale and property of acoustic frames. To cope this issue with the self-attention mechanism and achieve a simple, powerful and environmentrobust VAD, we first adopt the self-attention architecture in building up the modules for voice detection and boosted prediction. Our model surpasses the previous neural architectures in view of low signal-to-ratio and noisy real-world scenarios, at the same time displaying the robustness regarding the noise types. We make the test labels on movie data publicly available for the fair competition and future progress.