2021 IEEE International Conference on Acoustics, Speech and Signal Processing

6-11 June 2021 • Toronto, Ontario, Canada

Extracting Knowledge from Information

2021 IEEE International Conference on Acoustics, Speech and Signal Processing

6-11 June 2021 • Toronto, Ontario, Canada

Extracting Knowledge from Information

Technical Program

Paper Detail

Paper IDSPE-57.6
Paper Title END-2-END MODELING OF SPEECH AND GAIT FROM PATIENTS WITH PARKINSON'S DISEASE: COMPARISON BETWEEN HIGH QUALITY VS. SMARTPHONE DATA
Authors Juan Camilo Vásquez-Correa, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Germany; Tomás Arias-Vergara, Ludwig-Maximilians University, Germany; Philipp Klumpp, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Germany; Paula Andrea Perez-Toro, Juan Rafael Orozco-Arroyave, Universidad de Antioquia, Germany; Elmar Nöth, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Germany
SessionSPE-57: Speech, Depression and Sleepiness
LocationGather.Town
Session Time:Friday, 11 June, 14:00 - 14:45
Presentation Time:Friday, 11 June, 14:00 - 14:45
Presentation Poster
Topic Speech Processing: [SPE-ANLS] Speech Analysis
IEEE Xplore Open Preview  Click here to view in IEEE Xplore
Virtual Presentation  Click here to watch in the Virtual Conference
Abstract Parkinson's disease is a neurodegenerative disorder characterized by the presence of different motor impairments. Speech and gait signals have been analyzed to detect the presence of the disease and the severity in patients. However, most studies have been performed in controlled conditions using high quality data, which make those studies not suitable for a continuous at-home evaluation of the state of the patients. The developed technology should be evaluated in more realistic scenarios, for instance using smartphone data. We propose the use of state-of-the-art deep learning techniques to evaluate the speech and gait symptoms of patients. The proposed methods are evaluated in two scenarios to cover both high quality and smartphone data. The results indicate that it is possible to classify patients and healthy subjects with accuracies over 92% in both scenarios. The proposed methods are also promising to evaluate the severity of the speech symptoms and the global motor state of the patients.