JSAI2019

Presentation information

Interactive Session

[3Rin2] Interactive Session 1

Thu. Jun 6, 2019 10:30 AM - 12:10 PM Room R (Center area of 1F Exhibition hall)

10:30 AM - 12:10 PM

[3Rin2-32] An anti-noise performance comparison between acoustic features in detecting voice pathology using machine learning

〇Kouta Suzuki1,2, Shuji Shinohara2, Nobuhito Manome1,2, Kosuke Tomonaga1,2, Shunji Mitsuyoshi2 (1. SoftBank Robotics Corp., 2. Graduate School of Engineering, The University of Tokyo)

Keywords:Machine Learning, Voice Recognition, Voice Pathology

Developing communication robots requires to analyze human voice including various kinds of human biological information because the nonverbal information plays an important role in smooth communications between humans and robots. To analyze numerous voices available via the robot by using machine learning, we should take consideration of the existence of noises added to the voices. However, some acoustic features used for sensing human biological information is not designed for the noises. To validate the variation of the accuracy of classification when the voices includes the noises, we compare the classifications using voice indexes proposed for voice pathology estimation and using Mel-Frequency Cepstrum Coefficients(MFCC) in the classification problem of voice pathology as an early study. Experimental results show that classification using MFCC can detect voice pathology more precisely despite the noises while other voice indexes are adversely affected by the noises.