JSAI2019

Presentation information

Organized Session

Organized Session » [OS] OS-11

[4F2-OS-11a] 社会的信号処理とAI(1)

Fri. Jun 7, 2019 12:00 PM - 1:40 PM Room F (302B Medium meeting room)

岡田 将吾(北陸先端科学技術大学院大学)、石井 亮(NTT)

12:20 PM - 12:40 PM

[4F2-OS-11a-02] Estimating Verbal・Nonverbal Skills in Business Presentation

〇Yagi Yutaro1, Okada Shogo1, Shiobara Shota2, Sugimura Sota2 (1. Japan Advanced Institute of Science and Technology, 2. SoftBank Corp.)

Keywords:Multimodal interaction

This paper focuses on developing a model for estimating presentation skills of each participant from multimodal (verbal and nonverbal) features. For this purpose, we use a multimodal presentation dataset including audio signal data and body motion sensor data, text data of speech contents of participants observed in 58 presentation sessions. The dataset also includes the presentation skills of each participant, which is assessed by two external observers of the Human Resources Department. We extracted various kinds of features such as spoken uttetances, acoustic features, and the amount of body motion to estimate the presentation skills. We created a regression model to infer the level of presentation skills from these features using support vector regression to evaluate the estimation accuracy of the presentation skills. Experiment results show that the multimodal model achieved 0.59 in R2 as the regression accuracy of effective production elements.