JSAI2024

Presentation information

Organized Session

Organized Session » OS-29

[1O4-OS-29a] OS-29

Tue. May 28, 2024 3:00 PM - 4:40 PM Room O (Music studio hall)

オーガナイザ:北原 鉄朗(日本大学)、中村 栄太(京都大学)、浜中 雅俊(理化学研究所)

3:40 PM - 4:00 PM

[1O4-OS-29a-03] An Approach to Emotion-based Music Generation using Diffusion Model

〇Moyu Kawabe1, Ichiro Kobayashi1 (1. Ochanomizu University)

Keywords:Diffusion Model, Music Generation, Emotion

Diffusion process-based models have been attracting attention in the field of music generation in recent years due to their high quality and scalability.
Research has also been conducted to generate music on demand using diffusion models. However, it is not easy to control for complex attributes in diffusion models. In addition, there have not been many studies on music generation with an emphasis on emotion, which is closely related to music.
In this study, we aim to develop a method that can generate a variety of music using a diffusion model, taking emotion as an input and controlling it according to the musical attributes corresponding to the emotion. For the diffusion model, we used the Diffusion-LM method, which can be controlled by using a classifier at each time denoising stage, and the classifier uses musical attribute values to identify emotions and generate music based on the input emotion information.

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password