Videogames have consistently become one of the main forms of entertainment in recent years. Being a type of product that comprises several technological and artistical aspects, such as computer graphics, video and audio design, music composition, etc., a significant amount of research effort from various scientific fields has been devoted towards videogame-related problems. More recently, open-world videogames, characterized by nonlinear storylines and multiple gameplay scenarios and possibilities, have become increasingly popular. In such games, the music has to consider an enormous amount of events and variations, consequently it is extremely hard for human composers to write by themselves the score for every potential situation. Therefore, following the improvements in deep learning techniques, we propose a method for procedural music generation for videogames conditioned on emotions, especially suitable to open-world games. Specifically, we propose a pipeline where emotions, as represented on the valence-arousal plane, are extracted from gameplay videos, following the assumption that these values have a correspondence with the ones felt by the player. We then use the emotion-related information to condition a music transformer architecture in order to generate MIDI tracks, whose emotional content follows that of the game. We perform a perceptual experiment with human players in order to demonstrate the effectiveness of the proposed technique and to further investigate the applicability of this procedure to the videogame music generation research problem.

Procedural music generation for videogames conditioned through video emotion recognition

Comanducci L.;Zanoni M.;Bernardini A.;Antonacci F.;Sarti A.
2023-01-01

Abstract

Videogames have consistently become one of the main forms of entertainment in recent years. Being a type of product that comprises several technological and artistical aspects, such as computer graphics, video and audio design, music composition, etc., a significant amount of research effort from various scientific fields has been devoted towards videogame-related problems. More recently, open-world videogames, characterized by nonlinear storylines and multiple gameplay scenarios and possibilities, have become increasingly popular. In such games, the music has to consider an enormous amount of events and variations, consequently it is extremely hard for human composers to write by themselves the score for every potential situation. Therefore, following the improvements in deep learning techniques, we propose a method for procedural music generation for videogames conditioned on emotions, especially suitable to open-world games. Specifically, we propose a pipeline where emotions, as represented on the valence-arousal plane, are extracted from gameplay videos, following the assumption that these values have a correspondence with the ones felt by the player. We then use the emotion-related information to condition a music transformer architecture in order to generate MIDI tracks, whose emotional content follows that of the game. We perform a perceptual experiment with human players in order to demonstrate the effectiveness of the proposed technique and to further investigate the applicability of this procedure to the videogame music generation research problem.
2023
2023 4th International Symposium on the Internet of Sounds, ISIoS 2023
Human-centered computing
procedural music generation
Videogame audio
File in questo prodotto:
File Dimensione Formato  
Procedural_music_generation_for_videogames_conditioned_through_video_emotion_recognition.pdf

Accesso riservato

: Publisher’s version
Dimensione 1.53 MB
Formato Adobe PDF
1.53 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1260901
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact